Comments (2)
使用funasr里的镜像搭建环境,跑通了,可能跟我的环境有关。我的系统是CentOS Linux release 7.6.1810,我看镜像里的系统是Ubuntu 20.04.4 LTS
from funasr.
My system environment is:
system: Ubuntu 20.04.6 LTS
funasr.version == '1.0.16'
pytorch version = 2.2.1
I also encountered this problem, but setting nj to 0 in run.sh allowed me to run CMVN successfully.
However, I then encountered other errors in stage 4:
[2024-03-17 01:39:37,527] torch.distributed.run: [WARNING]
[2024-03-17 01:39:37,527] torch.distributed.run: [WARNING] *****************************************
[2024-03-17 01:39:37,527] torch.distributed.run: [WARNING] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
[2024-03-17 01:39:37,527] torch.distributed.run: [WARNING] *****************************************
If you want to use the speaker diarization, please pip install hdbscan
If you want to use the speaker diarization, please pip install hdbscan
{'model': 'Paraformer', 'model_conf': {'ctc_weight': 0.3, 'lsm_weight': 0.1, 'length_normalized_loss': False, 'predictor_weight': 1.0, 'sampling_ratio': 0.4, 'use_1st_decoder_loss': True}, 'encoder': 'ConformerEncoder', 'encoder_conf': {'output_size': 256, 'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 12, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'attention_dropout_rate': 0.0, 'input_layer': 'conv2d', 'normalize_before': True, 'pos_enc_layer_type': 'rel_pos', 'selfattention_layer_type': 'rel_selfattn', 'activation_type': 'swish', 'macaron_style': True, 'use_cnn_module': True, 'cnn_module_kernel': 15}, 'decoder': 'ParaformerSANDecoder', 'decoder_conf': {'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 6, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'self_attention_dropout_rate': 0.0, 'src_attention_dropout_rate': 0.0}, 'predictor': 'CifPredictor', 'predictor_conf': {'idim': 256, 'threshold': 1.0, 'l_order': 1, 'r_order': 1, 'tail_threshold': 0.45}, 'frontend': 'WavFrontend', 'frontend_conf': {'fs': 16000, 'window': 'hamming', 'n_mels': 80, 'frame_length': 25, 'frame_shift': 10, 'lfr_m': 1, 'lfr_n': 1, 'cmvn_file': '../DATA/data/train/am.mvn'}, 'specaug': 'SpecAug', 'specaug_conf': {'apply_time_warp': True, 'time_warp_window': 5, 'time_warp_mode': 'bicubic', 'apply_freq_mask': True, 'freq_mask_width_range': [0, 30], 'num_freq_mask': 2, 'apply_time_mask': True, 'time_mask_width_range': [0, 40], 'num_time_mask': 2}, 'train_conf': {'accum_grad': 1, 'grad_clip': 5, 'max_epoch': 150, 'keep_nbest_models': 10, 'avg_nbest_model': 5, 'log_interval': 50}, 'optim': 'adam', 'optim_conf': {'lr': 0.0005}, 'scheduler': 'warmuplr', 'scheduler_conf': {'warmup_steps': 30000}, 'dataset': 'AudioDataset', 'dataset_conf': {'index_ds': 'IndexDSJsonl', 'batch_sampler': 'RankFullLocalShuffleBatchSampler', 'batch_type': 'example', 'batch_size': 32, 'max_token_length': 2048, 'buffer_size': 1024, 'shuffle': True, 'num_workers': 4, 'preprocessor_speech': 'SpeechPreprocessSpeedPerturb', 'preprocessor_speech_conf': {'speed_perturb': [0.9, 1.0, 1.1]}}, 'tokenizer': 'CharTokenizer', 'tokenizer_conf': {'unk_symbol': '', 'token_list': '../DATA/data/zh_token_list/char/tokens.txt'}, 'ctc_conf': {'dropout_rate': 0.0, 'ctc_type': 'builtin', 'reduce': True, 'ignore_nan_grad': True}, 'normalize': None, 'train_data_set_list': '../DATA/data/train/audio_datasets.jsonl', 'valid_data_set_list': '../DATA/data/dev/audio_datasets.jsonl', 'output_dir': '/alghome/timmy.wan/whisper/lab/VariousLargeWhisper/FunASR/examples/aishell/paraformer/exp/baseline_paraformer_conformer_12e_6d_2048_256_zh_char_exp1'}
{'model': 'Paraformer', 'model_conf': {'ctc_weight': 0.3, 'lsm_weight': 0.1, 'length_normalized_loss': False, 'predictor_weight': 1.0, 'sampling_ratio': 0.4, 'use_1st_decoder_loss': True}, 'encoder': 'ConformerEncoder', 'encoder_conf': {'output_size': 256, 'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 12, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'attention_dropout_rate': 0.0, 'input_layer': 'conv2d', 'normalize_before': True, 'pos_enc_layer_type': 'rel_pos', 'selfattention_layer_type': 'rel_selfattn', 'activation_type': 'swish', 'macaron_style': True, 'use_cnn_module': True, 'cnn_module_kernel': 15}, 'decoder': 'ParaformerSANDecoder', 'decoder_conf': {'attention_heads': 4, 'linear_units': 2048, 'num_blocks': 6, 'dropout_rate': 0.1, 'positional_dropout_rate': 0.1, 'self_attention_dropout_rate': 0.0, 'src_attention_dropout_rate': 0.0}, 'predictor': 'CifPredictor', 'predictor_conf': {'idim': 256, 'threshold': 1.0, 'l_order': 1, 'r_order': 1, 'tail_threshold': 0.45}, 'frontend': 'WavFrontend', 'frontend_conf': {'fs': 16000, 'window': 'hamming', 'n_mels': 80, 'frame_length': 25, 'frame_shift': 10, 'lfr_m': 1, 'lfr_n': 1, 'cmvn_file': '../DATA/data/train/am.mvn'}, 'specaug': 'SpecAug', 'specaug_conf': {'apply_time_warp': True, 'time_warp_window': 5, 'time_warp_mode': 'bicubic', 'apply_freq_mask': True, 'freq_mask_width_range': [0, 30], 'num_freq_mask': 2, 'apply_time_mask': True, 'time_mask_width_range': [0, 40], 'num_time_mask': 2}, 'train_conf': {'accum_grad': 1, 'grad_clip': 5, 'max_epoch': 150, 'keep_nbest_models': 10, 'avg_nbest_model': 5, 'log_interval': 50}, 'optim': 'adam', 'optim_conf': {'lr': 0.0005}, 'scheduler': 'warmuplr', 'scheduler_conf': {'warmup_steps': 30000}, 'dataset': 'AudioDataset', 'dataset_conf': {'index_ds': 'IndexDSJsonl', 'batch_sampler': 'RankFullLocalShuffleBatchSampler', 'batch_type': 'example', 'batch_size': 32, 'max_token_length': 2048, 'buffer_size': 1024, 'shuffle': True, 'num_workers': 4, 'preprocessor_speech': 'SpeechPreprocessSpeedPerturb', 'preprocessor_speech_conf': {'speed_perturb': [0.9, 1.0, 1.1]}}, 'tokenizer': 'CharTokenizer', 'tokenizer_conf': {'unk_symbol': '', 'token_list': '../DATA/data/zh_token_list/char/tokens.txt'}, 'ctc_conf': {'dropout_rate': 0.0, 'ctc_type': 'builtin', 'reduce': True, 'ignore_nan_grad': True}, 'normalize': None, 'train_data_set_list': '../DATA/data/train/audio_datasets.jsonl', 'valid_data_set_list': '../DATA/data/dev/audio_datasets.jsonl', 'output_dir': '/alghome/timmy.wan/whisper/lab/VariousLargeWhisper/FunASR/examples/aishell/paraformer/exp/baseline_paraformer_conformer_12e_6d_2048_256_zh_char_exp1'}
tables:
----------- ** dataset_classes ** --------------
| register name | class name | class location |
| AudioDataset | AudioDataset | funasr/datasets/audio_datasets/datasets.py:8 |
| AudioDatasetHotword | AudioDatasetHotword | funasr/datasets/audio_datasets/datasets.py:103 |
| AudioLLMARDataset | AudioLLMARDataset | funasr/datasets/llm_datasets/datasets.py:273 |
| AudioLLMDataset | AudioLLMDataset | funasr/datasets/llm_datasets/datasets.py:147 |
| AudioLLMNARDataset | AudioLLMNARDataset | funasr/datasets/llm_datasets/datasets.py:8 |
| AudioLLMQwenAudioDataset | AudioLLMQwenAudioDataset | funasr/datasets/llm_datasets_qwenaudio/datasets.py:8 |
| AudioLLMVicunaDataset | AudioLLMVicunaDataset | funasr/datasets/llm_datasets_vicuna/datasets.py:8 |
----------- ** index_ds_classes ** --------------
| register name | class name | class location |
| IndexDSJsonl | IndexDSJsonlRankFull | funasr/datasets/audio_datasets/index_ds.py:73 |
| IndexDSJsonlRankFull | IndexDSJsonlRankFull | funasr/datasets/audio_datasets/index_ds.py:73 |
| IndexDSJsonlRankSplit | IndexDSJsonlRankSplit | funasr/datasets/audio_datasets/index_ds.py:12 |
----------- ** preprocessor_classes ** --------------
| register name | class name | class location |
| SpeechPreprocessSpeedPerturb | SpeechPreprocessSpeedPerturb | funasr/datasets/audio_datasets/preprocessor.py:18 |
| TextPreprocessRemovePunctuation | TextPreprocessRemovePunctuation | funasr/datasets/llm_datasets/preprocessor.py:20 |
| TextPreprocessSegDict | TextPreprocessSegDict | funasr/datasets/audio_datasets/preprocessor.py:38 |
----------- ** batch_sampler_classes ** --------------
| register name | class name | class location |
| BatchSampler | RankFullLocalShuffleBatchSampler | funasr/datasets/llm_datasets/samplers.py:88 |
| CustomDistributedBatchSampler | CustomDistributedBatchSampler | funasr/datasets/llm_datasets_vicuna/samplers.py:154 |
| CustomDistributedBatchSampler_fn | CustomDistributedBatchSampler_fn | funasr/datasets/llm_datasets_vicuna/samplers.py:145 |
| DistributedSamplerWarp | DistributedSamplerWarp | funasr/datasets/llm_datasets_vicuna/samplers.py:104 |
| DynamicBatchLocalShuffleSampler | BatchSampler | funasr/datasets/llm_datasets/samplers.py:9 |
| RankFullGlobalShuffleBatchSampler | RankFullGlobalShuffleBatchSampler | funasr/datasets/llm_datasets_vicuna/samplers.py:13 |
| RankFullLocalShuffleBatchSampler | RankFullLocalShuffleBatchSampler | funasr/datasets/llm_datasets/samplers.py:88 |
| RankFullLocalShuffleDynamicBatchSampler | RankFullLocalShuffleDynamicBatchSampler | funasr/datasets/llm_datasets/samplers.py:181 |
----------- ** frontend_classes ** --------------
| register name | class name | class location |
| DefaultFrontend | DefaultFrontend | funasr/frontends/default.py:21 |
| WavFrontend | WavFrontend | funasr/frontends/wav_frontend.py:78 |
| WavFrontendOnline | WavFrontendOnline | funasr/frontends/wav_frontend.py:216 |
| WhisperFrontend | WhisperFrontend | funasr/frontends/whisper_frontend.py:10 |
----------- ** joint_network_classes ** --------------
| register name | class name | class location |
| joint_network | JointNetwork | funasr/models/transducer/joint_network.py:12 |
----------- ** model_classes ** --------------
| register name | class name | class location |
| BAT | BAT | funasr/models/bat/model.py:35 |
| BiCifParaformer | BiCifParaformer | funasr/models/bicif_paraformer/model.py:37 |
| Branchformer | Branchformer | funasr/models/branchformer/model.py:6 |
| CAMPPlus | CAMPPlus | funasr/models/campplus/model.py:30 |
| CTTransformer | CTTransformer | funasr/models/ct_transformer/model.py:33 |
| CTTransformerStreaming | CTTransformerStreaming | funasr/models/ct_transformer_streaming/model.py:27 |
| Conformer | Conformer | funasr/models/conformer/model.py:8 |
| ContextualParaformer | ContextualParaformer | funasr/models/contextual_paraformer/model.py:40 |
| EBranchformer | EBranchformer | funasr/models/e_branchformer/model.py:6 |
| Emotion2vec | Emotion2vec | funasr/models/emotion2vec/model.py:34 |
| FsmnVADStreaming | FsmnVADStreaming | funasr/models/fsmn_vad_streaming/model.py:267 |
| LCBNet | LCBNet | funasr/models/lcbnet/model.py:26 |
| LLMASR | LLMASR | funasr/models/llm_asr/model.py:23 |
| LLMASRNAR | LLMASRNAR | funasr/models/llm_asr_nar/model.py:24 |
| LLMASRNARPrompt | LLMASRNARPrompt | funasr/models/llm_asr_nar/model.py:353 |
| MonotonicAligner | MonotonicAligner | funasr/models/monotonic_aligner/model.py:24 |
| OpenAIWhisperLIDModel | OpenAIWhisperLIDModel | funasr/models/whisper_lid/model.py:453 |
| OpenAIWhisperModel | OpenAIWhisperModel | funasr/models/whisper_lid/model.py:21 |
| Paraformer | Paraformer | funasr/models/paraformer/model.py:27 |
| ParaformerStreaming | ParaformerStreaming | funasr/models/paraformer_streaming/model.py:37 |
| Qwen-Audio | QwenAudioWarp | funasr/models/qwen_audio/model.py:17 |
| Qwen-Audio-Chat | QwenAudioChatWarp | funasr/models/qwen_audio/model.py:72 |
| Qwen/Qwen-Audio | QwenAudioWarp | funasr/models/qwen_audio/model.py:17 |
| Qwen/Qwen-Audio-Chat | QwenAudioChatWarp | funasr/models/qwen_audio/model.py:72 |
| Qwen/QwenAudio | QwenAudioWarp | funasr/models/qwen_audio/model.py:17 |
| Qwen/QwenAudioChat | QwenAudioChatWarp | funasr/models/qwen_audio/model.py:72 |
| QwenAudio | QwenAudioWarp | funasr/models/qwen_audio/model.py:17 |
| QwenAudioChat | QwenAudioChatWarp | funasr/models/qwen_audio/model.py:72 |
| QwenAudioChatWarp | QwenAudioChatWarp | funasr/models/qwen_audio/model.py:72 |
| QwenAudioWarp | QwenAudioWarp | funasr/models/qwen_audio/model.py:17 |
| SANM | SANM | funasr/models/sanm/model.py:13 |
| SCAMA | SCAMA | funasr/models/scama/model.py:38 |
| SeacoParaformer | SeacoParaformer | funasr/models/seaco_paraformer/model.py:43 |
| Transducer | Transducer | funasr/models/transducer/model.py:34 |
| Transformer | Transformer | funasr/models/transformer/model.py:20 |
| UniASR | UniASR | funasr/models/uniasr/model.py:25 |
| Whisper-base | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-base.en | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-large-v1 | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-large-v2 | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-large-v3 | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-medium | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-medium.en | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-small | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-small.en | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-tiny | WhisperWarp | funasr/models/whisper/model.py:16 |
| Whisper-tiny.en | WhisperWarp | funasr/models/whisper/model.py:16 |
| WhisperWarp | WhisperWarp | funasr/models/whisper/model.py:16 |
----------- ** predictor_classes ** --------------
| register name | class name | class location |
| CifPredictor | CifPredictor | funasr/models/paraformer/cif_predictor.py:15 |
| CifPredictorV2 | CifPredictorV2 | funasr/models/paraformer/cif_predictor.py:143 |
| CifPredictorV2Export | CifPredictorV2Export | funasr/models/paraformer/cif_predictor.py:378 |
| CifPredictorV3 | CifPredictorV3 | funasr/models/bicif_paraformer/cif_predictor.py:95 |
| CifPredictorV3Export | CifPredictorV3Export | funasr/models/bicif_paraformer/cif_predictor.py:340 |
----------- ** encoder_classes ** --------------
| register name | class name | class location |
| BranchformerEncoder | BranchformerEncoder | funasr/models/branchformer/encoder.py:294 |
| ChunkConformerEncoder | ConformerChunkEncoder | funasr/models/conformer/encoder.py:893 |
| ConformerEncoder | ConformerEncoder | funasr/models/conformer/encoder.py:287 |
| ConvBiasPredictor | ConvPredictor | funasr/models/lcbnet/encoder.py:358 |
| DFSMN | DFSMN | funasr/models/fsmn_vad_streaming/encoder.py:318 |
| EBranchformerEncoder | EBranchformerEncoder | funasr/models/e_branchformer/encoder.py:177 |
| FSMN | FSMN | funasr/models/fsmn_vad_streaming/encoder.py:181 |
| FSMNExport | FSMNExport | funasr/models/fsmn_vad_streaming/encoder.py:242 |
| FusionSANEncoder | SelfSrcAttention | funasr/models/lcbnet/encoder.py:232 |
| OpenAIWhisperEncoderWarp | OpenAIWhisperEncoderWarp | funasr/models/whisper_lid/encoder.py:17 |
| QwenAudioEncoder | QwenAudioEncoder | funasr/models/qwen_audio/audio.py:338 |
| RWKVEncoder | RWKVEncoder | funasr/models/rwkv_bat/rwkv_encoder.py:16 |
| SANMEncoder | SANMEncoder | funasr/models/sanm/encoder.py:161 |
| SANMEncoderChunkOpt | SANMEncoderChunkOpt | funasr/models/scama/encoder.py:162 |
| SANMEncoderChunkOptExport | SANMEncoderExport | funasr/models/sanm/encoder.py:487 |
| SANMEncoderExport | SANMEncoderExport | funasr/models/sanm/encoder.py:487 |
| SANMVadEncoder | SANMVadEncoder | funasr/models/ct_transformer_streaming/encoder.py:148 |
| SANMVadEncoderExport | SANMVadEncoderExport | funasr/models/ct_transformer_streaming/encoder.py:404 |
| TransformerEncoder | TransformerEncoder | funasr/models/transformer/encoder.py:139 |
| TransformerTextEncoder | TransformerTextEncoder | funasr/models/lcbnet/encoder.py:130 |
----------- ** decoder_classes ** --------------
| register name | class name | class location |
| ContextualParaformerDecoder | ContextualParaformerDecoder | funasr/models/contextual_paraformer/decoder.py:103 |
| ContextualParaformerDecoderExport | ContextualParaformerDecoderExport | funasr/models/contextual_paraformer/decoder.py:309 |
| DynamicConvolution2DTransformerDecoder | DynamicConvolution2DTransformerDecoder | funasr/models/sa_asr/transformer_decoder.py:703 |
| DynamicConvolutionTransformerDecoder | DynamicConvolutionTransformerDecoder | funasr/models/sa_asr/transformer_decoder.py:642 |
| FsmnDecoder | FsmnDecoder | funasr/models/sanm/decoder.py:198 |
| FsmnDecoderSCAMAOpt | FsmnDecoderSCAMAOpt | funasr/models/scama/decoder.py:197 |
| LightweightConvolution2DTransformerDecoder | LightweightConvolution2DTransformerDecoder | funasr/models/sa_asr/transformer_decoder.py:580 |
| LightweightConvolutionTransformerDecoder | LightweightConvolutionTransformerDecoder | funasr/models/sa_asr/transformer_decoder.py:519 |
| OpenAIWhisperDecoderWarp | OpenAIWhisperDecoderWarp | funasr/models/whisper_lid/decoder.py:15 |
| ParaformerDecoderSAN | ParaformerDecoderSAN | funasr/models/sa_asr/transformer_decoder.py:405 |
| ParaformerDecoderSANExport | ParaformerDecoderSANExport | funasr/models/paraformer/decoder.py:1093 |
| ParaformerSANDecoder | ParaformerSANDecoder | funasr/models/paraformer/decoder.py:979 |
| ParaformerSANMDecoder | ParaformerSANMDecoder | funasr/models/paraformer/decoder.py:220 |
| ParaformerSANMDecoderExport | ParaformerSANMDecoderExport | funasr/models/paraformer/decoder.py:637 |
| ParaformerSANMDecoderOnlineExport | ParaformerSANMDecoderOnlineExport | funasr/models/paraformer/decoder.py:823 |
| TransformerDecoder | TransformerDecoder | funasr/models/sa_asr/transformer_decoder.py:356 |
| rnn_decoder | RNNDecoder | funasr/models/transducer/rnn_decoder.py:86 |
| rnnt_decoder | RNNTDecoder | funasr/models/transducer/rnnt_decoder.py:14 |
----------- ** adaptor_classes ** --------------
| register name | class name | class location |
| Linear | Linear | funasr/models/llm_asr_nar/adaptor.py:6 |
| QFormer | EncoderProjectorQFormer | funasr/models/llm_asr/adaptor.py:31 |
----------- ** normalize_classes ** --------------
| register name | class name | class location |
| GlobalMVN | GlobalMVN | funasr/models/normalize/global_mvn.py:11 |
| UtteranceMVN | UtteranceMVN | funasr/models/normalize/utterance_mvn.py:8 |
----------- ** specaug_classes ** --------------
| register name | class name | class location |
| SpecAug | SpecAug | funasr/models/specaug/specaug.py:14 |
| SpecAugLFR | SpecAugLFR | funasr/models/specaug/specaug.py:104 |
----------- ** lid_predictor_classes ** --------------
| register name | class name | class location |
| LidPredictor | LidPredictor | funasr/models/whisper_lid/lid_predictor.py:5 |
----------- ** tokenizer_classes ** --------------
| register name | class name | class location |
| CharTokenizer | CharTokenizer | funasr/tokenizer/char_tokenizer.py:11 |
| HuggingfaceTokenizer | HuggingfaceTokenizer | funasr/tokenizer/hf_tokenizer.py:10 |
| SentencepiecesTokenizer | SentencepiecesTokenizer | funasr/tokenizer/sentencepiece_tokenizer.py:11 |
| WhisperTokenizer | WhisperTokenizer | funasr/tokenizer/whisper_tokenizer.py:9 |
[2024-03-17 01:39:43,764][root][WARNING] - Using legacy_rel_pos and it will be deprecated in the future.
[2024-03-17 01:39:43,776][root][INFO] - config.yaml is saved to: /alghome/timmy.wan/whisper/lab/VariousLargeWhisper/FunASR/examples/aishell/paraformer/exp/baseline_paraformer_conformer_12e_6d_2048_256_zh_char_exp1/config.yaml
[2024-03-17 01:39:43,780][root][WARNING] - Using legacy_rel_pos and it will be deprecated in the future.
[2024-03-17 01:39:43,800][root][WARNING] - Using legacy_rel_selfattn and it will be deprecated in the future.
[2024-03-17 01:39:43,816][root][WARNING] - Using legacy_rel_selfattn and it will be deprecated in the future.
No initialize method
No initialize method
[2024-03-17 01:39:44,941][root][INFO] - total_num of samplers across ranks: 120098
[2024-03-17 01:39:44,944][root][INFO] - total_num of samplers across ranks: 120098
[2024-03-17 01:39:44,996][root][INFO] - total_num of samplers across ranks: 14326
No checkpoint found at '/alghome/timmy.wan/whisper/lab/VariousLargeWhisper/FunASR/examples/aishell/paraformer/exp/baseline_paraformer_conformer_12e_6d_2048_256_zh_char_exp1/model.pt', does not resume status!
[2024-03-17 01:39:45,000][root][INFO] - total_num of samplers across ranks: 14326
No checkpoint found at '/alghome/timmy.wan/whisper/lab/VariousLargeWhisper/FunASR/examples/aishell/paraformer/exp/baseline_paraformer_conformer_12e_6d_2048_256_zh_char_exp1/model.pt', does not resume status!
rank: 0, Training Epoch: 1: 0%|�[34m �[0m| 0/1877 [00:00<?, ?it/s]
rank: 1, Training Epoch: 1: 0%|�[34m �[0m| 0/1877 [00:00<?, ?it/s]ERROR: Unexpected segmentation fault encountered in worker.
from funasr.
Related Issues (20)
- Funasr如何在Ubuntu18.04.6LTS上导出预训练模型的ONNX? HOT 1
- 微信群满了,超过200人加不进去了,求加 HOT 1
- 如何通过FunASR的docker部署方法支持SenseVoice模型?
- wss_server 启动连接modelscope 失败,模型已经下载到本地,为什么启动的时候要连接modelscope? HOT 3
- FunASR离线文件转写服务GPU版本docker部署seaco_paraformer报bug HOT 1
- 关于online模型导出onnx报错代码bug HOT 1
- paraformer在CUDA 12.1/cuDNN 8.9.2推理时出现“段错误 (核心已转储)”报错
- FunASR 离线 GPU版本 使用FunASR/runtime/run_server.sh 启动后,语音能成功解析,但是nvidia-smi看不到对应进程,怎么确定是真实使用gpu?? HOT 3
- funasr_onnx 还不支持 SenseVoice 模型吗 HOT 3
- Websocket 服务中针对不同音频识别结果差异巨大 HOT 4
- 多线程去做实时语音识别操作报错问题求助 HOT 2
- I can't find data prepared part recipe codes about sond diarization model HOT 7
- 实时语音识别结合标点检测效果不好,是正常的吗 HOT 5
- 一定要安装docker才能在服务器上部署funasr吗 HOT 2
- 使用离线paraformer-large-长音频版的模型进行推理,出现Segmentation fault HOT 2
- model not registered HOT 1
- 多线程实时语音识别,结果会互相干扰 [CPU] HOT 2
- AssertionError: C:\.cache\modelscope\hub\damo\speech_fsmn_vad_zh-cn-16k-common-pytorch is not registered HOT 1
- 【Error】AssertionError: choose a window size 400 that is [2, 0] HOT 2
- funasr-onnx-offline-rtf使用gpu推理遇到问题
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from funasr.