The program got stuck while executing the script test.en-x.st.sh
hus18 opened this issue · 1 comments
2023-07-12 20:20:51 | INFO | fairseq.models.hubert.hubert | HubertModel Config: {'_name': 'hubert', 'label_rate': 50.0, 'extractor_mode': default, 'encoder_layers': 12, 'encoder_embed_dim': 768, 'encoder_ffn_embed_dim': 3072, 'encoder_attention_heads': 12, 'activation_fn': gelu, 'layer_type': transformer, 'dropout': 0.1, 'attention_dropout': 0.1, 'activation_dropout': 0.0, 'encoder_layerdrop': 0.05, 'dropout_input': 0.1, 'dropout_features': 0.1, 'final_dim': 256, 'untie_final_proj': False, 'layer_norm_first': False, 'conv_feature_layers': '[(512,10,5)] + [(512,3,2)] * 4 + [(512,2,2)] * 2', 'conv_bias': False, 'logit_temp': 0.1, 'target_glu': False, 'feature_grad_mult': 0.1, 'mask_length': 10, 'mask_prob': 0.8, 'mask_selection': static, 'mask_other': 0.0, 'no_mask_overlap': False, 'mask_min_space': 1, 'mask_channel_length': 10, 'mask_channel_prob': 0.0, 'mask_channel_selection': static, 'mask_channel_other': 0.0, 'no_mask_channel_overlap': False, 'mask_channel_min_space': 1, 'conv_pos': 128, 'conv_pos_groups': 16, 'latent_temp': [2.0, 0.5, 0.999995], 'skip_masked': False, 'skip_nomask': False, 'checkpoint_activations': False, 'required_seq_len_multiple': 2, 'depthwise_conv_kernel_size': 31, 'attn_type': '', 'pos_enc_type': 'abs', 'fp16': True}
2023-07-12 20:20:56 | INFO | cress.tasks.speech_to_text_modified | pre-tokenizer: {'tokenizer': None}
2023-07-12 20:20:56 | INFO | cress.tasks.speech_to_text_modified | tokenizer: {'bpe': 'sentencepiece', 'sentencepiece_model': 'xxxx/data/mustc/en-de/spm_unigram10000.model'}
2023-07-12 20:20:56 | INFO | cress.datasets.speech_to_text_dataset | 'tst-COMMON' has 0.00% OOV
2023-07-12 20:20:56 | INFO | cress.datasets.speech_to_text_dataset | SpeechToTextDataset(split="tst-COMMON", n_samples=2_641, prepend_tgt_lang_tag=False, shuffle=False, transforms=None, n_frames_per_step=1
2023-07-12 20:20:59 | INFO | fairseq.tasks.fairseq_task | can_reuse_epoch_itr = True
2023-07-12 20:20:59 | INFO | fairseq.tasks.fairseq_task | reuse_dataloader = True
2023-07-12 20:20:59 | INFO | fairseq.tasks.fairseq_task | rebuild_batches = False
2023-07-12 20:20:59 | INFO | fairseq.tasks.fairseq_task | creating new batches for epoch 1
0%| | 0/154 [00:00<?, ?it/s]2023-07-12 20:20:59 | INFO | cress.tasks.speech_to_text_modified | pre-tokenizer: {'tokenizer': None}
2023-07-12 20:20:59 | INFO | cress.tasks.speech_to_text_modified | tokenizer: {'bpe': 'sentencepiece', 'sentencepiece_model': 'xxxx/data/mustc/en-de/spm_unigram10000.model'}
The program is stuck here, is there a problem with the writing of speech_to_text_modified.py?
Sorry for the late reply. Have you solved the problem? We haven't met this problem before.