Spaces:
Runtime error
Runtime error
| accum_grad: 4 | |
| adapter_type: gxl | |
| cmvn: null | |
| cmvn_conf: | |
| cmvn_file: null | |
| is_json_cmvn: null | |
| ctc_conf: | |
| ctc_blank_id: 50362 | |
| dataset: asr | |
| dataset_conf: | |
| batch_conf: | |
| batch_size: 26 | |
| batch_type: dynamic | |
| max_frames_in_batch: 2700 | |
| max_seq_in_batch: 2000 | |
| cycle: 100 | |
| emotion_en2zh_dict: conf/en2zh4emotion.json | |
| eod_id: 151643 | |
| feats_type: log_mel_spectrogram | |
| filter_conf: | |
| filter_no_extra_info: true | |
| max_length: 1000 | |
| max_seq_len: 950 | |
| min_length: 0 | |
| token_max_length: 360 | |
| token_min_length: 1 | |
| language_conf: | |
| limited_langs: | |
| - zh | |
| log_mel_spectrogram_conf: | |
| hop_length: 160 | |
| n_fft: 400 | |
| num_mel_bins: 80 | |
| padding: 0 | |
| resample_conf: | |
| resample_rate: 16000 | |
| shuffle: true | |
| shuffle_conf: | |
| shuffle_size: 1500 | |
| sort: true | |
| sort_conf: | |
| sort_size: 500 | |
| spec_aug: true | |
| spec_aug_conf: | |
| max_f: 10 | |
| max_t: 50 | |
| num_f_mask: 2 | |
| num_t_mask: 2 | |
| spec_sub: true | |
| spec_sub_conf: | |
| max_t: 30 | |
| num_t_sub: 3 | |
| spec_trim: false | |
| speed_perturb: false | |
| split_num: 10 | |
| decoder: transformer | |
| decoder_conf: | |
| activation_type: gelu | |
| attention_heads: 16 | |
| dropout_rate: 0.1 | |
| gradient_checkpointing: true | |
| input_layer: embed_learnable_pe | |
| key_bias: false | |
| linear_units: 4096 | |
| normalize_before: true | |
| num_blocks: 24 | |
| positional_dropout_rate: 0.1 | |
| self_attention_dropout_rate: 0.0 | |
| src_attention: true | |
| src_attention_dropout_rate: 0.0 | |
| tie_word_embedding: true | |
| use_output_layer: true | |
| downsample_rate: 4 | |
| dtype: bf16 | |
| encoder: transformer | |
| encoder_conf: | |
| activation_type: gelu | |
| attention_dropout_rate: 0.0 | |
| attention_heads: 16 | |
| dropout_rate: 0.1 | |
| gradient_checkpointing: true | |
| input_layer: conv1d2 | |
| key_bias: false | |
| linear_units: 4096 | |
| normalize_before: true | |
| num_blocks: 24 | |
| output_size: 1024 | |
| pos_enc_layer_type: abs_pos_whisper | |
| positional_dropout_rate: 0.1 | |
| static_chunk_size: -1 | |
| use_dynamic_chunk: false | |
| use_dynamic_left_chunk: false | |
| epoch: 11 | |
| fire_module: link_and_encoder_and_lora | |
| grad_clip: 5 | |
| init_step: false | |
| input_dim: 80 | |
| llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b | |
| log_interval: 10 | |
| lora_alpha: 32 | |
| lora_dropout: 0.1 | |
| lora_rank: 8 | |
| loss_dict: | |
| acc: 0.0 | |
| loss: 1.4107781417203814e-05 | |
| lrs: | |
| - 1.3727899783270902e-05 | |
| max_epoch: 100 | |
| model: llmasr | |
| model_conf: | |
| ctc_weight: 0 | |
| length_normalized_loss: false | |
| lsm_weight: 0.1 | |
| model_dir: /home/node54_tmpdata/xlgeng/ckpt/wenet_whisper_finetune_xlgeng/examples/wenetspeech/whisper/exp/qwen2_multi_task_4_6gpus_gxl_adapter/update_data/epoch_1_with_token | |
| optim: adamw | |
| optim_conf: | |
| betas: | |
| - 0.9 | |
| - 0.99 | |
| eps: 1.0e-06 | |
| lr: 5.0e-05 | |
| weight_decay: 0.01 | |
| output_dim: 151646 | |
| save_interval: 5000 | |
| save_states: model+optimizer | |
| save_time: 14/01/2025 20:35:10 | |
| scheduler: warmuplr | |
| scheduler_conf: | |
| warmup_steps: 8000 | |
| speech_token_num: 4097 | |
| step: 106124 | |
| tag: epoch_11 | |
| tokenizer: huggingface | |
| tokenizer_conf: | |
| llm_path: Qwen/Qwen2-7B #/home/node54_tmpdata/xlgeng/ckpt/qwen-7B-instruct/qwen2_7b | |
| train_engine: deepspeed | |
| use_amp: true | |
| use_lora: true | |
| vocab_size: 151646 | |