# This is an example that demonstrates how to configure a model file. # You can modify the configuration according to your own requirements. # to print the register_table: # from funasr.register import tables # tables.print() # network architecture model: WhisperWarp model_conf: lsm_weight: 0.1 length_normalized_loss: true hub: funasr # openai init_param_path: null # large-v2 or large-v3 if hub == "openai" # only use for hub == funasr, # if hub == openai, whisper_dims is automaticall download whisper_dims: 'n_mels': 80 'n_vocab': 51865 'n_audio_ctx': 1500 'n_audio_state': 1280 'n_audio_head': 20 'n_audio_layer': 32 'n_text_ctx': 448 'n_text_state': 1280 'n_text_head': 20 'n_text_layer': 32 # frontend related frontend: WhisperFrontend frontend_conf: fs: 16000 n_mels: 80 do_pad_trim: true tokenizer: WhisperTokenizer tokenizer_conf: language: null task: transcribe is_multilingual: true num_languages: 99 scope_map: ['none', "model."]