enh_inference_streaming.py
Less than 1 minute
enh_inference_streaming.py
Frontend inference
usage: enh_inference_streaming.py [-h] [--config CONFIG]
[--log_level {CRITICAL,ERROR,WARNING,INFO,DEBUG,NOTSET}]
--output_dir OUTPUT_DIR [--ngpu NGPU]
[--seed SEED]
[--dtype {float16,float32,float64}]
[--fs FS] [--num_workers NUM_WORKERS]
--data_path_and_name_and_type
DATA_PATH_AND_NAME_AND_TYPE
[--key_file KEY_FILE]
[--allow_variable_data_keys ALLOW_VARIABLE_DATA_KEYS]
[--output_format OUTPUT_FORMAT]
[--train_config TRAIN_CONFIG]
[--model_file MODEL_FILE]
[--model_tag MODEL_TAG]
[--inference_config INFERENCE_CONFIG]
[--enh_s2t_task ENH_S2T_TASK]
[--batch_size BATCH_SIZE]
[--ref_channel REF_CHANNEL]