Using Linux and deepspeech-gpu, I trained a model with 0.9.3 with the following command:
python3 DeepSpeech.py \
--alphabet_config_path data/alphabet.txt \
--beam_width 32 \
--checkpoint_dir $ckpt_dir \
--export_dir $ckpt_dir \
--scorer $scorer_path \
--n_hidden 128 \
--learning_rate 0.0001 \
--lm_alpha 0.75 \
--lm_beta 1.85 \
--train_batch_size 6 \
--dev_batch_size 6 \
--test_batch_size 6 \
--report_count 10 \
--epochs 500 \
--noearly_stop \
--noshow_progressbar \
--export_tflite \
--train_files /datasets/deepspeech_wakeword_dataset/wakeword-train.csv,\
/datasets/deepspeech_wakeword_dataset/wakeword-train-other-accents.csv,\
/datasets/deepspeech_wakeword_dataset/wakeword-train.csv,\
/datasets/india_portal_2may2019-train.csv,\
/datasets/india_portal_2to9may2019-train.csv,\
/datasets/india_portal_9to19may2019-train.csv,\
/datasets/india_portal_19to24may2019-train.csv,\
/datasets/brazil_portal_20to26june2019-wakeword-train.csv,\
/datasets/brazil_portal_26juneto3july2019-wakeword-train.csv,\
/datasets/japan_portal_3july2019-wakeword-train.csv,\
/datasets/mixed_portal_backups_14_16_17_18_19_visteon_wakeword_dataset-train.csv,\
/datasets/alexa-train.csv,\
/datasets/alexa-polly-train.csv,\
/datasets/alexa-sns.csv,\
/datasets/india_portal_ww_data_04282020/custom_train.csv,\
/datasets/india_portal_ww_data_05042020/custom_train.csv,\
/datasets/india_portal_ww_data_05222020/custom_train.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_train.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_test.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_train.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_test.csv,\
/datasets/ww_gtts_data_google_siri/custom_train.csv,\
/datasets/ww_gtts_data_google_siri/custom_dev.csv,\
/datasets/ww_polly_data_google_siri/custom_train.csv,\
/datasets/ww_polly_data_google_siri/custom_test.csv \
--dev_files /datasets/deepspeech_wakeword_dataset/wakeword-dev.csv,\
/datasets/india_portal_2may2019-dev.csv,\
/datasets/india_portal_2to9may2019-dev.csv,\
/datasets/india_portal_9to19may2019-dev.csv,\
/datasets/india_portal_19to24may2019-dev.csv,\
/datasets/brazil_portal_20to26june2019-wakeword-dev.csv,\
/datasets/brazil_portal_26juneto3july2019-wakeword-dev.csv,\
/datasets/mixed_portal_backups_14_16_17_18_19_visteon_wakeword_dataset-dev.csv,\
/datasets/alexa-dev.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_dev.csv,\
/datasets/india_portal_ww_data_05222020/custom_dev.csv,\
/datasets/ww_gtts_data_google_siri/custom_dev.csv,\
/datasets/ww_polly_data_google_siri/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_dev.csv \
--test_files /datasets/alexa-sns.csv,\
/datasets/india_portal_ww_data_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_04282020/custom_test.csv,\
/datasets/india_portal_ww_data_05042020/custom_dev.csv,\
/datasets/india_portal_ww_data_05042020/custom_test.csv,\
/datasets/india_portal_ww_data_05222020/custom_dev.csv,\
/datasets/india_portal_ww_data_06182020/custom_dev.csv,\
/datasets/india_portal_ww_data_06182020/custom_test.csv
I also previously trained a model with 0.6.1 with the following command using the same datasets for train, dev and test and keeping all the hyper parameters same:
python3 DeepSpeech.py \
--alphabet_config_path data/alphabet.txt \
--beam_width 32 \
--checkpoint_dir $ckpt_dir \
--export_dir $ckpt_dir \
--lm_binary_path $lm_path/lm.binary \
--lm_trie_path $lm_path/trie \
--n_hidden 128 \
--learning_rate 0.0001 \
--lm_alpha 0.75 \
--lm_beta 1.85 \
--train_batch_size 6 \
--dev_batch_size 6 \
--test_batch_size 4 \
--report_count 10 \
--epochs 500 \
--noearly_stop \
--noshow_progressbar \
--export_tflite \
--dev_files /datasets/deepspeech_wakeword_dataset/wakeword-dev.csv,\
/datasets/india_portal_2may2019-dev.csv,\
/datasets/india_portal_2to9may2019-dev.csv,\
/datasets/india_portal_9to19may2019-dev.csv,\
/datasets/india_portal_19to24may2019-dev.csv,\
/datasets/brazil_portal_20to26june2019-wakeword-dev.csv,\
/datasets/brazil_portal_26juneto3july2019-wakeword-dev.csv,\
/datasets/mixed_portal_backups_14_16_17_18_19_visteon_wakeword_dataset-dev.csv,\
/datasets/alexa-dev.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_dev.csv,\
/datasets/india_portal_ww_data_05222020/custom_dev.csv,\
/datasets/ww_gtts_data_google_siri/custom_dev.csv,\
/datasets/ww_polly_data_google_siri/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_augmented_05042020/custom_dev.csv \
--test_files /datasets/alexa-sns.csv,\
/datasets/india_portal_ww_data_04282020/custom_dev.csv,\
/datasets/india_portal_ww_data_04282020/custom_test.csv,\
/datasets/india_portal_ww_data_05042020/custom_dev.csv,\
/datasets/india_portal_ww_data_05042020/custom_test.csv,\
/datasets/india_portal_ww_data_05222020/custom_dev.csv,\
/datasets/india_portal_ww_data_06182020/custom_dev.csv,\
/datasets/india_portal_ww_data_06182020/custom_test.csv
But, the average WER on all these datasets is 21.26% for 0.6.1 and 44.41% for 0.9.3. The text corpus used for LM and scorer was the same in both the cases