hi dear friends:
Latest 0.7.3 version, chinese train. When I use below two commands to generate my own scorer file, there is one error message like below:
python3 generate_lm.py --input_txt vocabulary.txt --output_dir . --top_k 500000 --kenlm_bins /home/parallels/Desktop/ASR/kenlm/build/bin/ --arpa_order 5 --max_arpa_memory “85%” --arpa_prune “0|0|1” --binary_a_bits 255 --binary_q_bits 8 --binary_type trie
(deepspeech-0.7-train) parallels@parallels-Parallels-Virtual-Platform:~/Desktop/ASR/mozilla/DeepSpeech-0.7/data/lm$ python3 generate_package.py --alphabet alphabet.txt --lm lm.binary --vocab vocab-500000.txt --package chinese.scorer --default_alpha 0.931289039105002 --default_beta 1.1834137581510284
4173 unique words read from vocabulary file.
Looks like a character based model.
Using detected UTF-8 mode: True
Error when creating chinese.scorer
swig/python detected a memory leak of type ‘Alphabet *’, no destructor found.
But I see that chinese.scorer file is there, so I want to know what error it is? Is it harmless? Or what I should I do if I want to avoid it?
Thanks.