Hi everyone ! First of all, happy new year !
I got a new problem with this generate_lm.py let’s say that it’s the first of this year !
I tryed to follow the doc to validate my virtual environment but I got this :
(doc : https://deepspeech.readthedocs.io/en/latest/Scorer.html )
(venv) nathan@nathan-G771JM:~/PycharmProjects/DeepSpeech/DeepSpeech/data/lm$ python3 generate_lm.py --input_txt librispeech-lm-norm.txt.gz --output_dir . --top_k 500000 --kenlm_bins …/…/kenlm/build/bin/ --arpa_order 5 --max_arpa_memory “85%” --arpa_prune “0|0|1” --binary_a_bits 255 --binary_q_bits 8 --binary_type trie
Converting to lowercase and counting word occurrences …
| | # | 40418260 Elapsed Time: 0:15:38Saving top 500000 words …
Calculating word statistics …
Your text file has 803288729 words in total
It has 973673 unique words
Your top-500000 words are 99.9354 percent of all words
Your most common word “the” occurred 49059384 times
The least common word in your top-k is “corders” with 2 times
The first word with 3 occurrences is “zungwan” at place 420186Creating ARPA file …
=== 1/5 Counting and sorting n-grams ===
Reading /home/nathan/PycharmProjects/DeepSpeech/DeepSpeech/data/lm/lower.txt.gz
----5—10—15—20—25—30—35—40—45—50—55—60—65—70—75—80—85—90—95–100
******************************Traceback (most recent call last):
File “generate_lm.py”, line 210, in
main()
File “generate_lm.py”, line 201, in main
build_lm(args, data_lower, vocab_str)
File “generate_lm.py”, line 97, in build_lm
subprocess.check_call(subargs)
File “/usr/local/lib/python3.6/subprocess.py”, line 291, in check_call
raise CalledProcessError(retcode, cmd)
subprocess.CalledProcessError: Command ‘[’…/…/kenlm/build/bin/lmplz’, ‘–order’, ‘5’, ‘–temp_prefix’, ‘.’, ‘–memory’, ‘85%’, ‘–text’, ‘./lower.txt.gz’, ‘–arpa’, ‘./lm.arpa’, ‘–prune’, ‘0’, ‘0’, ‘1’]’ died with <Signals.SIGKILL: 9>.
I’m running this with :
Ubuntu 20.04.1 LTS (64 bits)
Intel® Core™ i5-4200H CPU @ 2.80GHz × 4
NV117 / Intel® HD Graphics 4600 (HSW GT2)
Python 3.6
Idk if it’s kenlm or generale_lm.py problem, if you could help !