Hello,
I am working on training/fine-tuning Deepspeech 0.7.0 (CUDA 10.0, cudnn 7.6.5, tensorflow 1.15.2, ubuntu 16.04 with Python 3.6.7) on RTX 2080 Ti 11GB 2 GPU machine. I have 2 such machines available. I was interested in knowing if anyone here has attempted distributed training with horovod. If yes, can you please share your experiences and what changes you had to make.
thanks