If I have a finished recording at hand and wanna get the transcription in case of Javascript I simply do:
result = model.stt(buffer.slice(0, buffer.length / 2))
If I use the streaming Interface it looks like this:
let stream = model.createStream()
model.feedAudioContent(stream, buffer.slice(0, buffer.length / 2));
let result = model.intermediateDecode(stream);
The second & third line obviously happen as long as there come in chunks of audio-data from another streaming interface like a websocket-connection. And when the stream is done I get the final transcription with mode.finishStream(stream)
The docs for intermediateDecode note that this is still a very expensive task, but in this completed and merged issue its mentioned that the decoder is already capable of streaming, while the docs say it isn’t.
Whats right, and is it good practice to use intermediateDecode to show the user whats currently being recognized?