Is Whisper any good for multi-lingual short command sentences or could it be better? #1430
StuartIanNaylor
started this conversation in
General
Replies: 2 comments 1 reply
-
#1229 might be helpful to some extend when the domain is super narrow and specific. As for training / finetuning - maybe in the future, but at least few months down the road. |
Beta Was this translation helpful? Give feedback.
1 reply
-
https://github.com/jmorganca/ollama |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
A question as much as a statement as if you use certain languages, with short context on smaller models the WER rockets.
I have a preference for smaller models and race-till-idle of a central ASR not the consumer device kind and being central the model.
When it comes to Whisper my preference is elsewhere are similar accuracy when doing domain n-gram based LM.
WeNet did a good write up and haven't actually used yet, but its a really good comparison as training and plugging the language / domain holes is much easier and you end up with much faster models that could even have better domain WER.
n-gram based LM and https://wenet.org.cn/wenet/context.html
I am spammed the above deliberately , to ask are there any similar methods with Whisper? Or is a matter of finetuning?
@ggerganov I thought I would ask, do you have any code plans for training small domain specific models.
I have been thinking a cluster of domain specific models, than bigger all-ones, but either, but wondering if training LLMs and Langchain/Agent type mechanisms could have similiar optimisations?
I have hunch this should be possible to train a domain specific small LLM on the output of the ASR to be used as a hybrid pair.
Also it output would likely be a Domain REST api of some form of json.
I thought I would ask as much dev seems larger all-in-one LLMs or very large domains such as coding and not seen. nor know if small domain efficient training is possible or what would be a goodbase model for finetuning with context injection to train to an api, where as above similar LM and context biasing can be implemented?
Beta Was this translation helpful? Give feedback.
All reactions