-
Notifications
You must be signed in to change notification settings - Fork 130
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adding multilingual support via fine-tuning code #23
Comments
Yes, I think releasing some fine-tuning code would be beneficial. I can add it to our TODO list. |
@evmaki Please provide the code for finetuning, would be beneficial for extending to other languages. |
Korean and Mandarin will boost their popularity in ASR use cases . it's great work guys |
+1 |
Results seem promising. Hope to have the multilingual support or some fine-tune code soon, thank you for your work :) |
Any updates on this? |
The reason why this (multilingual support) is extremely important can been see from this simple chart related to whisper fine-tune for downstream tasks of transcription (WER) and translation (BLEU) and the correlation to the amount (in hours) of audio transcribed or translated respectively. By example while Spanish (ES) exhibits the Best WER (2.5) for speech recognition it has the Lowest BLEU score (24) for translation, while German (DE) has balanced performance (WER: 4, BLEU: 35), or - viceversa Portuguese (PT) shows the High BLEU score (39) for translation but a relatively low WER (4) for speech recognition, etc. |
While both tiny and base model achieve impressive performances 💯 in terms of WER on academics datasets when compared to the "standard-de-facto" (e.g. Whisper), in a real world scenery a multi-lingual model would address more use cases.
Fine-tune train code could eventually enable support by the community.
The text was updated successfully, but these errors were encountered: