You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Just wondering if mixing language branches of small multi-lingual non-autoregressive models such as SenseVoice-Small reduces accuracy.
Being English my base language branch could be considered https://en.wikipedia.org/wiki/West_Germanic_languages and curious if stayed inside a singular language branch or at least only included similar say like North Germanic languages in this case than say what could be considered considerably different language branches that Mandarin, Cantonese, Japanese, and Korean may consider being part of.
Just curiosity but if English was omitted from SenseVoice-Small when trained would the accuracy for Mandarin, Cantonese, Japanese, and Korean increase or stay the same, or would it need to be purely Sino-Tibetan, wider or narrower in language branch?
Also when focussing on a language branch does this allow languages to be included that have much less resources?
Curiosity and as said was just wondering how a multi-lingual language branch model would perform...
The text was updated successfully, but these errors were encountered:
Just wondering if mixing language branches of small multi-lingual non-autoregressive models such as SenseVoice-Small reduces accuracy.
Being English my base language branch could be considered https://en.wikipedia.org/wiki/West_Germanic_languages and curious if stayed inside a singular language branch or at least only included similar say like North Germanic languages in this case than say what could be considered considerably different language branches that Mandarin, Cantonese, Japanese, and Korean may consider being part of.
Just curiosity but if English was omitted from SenseVoice-Small when trained would the accuracy for Mandarin, Cantonese, Japanese, and Korean increase or stay the same, or would it need to be purely Sino-Tibetan, wider or narrower in language branch?
Also when focussing on a language branch does this allow languages to be included that have much less resources?
Curiosity and as said was just wondering how a multi-lingual language branch model would perform...
The text was updated successfully, but these errors were encountered: