Per quick testing the 27b model seems very strong at least in natural language. It produces even good Finnish, in which smaller models tend to really struggle. Very promising.
Edit: Per even quicker testing the Finnish language performance degrades rapidly with the smaller models, as is usually the case. Would be great to have language specific distillations from larger models.
I’ve hoped for this too, but as a Swede. There’s been GPT-SW3 but it was poor. We could technically have very powerful, small language specific models. I think its unfortunately just a funding and resource issue.
DeepL is a lot better in spelling and grammar, but I didn't mean for translation but to interact directly in Finnish. Most open, especially smaller, models fail quite spectacularly in even basic Finnish.
Edit: Per even quicker testing the Finnish language performance degrades rapidly with the smaller models, as is usually the case. Would be great to have language specific distillations from larger models.