Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Do we need language specific LLMs? I can’t vouch for the data coverage or accuracy of Arabic in the leading models today, but I do know them to be highly cross-lingual capable.


Back when gpt 3.5 was released I was testing it for translating Tongan. A language that wasn't even on Google. It was doing okay. It lacks certain formalities and sort of contextual understandings in other languages like Tagalog or Spanish. But I noticed that if I put it into a character that is a native Filipino or Honduran. Or Tongan. It does do better. But gpt 5 really is leaps of evolution ahead compared to 3.5. It actually does really well now.


Makes me wonder who did the translation

"This is a translation of the Arabic article published on 3rd August 2025"

Full irony would be from an LLM


Such declarations have become pretty useless without any indicator of the translation method.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: