I asked a Finnish person how good an answer about the language example from ChatGPT was. It turned out to be a hallucination, a confidently sounding nonsense.
The quality of internet trained models degrade very fast with language material size
Hallucinations and output quality are two different problems. Hallucinations are usually expressed in perfect sounding sentences by the LLM, that's what makes it so convincing to end users.
Maybe things have improved since that time (about a year ago). But my question was along the lines of "is it true that in Finnish this word means that?" and ChatGPT promptly confirmed with nonsense
The big win for accessibility has already been won...3 years ago.