I'm not so sure that truth and trustability is something we can just hand-wave away as something they'll sort out in just a few more years. I don't think a complex concept like whether or not something is actually true can be just tacked onto models whose core function is to generate what they think the next word of a body of text is most likely to be.