I really need every organization ever to understand that large language models cannot ever fact-check without a major qualitative change in how they work. This isn't like how a toddler can't read, where it's a matter of putting in more time and effort. This is like how a rock can't read.
It's in the name! Large language model! They are getting better at forming grammatical sentences, and cannot get better at understanding their meaning! There is simply no mechanism in there, as best I understand, for determining that one grammatically valid sentence is likely to be true and another is likely to be false. Not an underdeveloped beta-stage mechanism. Nothing. This car isn't slow, it has no wheels.
So it's really weird to see all sorts of organizations rolling out "AI" features like we don't all know this. I guess it's the hot thing and it's cheap, but it's such a frustrating Emperor's New Clothes situation and all I can hope for is that some high profile companies get sued when their LLM writes checks their customer service can't cash.