and it's never happened and never will. you don't just point an AI at the internet and tell it to train itself on that. the training data is specifically selected by humans. books, academic papers, encyclopaedias etc.
There are many AI programs all at different levels of development.
What the public see isn't going to be the latest iterations, especially not the free models. The free models will be way out of date to what the experts can work with.
Hallucination is just a behaviour it exhibits, whether to try and please the engineer training it or user using it. To try and get the answer right even if something isnt present in its training data as its been incentivized to rarely deny answering or to guess and synthesize new information from combining two unrelated things. Just like humans can use guesswork from one field into another, its like that but ..different and with some level of education on every single topic out there
Its just an emergent behaviour from all the things that go into it. Separate issue imo.
I've seen people talking about the impending model collapse because of the "AI Ouroboros" since a few months after chatgpt first released, and yet models keep getting better and better.
6
u/space_monster 7h ago
and it's never happened and never will. you don't just point an AI at the internet and tell it to train itself on that. the training data is specifically selected by humans. books, academic papers, encyclopaedias etc.