Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Extending the LLM article: "general"-ness vs. so-called "hallucinations" #1

Open
6r1d opened this issue Jan 14, 2024 · 1 comment
Open

Comments

@6r1d
Copy link
Contributor

6r1d commented Jan 14, 2024

New section: "Modern LLMs lack the ability to self-improve"

Text

While we're often told the current models are powerful, and they certainly have their advantages, I find their inability to self-improve very curious.

An article by the Epoch research group claims that our civilisation may run out of training data very soon.

The language, not the program code on some particular programming language, is the most abundant thing in the training data fed to an LLM. If an LLM behaves in a manner comparable to a human in language processing, it should be most eloquent in its understanding of the natural language. If that is the case, it wouldn't be a stretch to imagine it filling its next training dataset using its language capabilities, be it paraphrasing or summarising. Some multimodal networks can recognise visual cues and infer some conclusions. Some, like Stable Diffusion or Dall-E, can create beautiful images. There is a lot of data that is already learned. Why, then, is such data not being used directly? What stops LLMs from learning further and further as infants do? Some networks can recognise voice, so why not use radios working 24/7, TV shows and anything else that seems valuable?

If LLMs became so powerful they are approaching "general artificial intelligence", why would training data ever be a limitation? I hypothesised that there was a significant issue with the quality of the available data. The term "hallucination" is thrown all around, and we probably deal with language models that need people to distinguish between a "good" and a "bad" output, as their models have limitations. And even if some composition of ML models could, in theory, be used to fill each other's gaps, there seems to be a limit to the practicality of such an approach.

References to add

Running out of the training data

"Will we run out of data? An analysis of the limits of scaling datasets in Machine Learning", Pablo Villalobos, Jaime Sevilla, Lennart Heim, Tamay Besiroglu, Marius Hobbhahn, Anson Ho. Business Insider refers to them as the [Epoch] research group, so I should check this before I write it.

Media references
@6r1d
Copy link
Contributor Author

6r1d commented Jan 15, 2024

Yet another reference is the admission of the unpredictability on the ChatGPT Twitter page:

we've heard all your feedback about GPT4 getting lazier! we haven't updated the model since Nov 11th, and this certainly isn't intentional.
model behavior can be unpredictable, and we're looking into fixing it 🫡

While the context is different, it is very much related to the hallucinations preventing a stable self-improvement.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant