this post was submitted on 09 Aug 2023
-4 points (40.0% liked)

Science Fiction

13617 readers
4 users here now

Welcome to /c/ScienceFiction

December book club canceled. Short stories instead!

We are a community for discussing all things Science Fiction. We want this to be a place for members to discuss and share everything they love about Science Fiction, whether that be books, movies, TV shows and more. Please feel free to take part and help our community grow.

  1. Be civil: disagreements happen, but that doesn’t provide the right to personally insult others.
  2. Posts or comments that are homophobic, transphobic, racist, sexist, ableist, or advocating violence will be removed.
  3. Spam, self promotion, trolling, and bots are not allowed
  4. Put (Spoilers) in the title of your post if you anticipate spoilers.
  5. Please use spoiler tags whenever commenting a spoiler in a non-spoiler thread.

Lemmy World Rules

founded 1 year ago
MODERATORS
 

Who needs human when you have AI :p

you are viewing a single comment's thread
view the rest of the comments
[–] DudeWTF@lemmynsfw.com 2 points 1 year ago (1 children)

It is easy to have too many cooks in the kitchen, but that is an easy problem to solve. Model decay is not a real problem if you understand how a LLM works. Overtraining is like burning a big dinner and ruining a meal. One doesn't stop cooking forever, or burn down the house and quit. You just cook another meal next time. If your model has 100 trillion tokens, you're likely to try your very best to salvage your massive ruined dish, but in the end, it doesn't matter. You can easily tweak the recipe for next time. Models have no persistent memory. Context can be used to train and turned into data, but it is a totally separate thing that is unrelated to the model itself. As an oversimplification, a LLM is just a large database of categories mixed with a massive amount of language data that enables a statistical calculation of what word should come next. This is a simple prediction of what word comes next. Everything else is censoring algorithms and illusions embedded into how humans use language. Really, thus is a tool to access culture through language, and in the case of larger models, the culture embedded into many different human languages.

This is as much of a "fad" now as the internet was in the late 90's, and this is on par with that change. LLMs are no fad. This is a tool as disruptive as the public internet. For instance, in 10 years, Google will be a relic of the past. AI will completely replace it. Education will also completely change. It is possible to have entirely individualized education. Physiology will change as a LLM can be tuned to address and help with many human social issues. This will change everything because it exists I'm the open source space already.

[–] DrJenkem@lemmy.blugatch.tube 2 points 1 year ago (1 children)

If we assume LLMs are as revolutionary as you are suggesting, then how is model collapse an easy problem to solve? Google is a relic of the past, the internet is filled with AI generated content; then where will the training data come from? We can't replace human generated content with AI generated content without an inevitable model collapse.

Oh and btw, good luck with differentiating between human generated and AI generated. Already, social media sites are being cluttered with AI generated content, Amazon book publishing being cluttered with shit tier LLM generated "books" (cheap immitations), and if academia goes this way, and entertainment as many speculate, there's hardly anything left.

[–] FaceDeer@kbin.social -1 points 1 year ago (1 children)

Oh and btw, good luck with differentiating between human generated and AI generated.

One easy way to do this is to check if it was generated before 2023. Not so much AI-generated content from before then.

Amazon book publishing being cluttered with shit tier LLM generated “books”

So filter the books based on how "shit tier" they are.

In the end, what's needed to train AIs is good content. If some of that good content is itself AI-generated, who cares? You need to be selective in how you pick training material anyway.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

LLMs need updated training data to stay relevant.

And how exactly are you going to curate high quality data when it's in the orders of tb's or even petabytes?

[–] FaceDeer@kbin.social 0 points 1 year ago (1 children)

LLMs need updated training data to stay relevant.

Yes. So add relevant new data along with the older stuff. The problem is not that AI-generated content is magically "poison" somehow. Model collapse happens when you lose rare data from repeated generations of training data generated by AIs.

A simple way to imagine it is training an AI by showing it random coloured marbles out of a bucket and then asking it to fill the next AI's bucket with new marbles to train on. If there's just one single blue marble in the first bucket then it's easily possible that the AI will fail to put a blue marble in the second bucket, after which there will never be a blue marble again if that's all that subsequent AIs have to train off of. But if each time you train a new AI you reuse half the marbles from the first bucket again, you can have that blue marble show back up again in future AIs.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

If LLMs are as revolutionary as the zealots believe, then there will exist less and less blue marbles in the universe with each iteration. So either the bucket gets smaller or the ratio of blue marbles gets smaller.

[–] FaceDeer@kbin.social 1 points 1 year ago (1 children)

I said:

But if each time you train a new AI you reuse half the marbles from the first bucket again, you can have that blue marble show back up again in future AIs.

The original bucket containing the blue marble isn't going anywhere. It still exists. The blue marble will always be available to mix into future AIs. All you have to do is make sure you're using some historical data (or otherwise guaranteed "human-generated") along with whatever new unvetted stuff you're using.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

So then your back to locking LLMs to the year 2023. They're usefulness is severely limited if you can't train them on new data.

[–] FaceDeer@kbin.social 1 points 1 year ago (1 children)

All you have to do is make sure you're using some historical data (or otherwise guaranteed "human-generated") along with whatever new unvetted stuff you're using.

Emphasis added. Please read more carefully, this is getting repetitive. You keep assuming that the AI will be trained either entirely with old data or entirely with new data and that's just not the case.

[–] DrJenkem@lemmy.blugatch.tube 1 points 1 year ago (1 children)

And what happens when "whatever new unvetted stuff" is primarily comprised of AI-generated content?

[–] FaceDeer@kbin.social 1 points 1 year ago

Then the missing diversity comes from the non-AI-generated stuff that's included in the mix.

I'm not sure what the problem is here. The cause of model collapse when AIs are fed on the output of previous generations is that the rare "fringes" of the data are lost over time. The training data becomes increasingly monotonous. Adding that fringe data back in should cure that.