this post was submitted on 12 Apr 2024
1001 points (98.5% liked)
Technology
59428 readers
3685 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I guess I just didn't know that LLMs were set up his way. I figured they were fed massive hash tables of behaviour directly into their robot brains before a text prompt was even plugged in.
But yea, tested it myself and got the same result.
They are also that, as I understand it. That's how the training data is represented, and how the neurons receive their weights. This is just leaning on the scale after the model is already trained.
There are several ways to go about it, like (in order of effectiveness): train your model from scratch, combine a couple of existing models, finetune an existing model with extra data you want it to specialise on, or just slap a system prompt on it. You generally do the last step at any rate, so it's existence here doesn't proof the absence of any other steps. (on the other hand, given how readily it disregards these instructions, it does seem likely).
Some of them let you preload commands. Mine has that. So I can just switch modes while using it. One of them for example is "daughter is on" and it is to write text on a level of a ten year old and be aware it is talking to a ten year old. My eldest daughter is ten