this post was submitted on 23 Jul 2023
41 points (95.6% liked)

Programming

17424 readers
27 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] polyfire@waveform.social 3 points 1 year ago (2 children)

I wonder what an ELI5 version of 'stored weights' would be in this context.

[–] Taival@suppo.fi 2 points 1 year ago* (last edited 1 year ago)

Not quite ELI5 but I'll try "basic understanding of calculus" level.

In very broad terms, the model learns complex relationships between words (or tokens to be specific, explained below) as probabilistic scores. At its simplest, this could mean the likelihood of one word appearing next to another in the massive amounts of text the model was trained with: the words "apple" and "pie" are often found together, so they might have a high-ish score of 0.7, while the words "apple" and "chair" might have a lower score of just 0.2. Recent GPT models consist of several billions of these scores, known as the weights. Once their values have been estabilished by feeding lots of text through the model's training process, they are all that's needed to generate more text.

Without getting into the math too much, this is how a GPT model then uses these numbers to come up with words:

  • The input prompt is first chopped up into tokens that are each assigned a number. For example, the OpenAI tokenizer translates "Hello world!" into the numbers [15496, 995, 0]. You can think of this as the A=1, B=2, C=3... cipher we all learnt as kids, but the numbers are also assigned to common words, syllables and punctuation.
  • These numbers are inserted into a massive system of equations where they are multiplied together with the billions of weights of the model in a specific manner. This calculation results in a probability score from 0 to 1 for each token known by the model, representing how likely that token is to appear next in sequences that look similar to your input.
  • One of the tokens with the highest scores is chosen as the model's output semi-randomly to provide variance.
  • This cycle is then repeated over and over, generating the text one token at a time.

In reality we're not quite so sure what the weights represent to the model exactly, but this is the gist of it. All we know is that they signify the importances or non-importances that the model places on some pattern that was present in the training data. Some of these patterns could be just simple two-word pairs, but many are probably much more complicated. Lots of researchers are currently trying to get a better idea of how these numbers are actually affecting the model's output.

[–] Lmaydev@programming.dev 1 points 1 year ago* (last edited 1 year ago)

How closely related words and their attributes are to other words.