by petervandijck on 3/21/23, 1:14 PM with 4 comments
by wsgeorge on 3/21/23, 1:26 PM
> How are you?
The most likely next word might be computed to be "I".
> How are you? I
Then, what's the word that's most likely to follow "How are you? I"? That might be "am"
> How are you? I am
...and what's the word that's most likely to follow "How are you? I am"?
BTW, I think it's helpful to put "Ask HN" before question posts. Makes it easier to find.
by 360macky on 3/21/23, 1:18 PM
by sharemywin on 3/21/23, 1:32 PM
and because it's transformers onto of transformers it can says phrases based on words and phrases
the smaller ones like to repeat themselves I'm not sure if they Reinforcement learning is what fixed that or just a larger model.
by PaulHoule on 3/21/23, 1:28 PM
For instance somebody posted the other day an example where they asked GPT-4 who wrote a snippet from the Stratechery blog and it replied (after a lot of boilerplate about the difficulty of the problem) "Ben Thompson".
Somewhere in the training data it saw something like "Author: Ben Thompson" close to a lot of text that uses words in a certain way and and it learned the conditional probability distribution for that.
That underlying "knowledge" is captured by the pre-training phase where it learns the statistical regularities of text.
The ability to access that knowledge through prompts and to have a personality, add boilerplate text, be helpful, and agreeable even when it refuses to write a Hitler speech are trained in a second stage using this technique
https://huggingface.co/blog/rlhf
that is, by seeing examples where somebody asks to identify the author of text it learns not only to access the knowledge it learned in pre-training the write way but to also sound equivocal about it, even though that problem is a baseball pitched right at the center of the box so far as it is concerned.