62
A prevailing sentiment online is that GPT-4 still does not understand what it talks about. We can argue semantics over what “understanding” truly means. I think it’s useful, at least today, to draw the line at whether GPT-4 has succesfully modeled parts of the world. Is it just picking words and connecting them with correct grammar? Or does the token selection actually reflect parts of the physical world?
One of the most remarkable things I’ve heard about GPT-4 comes from an episode of This American Life titled “Greetings, People of Earth”.
For the record, GPT4 specifically is non-deterministic. The current theory is because it uses MoE, but that’s just a theory. Maybe OpenAI knows why. Also, it’s not a random seed, it’s temperature. If you set that to 0, the model should always select the most probable next token because the probability becomes 1 for that token and 0 for all others. GPT3 and most others are basically deterministic at that level, but not GPT4.