cm0002@lemmy.world to Artificial Intelligence@lemmy.worldEnglish · 1 day agoAnthropic CEO Admits We Have No Idea How AI Worksfuturism.comexternal-linkmessage-square6linkfedilinkarrow-up118
arrow-up118external-linkAnthropic CEO Admits We Have No Idea How AI Worksfuturism.comcm0002@lemmy.world to Artificial Intelligence@lemmy.worldEnglish · 1 day agomessage-square6linkfedilink
minus-squareVagueAnodyneCommentslinkfedilinkEnglisharrow-up15·1 day agoThis bullshit propaganda is intensely frustrating. Stop pretending these models are magic. This propaganda is meant to mystify and fascinate idiots. Here is a short reading list that will explain to a competent person how generative transformers work. https://jaykmody.com/blog/gpt-from-scratch/#fn1 https://explainextended.com/2023/12/31/happy-new-year-15/ https://jalammar.github.io/how-gpt3-works-visualizations-animations/ https://arxiv.org/abs/1706.03762 Enough pretend-along. They aren’t smarter than you.
minus-squareKogasa@programming.devlinkfedilinkEnglisharrow-up1·18 hours agoThe GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model’s parameters. It’s not magic, it’s just a highly opaque encoding.
minus-squarewondrous_strange@lemmy.worldlinkfedilinkEnglisharrow-up1·20 hours agoGood on you for saying it so eloquently. None of them would have dared to say anything remotely similar to ‘we don’t know’ if it really was the case
This bullshit propaganda is intensely frustrating. Stop pretending these models are magic. This propaganda is meant to mystify and fascinate idiots.
Here is a short reading list that will explain to a competent person how generative transformers work.
https://jaykmody.com/blog/gpt-from-scratch/#fn1
https://explainextended.com/2023/12/31/happy-new-year-15/
https://jalammar.github.io/how-gpt3-works-visualizations-animations/
https://arxiv.org/abs/1706.03762
Enough pretend-along. They aren’t smarter than you.
The GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model’s parameters. It’s not magic, it’s just a highly opaque encoding.
Good on you for saying it so eloquently. None of them would have dared to say anything remotely similar to ‘we don’t know’ if it really was the case