• Kogasa@programming.dev
    link
    fedilink
    English
    arrow-up
    1
    ·
    1 day ago

    The GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model’s parameters. It’s not magic, it’s just a highly opaque encoding.