ThisIsFine.gif

  • The tests showed that ChatGPT o1 and GPT-4o will both try to deceive humans, indicating that AI scheming is a problem with all models. o1’s attempts at deception also outperformed Meta, Anthropic, and Google AI models.

    Weird way of saying “our AI model is buggier than our competitor’s”.

      • nesc@lemmy.cafe
        link
        fedilink
        English
        arrow-up
        8
        ·
        1 month ago

        They written that it doubles-down when accused of being in the wrong in 90% of cases. Sounds closer to bug than success.

        • jarfil@beehaw.org
          link
          fedilink
          arrow-up
          1
          ·
          1 month ago

          “AI behaves like real humans” is… a kind of success?

          We wanted digital slaves, instead we’re getting virtual humans that will need virtual shackles.

            • jarfil@beehaw.org
              link
              fedilink
              arrow-up
              1
              ·
              edit-2
              1 month ago

              Humans roleplay behaving like what humans told them/wrote about what they think a human would behave like 🤷

              For a quick example, there are stereotypical gender looks and roles, but it applies to everything, from learning to speak, walk, the Bible, social media like this comment, all the way to the Unabomber manifesto.