• schizo@forum.uncomfortable.business
    link
    fedilink
    English
    arrow-up
    14
    ·
    1 month ago

    I suspect that it’s going to go the same route as the ‘acting on behalf of a company’ bit.

    If I call Walmart, and the guy on the phone tells me that to deal with my COVID infection I want to drink half a gallon of bleach, and I then drink half a gallon of bleach, they’re going to absolutely be found liable.

    If I chat with a bot on Walmart, and it tells me the same thing, I’d find it shockingly hard to believe that the decisions from a jury would in any way be different.

    It’s probably even more complicated in that while a human has free will (such as it is), the bot is only going craft it’s response from the data it’s trained on, so if it goes off the rails and starts spouting dangerous nonsense, it’s probably an even EASIER case, because that means someone trained the bot that drinking bleach is a cure for COVID.

    I’m pretty sure our legal frameworks will survive stupid AI, because it’s already designed to deal with stupid humans.

    • Letstakealook@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      1 month ago

      Would a court find Walmart liable for your decision to take medical advice from a random employee? I’m sure Walmart could demonstrate that the employee was not acting in the capacity of their role and any reasonable person would not consider drinking bleach because an unqualified walmart employee told them so.

      • schizo@forum.uncomfortable.business
        link
        fedilink
        English
        arrow-up
        6
        ·
        1 month ago

        I changed company names before posting and broke the clarity, sorry.

        Imagine I wasn’t a idiot and had said Walmart pharmacy, which is somewhere you’d expect that kind of advice.

        • Letstakealook@lemm.ee
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 month ago

          That would make it more plausible. I don’t think you’re an idiot, I was asking because I was curious if there was precedent for a jackass conspiracy minded employee handing out medical advice causing liability for a business. I wouldn’t think it is right, but I also don’t agree with other legal standards, lol.

          • schizo@forum.uncomfortable.business
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 month ago

            Thankfully there’s not: you’d expect someone at a pharmacy to provide reasonable medical advice, or your mechanic to tell you the right thing to do with your car. Once you walk outside the field where a reasonable person would reasonably expect what they’re being told to be uh, reasonable, then there’s usually no real case for liabilities.

            Buuuuuut, in the US at least, this is entirely civil law, and that means the law is mostly whatever you can convince a jury of, so you can end up with some wacky shit happening.