Wikipedia has a new initiative called WikiProject AI Cleanup. It is a task force of volunteers currently combing through Wikipedia articles, editing or removing false information that appears to have been posted by people using generative AI.

Ilyas Lebleu, a founding member of the cleanup crew, told 404 Media that the crisis began when Wikipedia editors and users began seeing passages that were unmistakably written by a chatbot of some kind.

  • randon31415@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    ·
    19 minutes ago

    If anyone can survive the AI text apocalypse, it is wikipedia. They have been fending off and regulating article writing bots since someone coded up a US town article writer from the 2000 census (not the 2010 or 2020 census, the 2000 census. This bot was writing wikipedia articles in 2003)

  • kibiz0r@midwest.social
    link
    fedilink
    English
    arrow-up
    20
    ·
    3 hours ago

    Unleashing generative AI on the world was basically the information equivalent of jumping headfirst into Kessler Syndrome.

  • narc0tic_bird@lemm.ee
    link
    fedilink
    English
    arrow-up
    39
    ·
    6 hours ago

    Best case is that the model used to generate this content was originally trained by data from Wikipedia so it “just” generates a worse, hallucinated “variant” of the original information. Goes to show how stupid this idea is.

    Imagine this in a loop: AI trained by Wikipedia that then alters content on Wikipedia, which in turn gets picked up by the next model trained. It would just get worse and worse, similar to how converting the same video over and over again yields continuously worse results.

    • huginn@feddit.it
      link
      fedilink
      English
      arrow-up
      14
      ·
      4 hours ago

      See also: model collapse

      (Which is more or less just regression towards the mean with more steps)

    • 8uurg@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      2 hours ago

      A very similar situation to that analysed in this paper that was recently published. The quality of what is generated degrades significantly.

      Although they mostly investigate replacing the data with ai generated data in each step, so I doubt the effect will be as pronounced in practice. Human writing will still be included and even curation of ai generated text by people can skew the distribution of the training data (as the process by these editors would inevitably do, as reasonable text could get through the cracks.)

  • schizo@forum.uncomfortable.business
    link
    fedilink
    English
    arrow-up
    168
    ·
    8 hours ago

    Further proof that humanity neither deserves nor is capable of having nice things.

    Who would set up an AI bot to shit all over the one remaining useful thing on the Internet, and why?

    I’m sure the answer is either ‘for the lulz’ or ‘late-stage capitalism’, but still: historically humans aren’t usually burning down libraries on purpose.

    • rsuri@lemmy.world
      link
      fedilink
      English
      arrow-up
      1
      ·
      5 minutes ago

      Yeah but the other thing about humanity is it’s mostly harmless. Edits can be reverted, articles can be locked. Wikipedia will be fine.

    • Schmoo@slrpnk.net
      link
      fedilink
      English
      arrow-up
      64
      ·
      7 hours ago

      historically humans aren’t usually burning down libraries on purpose.

      How on earth have you come to this conclusion.

    • poszod@lemmy.world
      link
      fedilink
      English
      arrow-up
      79
      ·
      8 hours ago

      State actors could be interested in doing that. Same with the internet archive attacks.

    • Petter1@lemm.ee
      link
      fedilink
      English
      arrow-up
      2
      ·
      edit-2
      4 hours ago

      Maybe a strange way of activism that is trying to poison new AI models 🤔

      Which would not work, since all tech giants have already archived preAI internet

      • schizo@forum.uncomfortable.business
        link
        fedilink
        English
        arrow-up
        3
        ·
        4 hours ago

        Ah, so the AI version of the chewbacca defense.

        I have to wonder if intentionally shitting on LLMs with plausible nonsense is effective.

        Like, you watch for certain user agents and change what data you actually send the bot vs what a real human might see.

        • Dragonstaff@leminal.space
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 hours ago

          I suspect it would be difficult to generate enough data to intentionally change a dataset. There are certainly little holes, like the glue pizza thing, but finding and exploiting them would be difficult and noticing you and blocking you as a data source would be easy.

    • endofline@lemmy.ca
      link
      fedilink
      English
      arrow-up
      6
      ·
      7 hours ago

      It’s not about on purpose but usually most people don’t care about what’s not in their interest. Today interests are usually quite shallow what tiktok shows quite well. Libraries do require money for operating. Even internet archive and wikipedia

  • sbv@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    81
    ·
    8 hours ago

    As for why this is happening, the cleanup crew thinks there are three primary reasons.

    “[The] main reasons that motivate editors to add AI-generated content: self-promotion, deliberate hoaxing, and being misinformed into thinking that the generated content is accurate and constructive,”

    That last one. Ouch.

    • TimLovesTech (AuDHD)(he/him)@badatbeing.social
      link
      fedilink
      English
      arrow-up
      21
      ·
      6 hours ago

      “[The] main reasons that motivate editors to add AI-generated content: self-promotion, deliberate hoaxing, and being misinformed into thinking that the generated content is accurate and constructive,

      I think the main driver behind people misinformed about AI content comes from the fact that outside of tech people, most have no idea that AI will:

      1. 100% make up answers to things it doesn’t know because either the sample size of data they have ingested was to small or was bad. And it will do this with the same robot confidence you get for any other answer.

      2. AI that has been fed to much other AI generated content will begin to “hallucinate” and give some wild outputs, very similar to humans suffering from schizophrenia. And again these answers will be given as “fact” with the same robotic confidence.

    • BigDanishGuy@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      4
      ·
      7 hours ago

      Well, I was in doubt, so I asked the AI whether I could trust the answers and it told me not to worry about it. That must mean that I only get accurate answers, right? /s

  • lolola
    link
    fedilink
    English
    arrow-up
    32
    ·
    7 hours ago

    I hate to post because I have loved and trusted Wikipedia for years, but the fact that there are folks out there who equally trust what AI tools generate just baffles me.

    • Dragonstaff@leminal.space
      link
      fedilink
      English
      arrow-up
      2
      ·
      2 hours ago

      The signal to noise ratio is so low these days. There’s so much information out there but everyone wants to profit from you before you can get it. Even worse, the people with good information generally can’t buy as big a megaphone as the people who profit from lying to you.

      Honestly, I think humans have been more likely to believe an easy lie than a hard truth all along, but it’s easier than ever these days.

  • Aatube@kbin.melroy.org
    link
    fedilink
    arrow-up
    18
    ·
    7 hours ago

    Don’t worry, it’s not as bad as the title suggests. The attack on Internet Archive is far, far worse. It’s obviously a bit of a problem, though.