Apparently there are several narratives in regards to AI girlfriends.

  1. Incels use AI girlfriends given that they can do whatever they desire.
  2. Forums observing incel spaces agree that incels should use AI girlfriends to leave real women alone
  3. The general public having concerns towards AI girlfriends because their users might be negatively impacted by their usage
  4. Incels perceiving this as a revenge fantasy because “women are jealous that they’re dating AI instead of them”
  5. Forums observing incel spaces unsure if the views against AI girlfriends exist in the first place due to their previous agreement

I think this is an example of miscommunication and how different groups of people have different opinions depending on what they’ve seen online. Perhaps the incel-observing forums know that many of the incels have passed the point of no return, so AI girlfriends would help them, while the general public perceive the dangers of AI girlfriends based on their impact towards a broader demographic, hence the broad disapproval of AI girlfriends.

  • pavnilschanda@lemmy.worldOPM
    link
    fedilink
    arrow-up
    2
    ·
    edit-2
    7 months ago

    Thanks for your input. I agree with your overall comment. Within general narratives, incels aren’t usually included. As for the broad disapproval, it’s something that I tend to notice in the AI space.

    AI chatbot personas are generally seen as a hobby, a one-and-done thing compared to an “entity” that accompanies you for long periods of time; the latter part has more stigma attached to it. And given the AI boom only a few years ago, many people, including academic researchers, have only started to be aware of its existence and have made many uninformed assumptions about them. Not to mention the ethical minefelds that are yet to be explored, and increasingly so within the humanities such as psychology and anthropology, hence the Google Deepmind article that you shared. Given the sheer complexity surrounding AI companionship combined with the attention-based economy that has shaped our society, it makes sense that non-specialized places would adopt a binary approach when it comes to AI, artificial girlfriends included.

    There seems to be strong connections between inceldom and AI companionship, given that AI girlfriends are marketed for lonely men, and many of them just happen to be incels. But as you’ve said, AI companion users are very diverse, it’s just that the topic of incels or an incel-related topic would get brought up every now and then within the AI companionship discourse.

    • rufus@discuss.tchncs.de
      link
      fedilink
      arrow-up
      2
      ·
      edit-2
      7 months ago

      Hmmh. I’m pretty sure OpenAI and Google are very aware of this. I mean erotic roleplay is probably out of the question since they’re American companies. And the whole field of AI is a minefield to them starting with copyright to stuff like this. And they did their homework and made the chatbots not to present themselves as emotive. I percieve this as concensus in society, that we need to be cautious about the effects on human psyche. I wonder if that’s going to shift at some point. I’m pretty sure more research is going to be done and AI will become more and more prevalent anyways, so we’re going to see whether people like it or not.

      And as I heard lonelyness is on the rise. If not in western cultures, I think Japan and Korea are way ahead of us. And the South Koreans seem also to have a problem with a certain kind of incel culture, which seems to be way worse and more widespread amongst young men, there. I’ve always wanted to read more about that.

      I - myself - like AI companions. I think it’s fantasy. Like reading a book, playing video games or watching movies. We also explore the dark sides of humans there. We write and read murder mystery stories, detailing heinous acts. We kill people in video games. We process abuse and bad things in movies. And that’s part of being human. Doing that with chatbots is the next level, probably more addictive and without some of the limitations of other formats. But I don’t think it’s bad per se.

      I don’t know what to say to people who like to be cruel, simulate that in a fantasy like this. I think if they’re smart enough to handle it, I’m liberal enough not to look down on them for that. If being cruel is all there is to someone, they’re a poor thing in my eyes. Same for indulging in self-hatred and pity. I can see how someone would end up in a situation like that. But there’s so much more to life. And acting it out on (the broad concept of) women isn’t right or healthy. And it’s beyond my perspective. From my perspective there isn’t that big a difference between genders. I can talk to any of them and ultimately their interests and needs and wants are pretty much the same.

      So if an incel were to use a chatbot, i think it’s just a symptom for the underlying real issue. Yes it can reinforce them. But some people using tools for their twisted purposes, doesn’t invalidate other use cases. And it’d be a shame if that narrative were to dominate public perspective.

      I often disagree with people like Mark Zuckerberg, but I’m grateful he provides me with large language models that aren’t “aligned” to their ethics. I think combatting loneliness is a valid use case. Even erotic roleplay and exploring concepts like violence in fantasy scenarios ultimately is a valid thing to do in my eyes.

      There is a good summary on Uncensored Models by Eric Hartford which I completely agree with. I hope they don’t ever take that away from us.