• Sailor Sega Saturn@awful.systems
    link
    fedilink
    English
    arrow-up
    14
    ·
    edit-2
    5 months ago

    Oh hey a blog: https://www.encultured.ai/blog.htmlBecause of course they’re an rationalist AI alignment / AI gaming startup pivoting to healthcare

    Part of their bold vision: AI agents that heal not just your cells, but also your society :')

    Our vision for 2027 and beyond remains similar, namely, the development of artificial general healthcare: Technological processes capable of repairing damage to a diverse range of complex systems, including human cells, organs, individuals, and perhaps even groups of people. Why so general? The multi-agent dynamical systems theory needed to heal internal conflicts such as auto-immune disorders may not be so different from those needed to heal external conflicts as well, including breakdowns in social and political systems.

    We don’t expect to be able to control such large-scale systems, but we think healthy is the best word to describe our desired relationship with them: As a contributing member of a well-functioning whole.

    Translation: they don’t know the first thing about healthcare but want the big US healthcare grifting dollars anyway.

    • Soyweiser@awful.systems
      link
      fedilink
      English
      arrow-up
      10
      ·
      edit-2
      5 months ago

      Surely they can tell us what they actually worked on in the gaming space in the past few years. Right? Right?

      Edit: (I recall hearing stories about roguelike developers using AI to help produce roguelikes, the conclusion was (this was more than a year back btw) it works great if you just need some quick content, dialogue trees, stuff like that. But there also was one person who was trying (and failing) to create a roguelike in which all content was LLM created. Didn’t seem to work well. (Which is funny in a way, as this mirrors the experience of PCG (procedural content generation) in roguelikes a ~decade earlier. Good tool for small parts, doesn’t scale to properly interesting content. I should ask how the LLM stuff is doing now).

      • Sailor Sega Saturn@awful.systems
        link
        fedilink
        English
        arrow-up
        7
        ·
        edit-2
        5 months ago

        Here’s what they write:

        AI alignment via the power of videogames:

        We’re starting with a singular focus on video game development, because we think that will offer the best feedback loop for testing new AI models. Over the next decade or so, we expect an increasing number of researchers — both inside and outside our company — will transition to developing safety and alignment solutions for AI technology, and through our platform and products, we’re aiming to provide them with a rich and interesting testbed for increasingly challenging experiments and benchmarks.

        Healthcare pivot:

        Originally, when Encultured was founded as a gaming-oriented AI research company, our immediate goal was to make research progress on human–AI interaction that would ultimately benefit humanity well beyond the entertainment sector. Since then, we’ve considered healthcare as a likely next step for us after gaming.

        Couldn’t find any details beyond that. Perhaps one of them read way too much Friendship is Optimal but they didn’t actually having any gaming chops so never got anywhere.

        EDIT: More details here: https://www.lesswrong.com/posts/ALkH4o53ofm862vxc/announcing-encultured-ai-building-a-video-game

        • Soyweiser@awful.systems
          link
          fedilink
          English
          arrow-up
          9
          ·
          5 months ago

          More details.

          Wow, they actually succeeded at their plans. I’m impressed. “we expect to be much more careful than other companies to ensure that recursively self-improving intelligent agents don’t form within our game and break out onto the internet!” Well done!

          • Sailor Sega Saturn@awful.systems
            link
            fedilink
            English
            arrow-up
            9
            ·
            5 months ago

            It’s real, all of it. John Titor the time traveler? He’s real. AI gods? We could build them.

            John Titor came back in time to stop the creation of a superintelligence. He does this by secretly founding, co-founding, or co-co-founding various silicon valley startups that don’t actually do anything; but that sound good to venture capitalists with too much money.

            The money is secretly funneled to good causes like food banks, adopting puppies, and maintaining the natural habitat of burrowing owls. Thus averting the end of the world. Encultured AI is part of this plan. They do nothing-- for the good of the earth.