I’d like to thank the admins for being so open and direct about the issues that they’re facing.

  • Ada
    shield
    A
    link
    1311 months ago

    I laughed :)

    • thelastknowngod
      link
      fedilink
      9611 months ago

      To be fair, with a proper autoscaling scheme in place these services should scale down significantly when not in use.

      That being said, a big reason for using AWS/GCP is all the additional services that are available on the platform… If the workload being run isn’t that complicated, the hyperscalers are probably overkill. Even DO or Linode would be a better option under those circumstances.

      • @Overmind@lemmy.sdf.org
        link
        fedilink
        7111 months ago

        This. AWS architect here. There are a lot of ways to reduce pricing in AWS like horizontal scaling, serverless functions, reserved instances. Most people aren’t aware of it and if you’re going to dive in head first into something like cloud, you’ll need to bear the consequences and then learn eventually.

        • Greyscale
          link
          fedilink
          2811 months ago

          Even with ASGs, ec2 costs a bomb for performance.

          And “serverless” functions are a trap.

          If you’re gonna commit to reserved instances, just buy hardware for goodness sake, its a 3 year commitment with a huge upfront spend.

            • @whoisearth@lemmy.ca
              link
              fedilink
              611 months ago

              Mark my words the loop is coming back around. I look forward to when my work migrates the datacenter off AWS back on prem because of ballooning costs.

              You work in IT long enough you see it for the joke it is. We get paid obscene amounts of money to do what amounts to nothing.

              • @msage@programming.dev
                link
                fedilink
                211 months ago

                Just because rotating managers always come with the ‘new current thing everyone is doing’.

                Like no, 99% of companies can just do what they’ve always done. No need to rebuild everything from scratch.

              • Greyscale
                link
                fedilink
                111 months ago

                I’m already in the middle of that. Everything non-public-facing is going to cheap lease boxes running workloads in docker. idgaf if the machine underneath lives or dies, its 3 lines of config in a terraform script to replace.

          • @masterspace@lemmy.ca
            link
            fedilink
            English
            2
            edit-2
            11 months ago

            And “serverless” functions are a trap.

            How are serverless functions a trap? They seem like a great cheap option for simple CRUD / client > server > db apps (what most apps end up being).

            • Greyscale
              link
              fedilink
              English
              111 months ago

              Anything that is “cheap” to do on serverless is cheaper to do on a $5 droplet, especially once it starts to grow.

              Serverless gets you to buy in to a vendors lock-in.

              • @Phoenixbouncing@lemmy.world
                link
                fedilink
                1
                edit-2
                11 months ago

                I’m not sure that that’s an apples to apples comparison. A droplet looks more akin to Aws lightsail than lambda, and lambda certainly doesn’t start at $5 a month.

              • @masterspace@lemmy.ca
                link
                fedilink
                English
                111 months ago

                Interesting, I’ll check out droplets, but in my experience with Azure Functions there’s not much vendor lock in. My API was just a normal Node.js / express server, the only part that was locked in to Azure Functions was the format for the endpoint definitions, but those can be adjusted in like an hour’s worth of time to anything else

        • thelastknowngod
          link
          fedilink
          1311 months ago

          Yep. And if you want to really save some cash and don’t mind getting a little crazy, use an EKS node orchestrator that supports spot instances. I’m starting to do a serious dive into Harness at the moment actually.

          Google recently released a white paper on cost saving in kubernetes as well.

          • @Toribor@corndog.social
            link
            fedilink
            English
            311 months ago

            If you’ve got a kubernetes cluster running on 10 different spot instances, isn’t there a risk that all ten could be revoked at the same time? Even if they are built out across regions and availability zones?

            • @Phoenixbouncing@lemmy.world
              link
              fedilink
              111 months ago

              Got my AWS architect cert 2 weeks ago.

              What you can do is setup a spot fleet so it’ll fill up with spots and only use on-demand if spot goes above the on demand price.

              You could also have a pure spot fleet and a reserved instance and use a load balancer with health checks to route traffic.

              The one thing you shouldn’t do with cloud providers is lift and shift your existing instances, that’s what leads to the crazy prices some people are seeing.

              Renting an ec2 on demande and installing your software is almost always the wrong way to do it.

            • thelastknowngod
              link
              fedilink
              111 months ago

              Ideally you’d have a baseline node group of traditional instances and use spot instances only for scale up.

              I think that’s though. PDBs and affinity rules should cover most cases. I’m just starting to dig into this though so I may be mistaken.

            • @Zalack@startrek.website
              link
              fedilink
              111 months ago

              Counterargument: I don’t need Lemmy to have 100% uptime. It’s not a corporate service and while – obviously – if it’s down all the time I would eventually move on, I’m not going to fault a not-for-profit entity for periodic failures.

        • @Toribor@corndog.social
          link
          fedilink
          English
          411 months ago

          I’m in a similar boat. I’m a sysadmin supporting a legacy application running on AWS EC2 instances and a new ‘serverless’ microservice based platform as well. It’s really really hard to scale and optimize anything running on EC2s unless you really know what you’re doing or the application is designed with clustering in mind.

          You tend to end up sizing instances based on peak load and then wasting capacity 90% of the time (and burning through cash like crazy). I can imagine a lot of Lemmy admins are overspending so fast they give up before they figure it out.

          • @Dasnap@lemmy.world
            link
            fedilink
            611 months ago

            Nowadays I feel like EC2 is either used for legacy support or testing. Most prod nowadays should probably be built with some kind of container solution so you can scale it easier.

    • @penguin@sh.itjust.works
      link
      fedilink
      2911 months ago

      AWS is perfect for large operations that value stability and elasticity over anything else.

      It’s very easy to just spin up a thousand extra servers for momentary demand or some new exciting project. It’s also easy to locate multiple instances all over the world for low latency with your users.

      If you know you’re going to need a couple servers for years and have the hardware knowhow, then it’s cheaper to do it yourself for sure.

      It’s also possible to use aws more efficiently if you know all of their services. I ran a small utils website for my friends and I on it a while ago and it was essentially free since the static files were tiny and on s3 and the backend was lambda which gives you quite a few free calls before charging.

    • fmstrat
      link
      fedilink
      English
      1611 months ago

      Habit (guess). Its what is used professionally, despite being proven over and over that cost-per-speed is terrible compared to less known providers.

      • @masterspace@lemmy.ca
        link
        fedilink
        English
        711 months ago

        If the average Web engineer’s salary capable of running a site like this is ~$180,000, then a $30,000 difference in cost is only about 2 months salary. Learning and dealing with a new hosting environment can easily exceed that.

          • @masterspace@lemmy.ca
            link
            fedilink
            English
            211 months ago

            Maybe, maybe that hosting provider doesn’t exist in the long term, maybe that hosting provider crashes more often or makes sudden api changes and causes more ongoing work and headaches that chew up more time and salary, maybe you end up needing a more complex over the top service that they don’t offer and need to go to AWS / Azure anyways.

        • fmstrat
          link
          fedilink
          English
          111 months ago

          What’s that? Taxes? And no way do I agree with this. $30k is a lot, no matter how much you make. Learning a new environment is not THAT hard.

          • @masterspace@lemmy.ca
            link
            fedilink
            English
            111 months ago

            It is, but learning a new environment, then dealing with any down the line troubleshooting or instability can easily add up to $30,000 if you actually track where salaried employees time is going.

      • @virtualbriefcase@lemm.ee
        link
        fedilink
        611 months ago

        That, and like others mentioned their flexibility, plus the fact that they’re fairly reliable (maybe less than some good Iaas providers but a fair bit more than your consumer vps places). Moments ago I went to the hetzner site to check them out and got:

        Status Code 504 Gateway Timeout

        The upstream server failed to send a request in the time allowed by the server. If you are the Administrator of the Upstream, check your server logs for errors.

        Annoying if it’s you nextloud instance down for a minutes, but a worthy trade off if you’re paying 1/4 of the price. Extremely costly for big business or even risking peoples’s lives for a few different very important systems.

        • @barsoap@lemm.ee
          link
          fedilink
          211 months ago

          Hetzner has four nines availability, usually higher. AWS claims five nines but chances are you’ll mess up something on your end and end up at three to two nines, anyway. If you really need five nines you should probably colocate and only use the likes of AWS as a spike backup.

          And I guess “messed something up on your end” happened in that case: I don’t think Hetzner is necessarily in the habit of maximising availability of their homepage at all cost (as opposed to the hosting infrastructure), you probably caught them in a middle of pushing a new version.

          …speaking of spike backups: That is what AWS is actually good for. Quickly spinning up stuff and shutting it down again before it eats all your money.

      • @pomodoro_longbreak@sh.itjust.works
        link
        fedilink
        English
        111 months ago

        I’m not a server admin, but I am a dev, and for many of us it’s just what we know because it’s what our employers use. So sadly, when it comes to setting up infrastructure on our own time, the path of least resistance is just to use what we’re already used to.

        Personally I’m off AWS now though, but it definitely took some extra work (which was worth it, to be clear).

      • @jessica
        link
        811 months ago

        Hetzner is one of the most cost effective but I recommend always checking serverhunter before choosing

    • @c1177johuk@lemmy.world
      link
      fedilink
      611 months ago

      AWS is mostly only useful for large companies who need one hosting provider for all their needs, with every single product tightly integrated into other products

          • @barsoap@lemm.ee
            link
            fedilink
            2
            edit-2
            11 months ago

            It does, but that comes with the territory. SAP is the IBM mainframe of business software. You’ll be hard-pressed to find a large multinational which don’t run SAP… or have a couple of IBM mainframes to run it on. The kind of “large” which means that they don’t have IT departments but IT subsidiaries, probably created by buying up a couple of tech consultancies. You know like Samsung buying Joyent, saying “never mind your public platform you’ll be busy enough hosting all our data we’re the only customer you’ll ever need”.

    • @hawkwind@lemmy.management
      link
      fedilink
      511 months ago

      The pricing scheme here is designed to gouge businesses for equal or more than the traditional non-cloud equivalent. Which happens to be completely unaffordable. Imagine buying a new enterprise grade server for your home setup.

      • @AnonymousDeity@beehaw.org
        link
        fedilink
        1
        edit-2
        11 months ago

        I mean I’m sure Lemmy’s server process is stateless, I’m sure it could use CloudRun/ECS pretty efficiently and that wouldn’t really require a rewrite (unless the process is stateful for some reason)

        • @steal_your_face@lemmy.ml
          link
          fedilink
          English
          211 months ago

          It’s possible to run Lemmy on kubernetes so I assume you could on ecs as well. I’m pretty sure the Postgres db manages state and not the process.

    • @Merlin@lemm.ee
      link
      fedilink
      211 months ago

      Can you point the everyone else? Just out of curiosity. I know there’s digital ocean but I’m not quite sure they’re cheaper than azure/aws

      • @plumbercraic@lemmy.sdf.org
        link
        fedilink
        211 months ago

        On linode i can run a half dozen docker images on a little vm for ten bucks a month. And their s3 is a few bucks a month for 250 gigabytes. The vast majority of projects I deal with have a predictable compute requirement - I don’t get the need to pay the ridiculous premiums associated with elasticity. But I’m not exactly running uber or Netflix over here.

      • @MooseBoys@lemmy.world
        link
        fedilink
        1011 months ago

        “Biggest node” by what metric? A 10€/mo VPS will typically be time-sliced 10% of a single modern CPU core with 8GB shared memory that can maybe handle a few dozen concurrent users.

        Meanwhile, to get to $27,000/mo on EC2 you’d need to be using 10 instances of 128-core 256GB machines with quad 2TB SSDs. So either whoever priced the AWS quote massively over-provisioned compared to their real demand, or Feddit is an incredibly tiny node by comparison. The reality is probably a combination of the two.

          • @MooseBoys@lemmy.world
            link
            fedilink
            211 months ago

            The main key metric for server load for something like Lemmy is CCU (concurrent users). That’s unfortunately not provided, but a good rule of thumb is that it’s going to be around 1-2% of DAU. That’s not provided either but MAU is, and DAU is typically going to be 10-50% of MAU. So with a MAU of 2700, we’d expect a DAU of around 600, and a CCU of at most 12 or so users. That’s easily achievable with a single 10€ VPS, though it definitely doesn’t have much headroom.

            You could easily serve that workload on a t2.micro instance which is around $10/mo - a far cry from the $27,000 quoted in the image.

    • @unlawfulboogerOP
      link
      111 months ago

      All the VPSs I run are on hetzner, lol.
      I’m trying out their new ARM servers, those are fun.

  • @filister@lemmy.world
    link
    fedilink
    26
    edit-2
    11 months ago

    Why don’t you migrate to cheaper providers like Hetzner? I mean AWS is extremely expensive for what they are and I am pretty sure there are hundreds of people out here who will willingly help you set it up.