I’d like to contribute to the Lemmy community. I’ve been running my own private Linux servers for more than 25 years for things like email (years ago before all the spam), and as file servers, backup, etc. It’s an old, not very powerful computer, running Ubuntu server, in a corner in my house. Is it worth running a Lemmy instance on such a machine? I suppose there’d also be issues of how much data is going in and out, and how that would impact my internet cable usage. Thoughts?

  • Slashzero@hakbox.social
    link
    fedilink
    arrow-up
    3
    ·
    edit-2
    1 year ago

    I’m running my instance on 2 cores, 2GB of RAM. Of course I’m the only one on there at the moment, but it’s running great, and I think it might even be fine with a single core.

    As other have said, if you are planning to use it for your own “user’s home instance,” that should be fine. I’ve read a few people are running their instances on Raspberry PIs, which is pretty neat. While I have one I could use, I opted to setup a new droplet in DigitalOcean instead (I also run my own servers like you). A 2 core / 2GB RAM / 50GB SSD disk droplet on DigitalOcean is about $18 (USD) a month, while a single core droplet is about $12 (USD) per month.

    If you plan to run an instance for others to use, be aware the federation is going to be chatty on your home network, and could impact other devices on your network. Probably not ideal, which is why I opted for a droplet in DigitalOcean instead.

    • mike@reddthat.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      1 year ago

      Any recommended guides? I consider myself pretty savvy with tech as a software engineer, but I’d really like some sort of docker image to just spin up on my unraid server. I’m pretty lazy playing the whole sys admin role…

    • CaptManiac@lemmy.sdf.orgOP
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      It did cross my mind if I could have one of my raspberry pis run it. Actually, if it is possible I’d do on an Odroid N2+. Hmm…

    • KonQuesting@lemmy.sdf.org
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      How much headroom do you have left on that? I’m considering starting up a public instance and would love to get an estimate for per-user workload on a federated instance.

      • Slashzero@hakbox.social
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        1 year ago

        With just me on the system, CPU is barely ever over 2 -3%. Load average looks good. Memory usage looks fine. You know what? Let me post some graphs for the past 24 hours, which, I’ve pretty much been on here for 24 hours straight. Again, I’m the only user on my instance.

        I’ve mentioned this in a few other threads, but I’m tempted to fire up jmeter and push some load through my instance just to see how it behaves if I slam the system via the API. I just don’t feel like learning the internal API endpoints and all that right at this time though.

        • KonQuesting@lemmy.sdf.org
          link
          fedilink
          English
          arrow-up
          0
          ·
          1 year ago

          Awesome, this is super helpful! I’d be using a very similar setup. It might be best to start small, invite a couple people on, and see how that memory scales. I’ll be avoiding any auto-scaling unless it becomes a much bigger project.

          • Slashzero@hakbox.social
            link
            fedilink
            English
            arrow-up
            1
            ·
            1 year ago

            Well, ideally each service would have their own dedicated resources to begin with. But, given all of the lemmy services + Postgres are running on 2 cores with 2GB of RAM, that’s pretty impressive.

            Anyway, autoscaling doesn’t necessarily solve scaling issues without a lot of thought and planning. It’s not always as simple as throwing more hardware at the problem, as I’m sure you already know.