I just spun up Lemmy on my Kubernetes cluster with nginx-unprivileged and ingress-nginx. All is well so far! I’m thinking about posting the Kustomization manifests and continuing to maintain and publish OCI’s per version release of Lemmy.

  • dudeami0@lemmy.dudeami.win
    link
    fedilink
    English
    arrow-up
    7
    ·
    1 year ago

    I currently am running the instance I am responding from on kubernetes. I published a helm chart, and others are working on them too. I feel being able to quickly deploy a kubernetes instance will help a lot of smaller instances pop up, and eventually be a good method of handling larger instances once horizontal scaling is figured out.

      • Andreas@feddit.dk
        link
        fedilink
        English
        arrow-up
        4
        ·
        1 year ago

        Saved this comment. It claims that the Lemmy frontend and backend are stateless and can be scaled arbitrarily, as can the web server. The media server (pict-rs) and Postgres database are the limitations to scaling. I’m working to deploy Lemmy with external object storage to solve media storage scaling and there’s probably some database experts figuring out Postgres optimization and scaling as well. None of the instances are big enough to run into serious issues with vertical scaling yet, so this won’t be a problem for a while.

        • blazarious@mylem.me
          link
          fedilink
          English
          arrow-up
          2
          ·
          1 year ago

          I’ve got my pictrs backed by an S3, so that should scale well.

          I had some issues with the image server, though, and I had multiple of them running at the same time at some point, so that may have been the cause.

      • dudeami0@lemmy.dudeami.win
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        I’m not sure if there really is issues, I think it’s just new ground since most lemmy instances have been able to run on a single node due to the low populations. It seems most large public instances are just adding bigger servers to deal with the problem short term.

        From what I can tell (I am not an expert in this field), it seems most of the architecture would spread horizontally without much issue. I haven’t seen anywhere this is done yet, but I could be missing the obvious.

        The lemmy backend api just takes HTTP requests (and at the present websockets, but this is changing in 0.18 to only HTTP requests), and it uses postgres as the backend storage. Using a kubernetes postgres operator to scale the database and then running multiple lemmy backend api instances (and frontend as needed) seems like it would work, or would require minimal work to get running.

        • tj111@lemmy.fmhy.ml
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          Thanks for the input, yeah just rest apis over a postgres db is pretty standard k8s setup, so unless there’s something weird in the middle it would work fine. Curious why the OG design leaned into sockets over rest though, that’s an interesting choice.

    • anthr76@lemmy.kutara.ioOP
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 year ago

      Very neat! I also considering writing a helm-chart with my close friend’s amazing helm library. In the end I decided against it since this is a pretty simple deployment as of today. Tomorrow I will clean up the Kustomize manifests and some CI with a non-federated config file and post it :)