Edited: this post to be the Lemmy.World federation issue post.

We are now ready to upgrade to postgres 16. When this post is 30 mins old the maintenace will start


Updated & Pinned a Comment in this thread explaining my complete investigation and ideas on how the Lemmy app will & could move forward.

  • Tiff@reddthat.comOPM
    link
    fedilink
    English
    arrow-up
    4
    ·
    8 months ago

    Yes… it is VERY annoying. We have so much resources available and lemmy/postgres will not use them

      • Tiff@reddthat.comOPM
        link
        fedilink
        English
        arrow-up
        4
        ·
        8 months ago

        Yes. Unfortunately the information gleamed boils down to two reasons:

        • their db was slow to respond
        • their db server ended up being 25ms away from their backend servers which caused the slowness.

        Our db server is occasionally slow to respond, but most requests from LW complete in less than 0.1 second. Unfortunately there are times when they take longer. These longer ones are going to be the problem (I believe). As all activities are sequential servers can only catch up as fast as they can process them.

        What i’ve found in the past 30 seconds is that it is not necessarily out database that is the problem but possibly the way lemmy handles the federation. I’m chatting with some of the admins on reddthat and making pretty graphs while looking at walls of logs.