this post was submitted on 03 Apr 2025
131 points (97.1% liked)

Fediverse

32395 readers
346 users here now

A community to talk about the Fediverse and all it's related services using ActivityPub (Mastodon, Lemmy, KBin, etc).

If you wanted to get help with moderating your own community then head over to !moderators@lemmy.world!

Rules

Learn more at these websites: Join The Fediverse Wiki, Fediverse.info, Wikipedia Page, The Federation Info (Stats), FediDB (Stats), Sub Rehab (Reddit Migration)

founded 2 years ago
MODERATORS
 

Every time I go to the piefed frontpage I'm blown away by how much more polished it is. It has all the bells and whistles that lemmy is sometimes missing.

Whats the catch? Why aren't we recommending everyone goes to piefed instead of lemmy?

App support is one thing I can think of.

you are viewing a single comment's thread
view the rest of the comments
[–] poVoq@slrpnk.net 19 points 1 day ago (2 children)

While theoretically true, the main bottleneck with Lemmy seems to be the database performance, so with both projects depending on PostgreSQL for that, I somewhat doubt that Piefed being written in Python will have much noticeable effect in reality.

[–] nickwitha_k@lemmy.sdf.org 3 points 1 day ago

the main bottleneck with Lemmy seems to be the database performance, so with both projects depending on PostgreSQL

Postgres being a bottleneck is a first for me. Not saying it's not possible, just... It's postgres. Wondering if it's more an issue with ORM, etc.

[–] msage@programming.dev 4 points 1 day ago (1 children)

Postgres is so quick if you know how to use it...

[–] nickwitha_k@lemmy.sdf.org 2 points 1 day ago (1 children)

You don't even need to know how to use it very well, in my experience.

[–] msage@programming.dev 2 points 1 day ago

Really depends on many factors. If you have everything in RAM, almost nothing matters.

If your dataset outgrows the capacity, various things start to matter, based on your workload. Random reads need to have good indices (also writes with unique columns), OLAPs benefit from work_mem, >100M rows will need good partitioning, OLTP may even need some custom solutions if you need to keep a long history, but not for every transaction.

But even with >B of rows, Postgres can handle it with relative ease, if you know what you're doing. Usually even on a hardware you would consider absolutely inadequate (last year I migrated our company DB from MySQL to Postgres, and with even more data and more complex workflows we downsized our RAM by more than half).