this post was submitted on 06 Nov 2023
178 points (95.4% liked)

Open Source

30349 readers
2247 users here now

All about open source! Feel free to ask questions, and share news, and interesting stuff!

Useful Links

Rules

Related Communities

Community icon from opensource.org, but we are not affiliated with them.

founded 5 years ago
MODERATORS
 

Most of the problems in the current internet landscape is caused by the cost of centralized servers. What problems are stopping us from running the fediverse on a peer to peer torrent based network? I would assume latency, but couldn't that be solved by larger pre caching in clients? Of course interaction and authentication should be handled centrally, but media sharing which is the largest strain on servers could be eased by clients sending media between each other. What am I missing? Torrenting seems to be such an elegant solution.

top 50 comments
sorted by: hot top controversial new old
[–] taladar@sh.itjust.works 73 points 10 months ago (1 children)

Authentication and authorization can not be handled centrally if the host performing the actual action you want to apply those to can not be trusted.

Media sharing is mainly a legal problem. With decentralized solutions you couldn't easily delete illegal content and anyone hosting it would potentially be legally liable.

[–] uis@lemmy.world 2 points 10 months ago

and anyone hosting it would potentially be legally liable.

Bullshit in many countries. Well, I know two of them: USA and Russia. First requires intention, second has explicitly excludes P2P from liability.

[–] TCB13@lemmy.world 34 points 10 months ago (4 children)

Better question: how come we aren't all using XMPP for chat instead of the current mess we have? Why can’t we just bring back XMPP with encryption? I don’t get people, all for federation and shit but when it comes to messages everyone suddenly forget that XMPP is the original and truly open messaging solution. You can message anyone by email no matter what’s or where’s their server. Pretty much like lemmy, true interoperability.

[–] Asudox@lemmy.world 26 points 10 months ago (1 children)

Currently, attempts at bringing back XMPP has been horrible. Some instances have different plugins, which often break things. Some have e2ee some don't. XMPP is great but the current attempts are bad. Matrix is good enough for now.

[–] TCB13@lemmy.world 11 points 10 months ago* (last edited 10 months ago) (6 children)

I believe the worst part of XMPP isn't the instances but the lack of a decent cross platform client that actually supports everything and has a decent UI. For eg. iOS clients are all shit. Without decent clients and push notifications people won't be using XMPP ever.

Matrix is good enough for now.

Questionable...

[–] Apollo2323@lemmy.dbzer0.com 13 points 10 months ago (1 children)

What's wrong with Matrix in your opinion? I found it works fine for chat and group chat. Maybe video and audio calls lack but other than that it works fine.

[–] Asudox@lemmy.world 3 points 10 months ago (3 children)

There are some things such as metadata leakage and the server isn't the best at being lightweight.

[–] Apollo2323@lemmy.dbzer0.com 3 points 10 months ago

I have read that they have improve the metadata leakage. And the server being lightweight is being working on too , on the new version.

[–] TCB13@lemmy.world 1 points 10 months ago

Exactly my concerns, also Matrix isn't an open standard, truly standardized. It is way more prone to be taken over by some company or ecosystem later on. For what's worth what is even the Matrix Foundation? Where does the money come from?

load more comments (1 replies)
load more comments (5 replies)
[–] mark@programming.dev 22 points 10 months ago* (last edited 10 months ago) (1 children)

I feel the same way about RSS feeds. It's a technology meant to keep up with updates on nearly anything across the internet. Even social media sites. It's been available for ages. But no one is pushing for sites to provide them. 🤷‍♂️

[–] TCB13@lemmy.world 5 points 10 months ago

That's why I use FreshRSS.

[–] NeryK@sh.itjust.works 13 points 10 months ago
[–] uis@lemmy.world 1 points 10 months ago (1 children)

SMTP is federation too. But certain megacorp basically fenced off huge chunk of users.

[–] TCB13@lemmy.world 2 points 10 months ago

Yes, that's exactly by point. SMTP is federation, you can setup your own server and have interoperability with others and have 100% of its features working right, so you aren't locked in to those megacorps. Chat applications should use XMPP to get the same for chat/video.

[–] turkalino@lemmy.yachts 33 points 10 months ago (1 children)

Torrenting requires way more resources than people realize. It's easy to look at your torrents' download speeds and think "oh, that's less than a normal download, like from Steam, so it must not take nearly as many resources" -- it's not all about bandwidth. The amount of encryption and hashing involved in torrenting is fairly CPU heavy (every ~4 MB piece has to be hashed and verified), especially if your CPU doesn't have onboard encryption hardware (think mobile devices). The sheer number of connections involved even in just one torrent can also bog down a network like you wouldn't believe -- anyone who runs a home seedbox can attest.

[–] uis@lemmy.world 3 points 10 months ago

"oh, that's less than a normal download, like from Steam, so it must not take nearly as many resources"

For me it's always more.

The amount of encryption and hashing involved in torrenting is fairly CPU heavy

Same amount of encryption https requires. Hashing is completely optional and is not required for operation, encription is optional too, but other peers may require it.

every ~4 MB piece has to be hashed and verified

Which is same. I'm not sure, but steam probably verifies file at some stage too.

The sheer number of connections involved even in just one torrent can also bog down a network like you wouldn't believe -- anyone who runs a home seedbox can attest.

There is not difference between 4kpkts for 500 connections vs 4kpkts for 1 connection for network itself. IP and UDP don't even have such concept. Network is stateless. But shitty routers with small conntrack table on the other hand...

[–] makeasnek@lemmy.ml 28 points 10 months ago* (last edited 10 months ago)

The short answer is that while torrents show great possibility for content distribution (as an alternative to CDNs for example), they inherently rely on some centralized resources and don't make sense for a lot of use cases. Most websites are a bunch of small files, torrenting is really much more useful for offloading large bandwidth loads. On small files, the overhead for torrents is a waste. That's why your favorite linux ISO has a torrent but your favourite website doesn't.

One major issue is difficulty in accurately tracking the contribution of each member of the swarm. I download a file and I seed it to the next person, sounds great right? But what if the next person doesn't come along for a long time? Do I keep that slot open for them just in case? How long? How I prove I actually "paid my dues" whether that was waiting for peers or actually delivering to them? How do we track users across different swarms? Do we want a single user ID to be tracked across all content they've ever downloaded? When you get into the weeds with these kinds of questions you can see how quickly torrenting is not a great technology for a number of use cases.

Being somewhat centralized, by the way, is how BitTorrent solved the spam issue which plagued P2P networks prior to it. Instead of searching the entire network and everything it contains (and everything every spammer added it to it), you instead rely on a trusted messenger like a torrent index to find your content. The torrent file or magnet link points to a link in a DHT and there you go, no need to worry about trusting peers since you are downloading a file by hash not by name. And you know the hash is right because some trusted messenger gave it to you. Without some form of centralization (as in previous P2P networks), your view of the network was whatever your closest peers wanted it to be, which you essentially got assigned at random and had no reason to trust or not trust. You couldn't verify they were accurately letting you participate in the wider network. Even a 100% trustworthy peer was only as good as the other peers they were connected to. For every one peer passing you bad data, you needed at least two peers to prove them wrong.

Blockchain gets us close to solving some of these problems as we now have technology for establishing distributed ledgers which could track things like network behavior over time, upload/download ratio, etc. This solves the "who do I trust to say this other peer is a good one?" problem: you trust the ledger. But an underlying problem to applying Blockchain to solve this problem is that ultimately people are just going to be self-reporting their bandwidth. Get another peer to validate it, you say? Of course! But how do we know that peer is not the same person (how do we avoid sybil attacks)? Until we have a solid way to do "proof of bandwidth" or "proof of network availability", that problem will remain. There are many people working on this problem (they've already solved proof of storage so perhaps this could be solved in a similar way) but as of right now I know of no good working implementation that protects against sybil attacks. Then again, if you can use blockchain or some other technology to establish some kind of decentralized datastore for humanity, you don't need torrents at all as you would instead be using that other base layer protocol for storage and retrieval.

IPFS was intended as a decentralized replacement for much of the way the the current internet works. It was supposed to be this "other protocol", but the system is byzantinely complex and seems to have suffered from a lack of usability, good leadership, and promotion. When you have an awesome technology and nobody uses it, there are always good reasons for lack of adoption. I don't know enough about those reasons to really cover them here, but suffice to say they actually do exist. Then again, IPFS has been around for a while now (15 years?) and people use it for stuff so clearly it has some utility.

That said, if you want to code on this problem and contribute to helping solve data storage/transmission problems, there are certainly many OSS projects which could use your help.

[–] onlinepersona@programming.dev 27 points 10 months ago (1 children)

You're basically talking about IPFS. I think their problem is that they gave it a local HTTP interface and the documentation is... in need of improvements.

[–] Acters@lemmy.world 7 points 10 months ago (1 children)

I always shy away from newer tech because of lackluster documentation and poor leadership. The latter is rare enough. Without proper documentation, I feel like I have to read the code and make my own notes to put into their documentation platform. Which is not what I want to do when I use it. Contributing is nice, but when doing something a core member would do without credit, it will dissuade me from participating.

[–] onlinepersona@programming.dev 1 points 10 months ago

I know that feeling. Curiosity often gets the better of me though: I'm a nix / NixOS user - amongst the worst documented projects I've come across.

[–] Eggymatrix@sh.itjust.works 12 points 10 months ago (1 children)

I would not want to neither deal with security issues nor pay the data costs associated with some an app being able to connecting to my phone to download media

[–] henrikx@lemmy.dbzer0.com 2 points 10 months ago (1 children)

What if not everyone had to be a seeder?

[–] SomeoneSomewhere@lemmy.nz 19 points 10 months ago (2 children)

You rapidly end up with a freeloader issue.

[–] henrikx@lemmy.dbzer0.com 8 points 10 months ago

Don't see how this is much different to today's way of doing things where pretty much everyone is a freeloader to the centralized server. The major benefit is that it doesn't have to be just one server anymore.

[–] whyNotSquirrel@sh.itjust.works 8 points 10 months ago (1 children)

not if you give benefits to seeder, just like private trackers

[–] folkrav@lemmy.ca 5 points 10 months ago

The trackers themselves are centralized. The .torrent file you download from a private tracker has a unique private ID tied to your account, which the torrent client advertises to the tracker when it phones home to the announce URL, alongside your leech/seed metadata.

[–] andruid@lemmy.ml 11 points 10 months ago (2 children)

For me I've had issues with getting organzational support for use anything close to p2p, with things like "keep that bot net off my system" being said. On personal side I had issues with ISPs assuming traffic was illegal in nature and sending me bogus cease and desist notices.

Agreed though. At least webrtc has a strong market. IPFS and other web3 things also have tried to find footholds in common use, so the fight isn't over for sure!

[–] andruid@lemmy.ml 3 points 10 months ago

Another good example in the fediverse space is peertube too!

[–] uis@lemmy.world 2 points 10 months ago

On personal side I had issues with ISPs assuming traffic was illegal in nature and sending me bogus cease and desist notices.

On the other hand check if you can sue them for bogus cease and desists. Of you can, do it after changing ISP.

[–] derrickoswald@sh.itjust.works 10 points 10 months ago

In my mind, Veilid (https://veilid.com/) is trying to solve this exact problem.

It's only been public since the last DefCon, so it's a little raw at the moment. That said, it is also only a framework - specific applications need to be built.

See also their Discord: https://discord.com/channels/1077244355439509565/ and GitLab: https://gitlab.com/veilid/

[–] einlander@lemmy.world 6 points 10 months ago (3 children)
[–] xoggy@programming.dev 21 points 10 months ago (1 children)

Gotta love the heavy use of buzzword technologies and no actual information on what is actual is. Then you click the "How does it work?" button and it takes you to a Google powerpoint... so much for the sleek website design.

[–] ICastFist@programming.dev 3 points 10 months ago

A side effect that is very common with anything related with cryptocurrencies

[–] blackfire@lemmy.world 3 points 10 months ago

This is the only system i am aware of using torrent based content sharing. Its not a great system though as you are essentially downloading a whole archive everytime you connect so it just grows and grows unless you set some retention

[–] jonuno@lemmy.ml 2 points 10 months ago

Nice, a loading screen app

[–] Dunstabzugshaubitze@feddit.de 6 points 10 months ago

Most clients are web browsers and support for torrents in http is the same as for every other file.

So that would only give us a use for torrents as a form of content distribution plattform to get the actual files closer to the client.

In cases where we have actual non browser clients: i like to curate what i am distributing and don't want to distribute anything i happen stumble upon or would you be willing to store and more importantly share everything you find on 4chan or that might show up in your mastodon feed?

[–] gila@lemm.ee 3 points 10 months ago (1 children)

The instance is the aggregator, if it's P2P then the aggregation is done by the client. In a torrent swarm you contribute bandwidth, not processing power

[–] bigboismith@lemmy.world 2 points 10 months ago (2 children)

I'm not necessarily experienced with server hosting, but isnt bandwidth the primary cost of fediverse instances for example? There shouldn't be to much logical work compared to delivering content?

[–] themoonisacheese@sh.itjust.works 5 points 10 months ago (1 children)

Fediverse instances with image hosting are bandwidth limited, but that's just a normal result of image hosting. If you remove image hosting then the bottleneck becomes processing power again.

[–] cogman@lemmy.world 2 points 10 months ago (1 children)

AFAIK, the main bottleneck is data storage. Related to processing power, but also IO and having a central source of truth.

[–] themoonisacheese@sh.itjust.works 2 points 10 months ago

Sure, but data storage is quite cheap these days. I'm not saying it isn't a problem, but a 12 Tb raid goes a really long way, or AWS s3 charges pennies per GB per month and solves all your problems if you're prepared to spend tens of dollars per month.

Bandwidth on the other hand is either inaccessible (read: you have a normie ISP that has at most 2 speeds to sell you and neither of them have guarantees), or extremely expensive, on the order of thousands per month. On top of that, if you happen to pay AWS for storage, each request must be forwarded to AWS, converted in some way by your server then sent to the client, which means it eats both up and down bandwidth. Of course, if you know what you're doing you can use Amazon's CDN but at this point administering your instance is a full time job and your expenses are that of small company.

[–] gila@lemm.ee 3 points 10 months ago* (last edited 10 months ago)

Sure, but this is largely because currently each client doesn't need to aggregate the whole fediverse. In a decentralised network, you can't split the sum total of processing required to run the fediverse equally amongst peers. Each peer would need to do more or less the same aggregation job, so the total processing required would be exponentially more than with the current setup. You could still argue it's a negligible processing cost per client, but it's certainly way less efficient overall even if we assume perfect i/o etc in the p2p system and even if the client only needs to federate the user selected content

Also just practically deploying a client app that can federate/aggregate constantly in the background (kinda required for full participation) and scale with the growth of fedi without becoming a resource hog I imagine would be pretty tough, like maybe possible yeah but I feel like it makes sense why it isn't like that also

[–] neo@lemmy.comfysnug.space 1 points 10 months ago

Because it became heavily associated with illegal activity.

load more comments
view more: next ›