• 8 Posts
  • 36 Comments
Joined 2 months ago
cake
Cake day: February 3rd, 2026

help-circle
  • Honestly, I think your friend is right, it’s a question of economy of scale. As you scale up there will be less and less wasted resources in overhead. Once you reach the scale where you need hundreds or thousands (or hundreds of thousands) of servers to operate your site you’d likely be able to fairly efficiently dimension the amount of servers you have so that each server is pretty efficiently utilized. Youd only need to keep enough spare capacity to handle traffic bursts, which would also become smaller compared to the baseline load the larger your site becomes.

    Realistically most self-hosted setups will be mostly idle in terms of CPU capacity needed, with bursts as soon as the few users accesses the services.

    As for datacenters using optimized machines there is probably some truth to it. Looking at server CPUs they usually constrain power to each core to add more cores to the CPU. Compared to consumer CPUs where at least high-end CPUs crank the power to get the most single-core performance. This depends heavily on what kind of hardware you are self-hosting on though. If you are using a raspberry-pi your of course going to be in favor, same is probably true for miniPCs. However if you’re using your old gaming computer with an older high-end CPU, your power efficiency is very likely sub-optimal.

    As a “fun” fact/anecdote, I recently calculated that my home server which pulls ~160W comes out as 115kWh in a month. This is a bit closer than I would like to the 150-200 kWh I spend on charging my plug-in hybrid each month… To be fair though I had not invested much in power efficiency of this computer, running the old gaming computer approach and a lot of HDDs.

    That said there is plenty of other advantages with self-hosting, but I’m not sure the environmental angle works out as better overall.





  • It’s extremely unlikely that they are going to do any kind of deep traffic inspection in the router/modem itself. Inspecting network traffic is very intensive though and gives very little value since almost all traffic is encrypted/HTTPS today, with all major browsers even showing scare warnings if’s regular unencrypted HTTP. Potentially they could track DNS queries, but you can mitigate this with DNS over TLS or DNS over HTTPS (For best privacy I would recommend Mullvad: https://mullvad.net/en/help/dns-over-https-and-dns-over-tls)

    And of course, make sure that anything you are self-hosting is encrypted and using proper HTTPS certificates. I would recommend setting up a reverse proxy like Nginx or Traefik that you expose. Then you can route to different internal services over the same port based on hostname. Also make sure you have a good certificate from Letsencrypt



  • It’s the Network effect. At the end of the day, Lemmy is still a lot smaller than Reddit, and of course that means there is less content, especially in more niche communities. The only way to really improve the situation is to grow the network with more users and more diversity.

    I feel like the activity level is good enough to use it mostly, but I also still check in on Reddit regarly because some communities are simply very inactive on here.

    I guess what you can do as an individual is to post more content, participate in the community, and help spread the word that the fediverse even exists.








  • I’m like 90% sure that this post is AI Slop, and I just love the irony.

    First of all, the writing style reads a lot like AI… but that is not the biggest problem. None of the mitigations mentioned has anything to do with the Huntarr problem. Sure, they have their uses, but the problem with Huntarr was that it was a vibe coded piece of shit. Using immutable references, image signing or checking the Dockerfile would do fuck-all about the problem that the code itself was missing authentication on some important sensitive API Endpoints.

    Also, Huntarr does not appear to be a Verified Publisher at all. Did their status get revoked, or was that a hallucination to begin with?

    To be fair though the last paragraph does have a point, but for a homelab I don’t think it’s feasible to fully review the source code of everything you install. It would rather come down to being careful with things that are new and doesn’t have an established reputation, which is especially a problem in the era of AI coding. Like the rest of the *arr stack is probably much safer because it’s open source projects that have been around for a long time and had had a lot of eyes on it.


  • Worth noting that despite the headline this does not have anything to do with the huge outage in the end of 2025.

    The company said the incident in December was an “extremely limited event” affecting only a single service in parts of mainland China. Amazon added that the second incident did not have an impact on a “customer facing AWS service.”

    Neither disruption was anywhere near as severe as a 15-hour AWS outage in October 2025 that forced multiple customers’ apps and websites offline—including OpenAI’s ChatGPT.

    I would also have felt some level of schadenfreude if it turned out that any of the really big incidents in the end of 2025 was a result of managements aggressive pushes for AI coding. Perhaps that would cool off the heads of executives a bit if there were very real examples pf shit properly hitting the fan…


  • The free version is mainly just a number of user and device limit. Although the relaying service might be limited as well, but that should only matter if both of your clients have strict NAT, otherwise the Wireguard tunnels gets directly connected and no traffic goes through Netbirds managed servers.

    You can also self-host the control plane with pretty much no limitations, and I believe you no longer need SSO (which increased the complexity a lot for homelab setups).






  • This article feels a bit like ragebait.

    Yes, this happened once with a company that went bankrupt 2 years after launching their product. They seem to have designed an exceptionally poor product. How does this mean that the enormous engineering failures of this small startup applies to all other car brands?

    Most cars have a very clear separation between core driving software and the infotainment, and the vast majority will never have any software updates so what works, will continue to work (or the other way around). At worst you’ll loose stuff like remote commands, wheatear info, list of charging points/map updates… Things that are kind of dynamic and needs to be regularly updated.