• 0 Posts
  • 33 Comments
Joined 5 months ago
cake
Cake day: January 13th, 2025

help-circle
  • Battery circuits come on enough to be a load that needs to be considered and will show up if you measure load on the device vs load consumed by the components connected to the power supply. In terms of low power devices, it is significant, though not the primary concern. But compared to the pi PSU, the charger not to mention the battery and internal PSU of a laptop, consume way more power and produce way more heat.

    All of the rest assumes needing always on, heavy load processing which isn’t what the post I replied to was talking about. I was specifically replying to idle power load. And in my case, even with a bunch of self hosted applications, most of the time my servers are idling. If I was running a virtualization farm or something that was always under heavy load, then yes, as I mentioned, a single board server isn’t ideal.

    As for disks, I don’t use SSDs on my pis except one that actually does a lot of local data processing. Everything else runs in memory and stores persistent data on my NAS, including logging. Virtual memory/swap is disabled on all and things that need temporary storage/cache of small amounts of data is cached on RAM disks where applications can’t be configured to not use disk caching. The only need for the SD card is for boot and some minimal IO needed for local OS operation. I have a Raspberry Pi 3 B i got about 8 or 9 years or so ago with the same SD card in it.

    They aren’t what I use as a database server, obviously, but they are extremely low power compared to what an old laptop would need and work great for things like pihole, and other network applications as well as being a part if my home kubernetes cluster and run the majority of the cluster’s processes on demand.



  • Not quite. Unless the system has pretty advanced power management and is using very recent technology with high density, it’s unlikely that an x64 chipset will use less power than a comparably powered arm64 chipset. Not just the processor, but the smaller board is actually a power saver and allows it to generate less heat meaning both less power wasted and dissipated as heat as well as less power needed for fans to properly dissipate the heat. I’ve never seen a laptop use 3W at idle when considering the whole device, maybe just the CPU, but not if you include the rest of the components like RAM and disks and power supply. And especially true in a laptop that is old enough that it’s being recycled. Heck, the power supply and charger alone might be using 3W at idle with full battery.

    With a raspberry pi 4, the typical power usage for the 2GB RAM model is 5W under load for the whole device and about half that for idle. Add a couple of watts for the extra memory and wider bus on the 8GB model and other things can add to that, but that’s mostly accurate. The pi 5 is a little more and the 3 is a little less. Of course, the efficiency of the laptop at full load might end up being better than a comparable number of raspberry pis it would take to do the same amount if work, but comparing a single pi or any other reputable arm-based, single board computer to a single laptop at idle is always going to be that way.



  • Problem with Manjaro is they have their own opinionated repository that is not always in sync with Arch because they try to introduce more “stability”. I found this actually caused the opposite in most cases as there are a lot of dependencies that end up being behind and so you can’t install more stable versions of a lot of software. With the complexity of modern software dependencies, it has become a big problem. Also, they have in the past caused lots of problems with AUR and have let their SSL certs expire multiple times. Overall, they just haven’t been reliable IMHO, so I moved to Fedora a while back.


  • Same. I had been using Ubuntu for over a decade for all of my Desktops, but had used CENTOS/Rocky for servers. Now I switched to Fedora for desktop which simplifies things since now only my Raspberry Pis use deb vs rpm.

    Snap is super frustrating and the gate-keeping of updates and features behind the Pro subscription is annoying. I don’t want to have an account if I dint have to. It’s just one more privacy violation waiting to happen with no real benefit to me even if it is free for personal use.


  • I had to get away from Ubuntu because of the recent performance issues and the requirement to have an account to get updates faster. I have used CENTOS and more recently Rocky Linux on all my servers for over a decade, so Fedora was the obvious choice for me. I had problems with desktop applications being missing from rpm repos in the past, but that has greatly reduced improved and flatpack has helped with some stragglers. But I’m still not a fan of RedHat, but Fedora is a little more separate from them than Ubuntu is from Canonical.

    I tried Debian, but it’s not easy to get up and running on a more modern laptop or desktop without a lot of tweaking and kernel mods. It’s a good base OS but not good out of the box desktop OS. Same issue with Arch based distros.




  • I have a separate boot partition so the rest can be encrypted with luks. That’s all that’s needed in a large majority of scenarios. Most other setups end up needing to resize something at some point which in many cases is a total pain.

    On my primary PC I do have a second hard drive for documents and other long term storage files that I want to access more often than on the NAS. This way it’s nearly impossible to lose those files of I reinstall something and it can act as a temporary backup storage for settings files when I do reinstall stuff rather than having a partition that wastes space or runs out of space.






  • Not offering a solution here exactly, but as a software engineer and architect, this is not a Linux only problem. This problem exists across all software. There are very few applications that are fully self contained these days because it’s too complex to build everything from scratch every time. And a lot of software depends on the way that some poorly documented feature worked at the time that was actually a bug and was eventually fixed and then breaks the applications that depended on it, etc. Also, any time improvements are made in a library application it has potential to break your application, and most developers don’t get time to test the every newer version.

    The real solution would be better CI/CD build systems that automatically test the applications with newer versions of libraries and report dependencies better. But so many applications are short on automated unit and integration tests because it’s tedious and so many companies and younger developers consider it a waste of time/money. So it would only work in well maintained and managed open source types of applications really. But who has time for all that?

    Anyway, it’s something I’ve been thinking about a lot at my current job as an architect for a major corporation. I’ve had to do a lot of side work to get things even part of the way there. And I don’t have to deal with multiple OSes and architectures. But I think it’s an underserved area of software development and distribution that is just not “fun” enough to get much attention. I’d love to see it at all levels of software.



  • Most could, but most are also designed not to because adding a virtualization type of layer allows for ways to circumvent it. Anticheat needs to trust the environment it is running in so it can rely on the information. Wine is designed to replicate things it trusts in Windows, but not actually necessarily replicate the way the kernel actually does those things, so the things they are relying on might not mean the same thing as the do in Windows. So they’d need to analyze and possibly implement things a bit differently. This takes time and money and for companies like this, the customer isn’t the user, so they have little reason to cater to users needs. Pro gaming and a few online game companies are their primary customers and they generally don’t want to support Linux anyway.


  • Nope it doesn’t add anything for me. The _netdev option tells mount to wait until the network is connected before attempting to mount. And the nofail option tells it not to error or block the process that called it if the mount doesn’t work or is delayed.

    Now if the mount contains your etc or other critical config files, it could cause problems and maybe you want to wait, so don’t want the nofail. And of course this kind of thing is somewhat OS specific depending on what boot system and service manager, etc., is used, so YMMV, but on Fedora, Rocky, and Ubuntu, it has worked for me for many years.



  • I haven’t used OpenSUSE before, but I don’t really experience those issues, though I don’t use caps lock that way. I use Fedora with Plasma for desktop these days since Ubuntu is heading too corporate for my taste and plain Debian is missing too much hardware support. I’m sure Fedora will eventually, too, but I also use Rocky on all of my server installs so I prefer RHEL-based over Debian-based, for consistency anyway. Install and setup has always been smooth for me. The Discover app is there for installing stuff. It lags a lot, but otherwise makes installing things pretty easy. I’m sure there must be an equivalent for OpenSUSE. That said, Linux does rely on the command line a lot more than windows. In Windows the command line is bolted on, but in Linux it’s more that the GUI is bolted on, though that has smoothed quite a bit and even on Windows the v7 powershell has smoothed out command line a little bit even if powershell commands aren’t that intuitive IMHO. At least this version understands some dos formatted commands. I use Windows 11 for work.