That’s great if it’s your experience.
I’m just saying me and others have consistently had different experiences, and OP can get a better experience at half the price, with the same (or better) energy consumption, all while supporting the Linux ecosystem directly.
That may be, but buying a Mac Mini is like buying a device made from the ground up for Windows, where any other operating system has to reverse engineer 100% of the things to work well, or you have to emulate another OS on it (which comes with its own pitfalls), and it’s 200+€ more expensive than its nearest equivalent.
Every single company I’ve worked at which introduced Apple Silicon to its developers has had headaches with compatibility. The worst I’ve seen was it taking a developer a month to get up and running because the specific component we used didn’t have a build for the specific ARM architecture. Multipass, UTM, podman, docker desktop, all didn’t work until colima and forcing the VM to emulate x86 + forcing docker in the VM to use the x86 image worked. There was a persistent problem with disk IO since it used 9p or whatever. Installing dependencies from scratch meant waiting 30 minutes on the M2.
Why pay a premium for less compatibility and worse specs? Just get yourself something that works, which is cheaper, maybe even supports a company that invests in Linux and its ecosystem, and be able to ask an existing developer community instead of asking the subsection of linux users that run your specific app on however you’re running linux on Appe hardware.
The problem with Mac hardware is that it’s ARM and vertically integrated with everything Apple. Not all hardware is supported by Linux because Apple won’t write any linux drivers and everything is reverse engineered. You’re better off buying something non-Apple which linux properly supports.
If power consumption is an issue for you get, a R9 7950X consumes as much and at times less power than an M1/M2 (I think even M3). Check out GamerNexus’s charts. IINM AMD in W/Ghz performs better than Intel across the board.
No idea where you are, but you can get a small factor PC from one of the vendors that preload linux, or configure a small form factor PC of your liking for cheap and put linux on it. You’ll get more out of your money for the same or better performance with about the same energy consumption (or a bit more).
Somebody I know who happens to live in Hungary got himself this cheap beauty. They deliver all over Europe, but if you live elsewhere on this planet, there probably is something similar like this out there.
A single folder and power consumption is important --> syncthing. It doesn’t have great power consumption, but since the devices aren’t constantly on, you can just start syncthing up on the portable devices when needed. You can configure syncthing to sync when connecting to a specific Wifi, when power saving mode is turned off, I think even specific times.
It’ll run fine on a server and can be configured .
Sail the seas with I2P and anonymous torrents. They can’t stop it.
If this had IPFS support, it would be so cool!
🤔 maybe GTK4 apps look dated to me. Didn’t realise that’s what GTK4 looks like.
Dino has an intentionally simplistic design, but it doesn’t look “dated” at all.
That’s highly subjective, but I’ve shown some Gtk3 apps to people at work and the most expressive first reaction I got was “ew”. Dino and others getting that exact reaction wouldn’t be surprising.
it just doesn’t have a major centralized provider of them that in exchange siphons up all your personal data like Element & Beeper does. But you can easily self-host the available bridges for XMPP
And this is another reason why it isn’t prominent. “Grandma, all you need to do is host an XMPP server. It’s incredibly easy”.
I looked at Dino and another one mentioned here and they look dated. Windows 95 feel with better anti-aliasing, rounder corners, but same colors? Gtk 2 or something?
Pidgin was my messenger of choice to communicate with people on facebook, gmail, and a few other protocols back then.
But yeah, my experience with XMPP wasn’t good and if they don’t have bridges, there isn’t much of a reason for me to switch right now. It doesn’t seem to provide any advantages over Matrix.
Matrix clients are simple, easy, and nice to look at. The matrix server might need more resources, but it comes with everything out of the box. There’s no need to fiddle with extensions and their weird naming, and hope that the other server/client also supports the extension. Also, are there bridges to other protocols?
I remember trying to get encryption working on Pidgin and it was all around a bad experience.
XMPP might be as powerful or more powerful than matrix, but nothing about it screams modern. It’s like IRC for Gen X’ers.
@Two9A@hachyderm.io
Missed the XMPP acronym. Probably needs to scan the title too
The original service won’t know if it’s you accessing them through the server IP or not. What you could do is add your service to the public instance list and generate organic traffic by virtue of it being used by others.
I did have to make sure some services were fault tolerant if an encrypted volume was unavailable when the OS booted
How did you achieve that? systemd dependency?
Why? What would be the problem?
On linux, you’re probably using LUKS. That has a header with the keys at the beginning of each encrypted volume. If those keys (or key if you only have one) is corrupted and you don’t have a backup of that, you’re fucked.
The next problem is that data recovery tools mostly don’t support decryption. They scan regions or the entire drive for recognizable things like partition headers, partition tables, file types, etc. if those are encrypted, well…
If you are able to decrypt a partition, then it might work as it will show up like any other device in /dev/mapper/
and you could do recovery /dev/mapper/HDD
. However, I have no idea what data corruption does to encryption algorithms. If one part of what is being decrypted is faulty, what does that do to the entire thing?
This mostly comes from a lack of knowledge on my part. IIRC encryption depends on hashsums -> if you change what’s being decrypted/encrypted, the entire hashsum is incorrect and thus all the data shouldn’t be able to be decrypted. But I might be wrong - I’ll gladly be wrong on this.
I want to, but haven’t found the time to make a strategy on how to move over the data. It would take a bunch of shuffling as all drives are in use. The next problem is decrypting at boot and securely storing the decryption key - if I choose to use a decryption key at all. Maybe it’ll be a usb key that I have to plug into the server when starting it, or I have to setup decryption of the system over SSH, but that means automated restarts are… difficult.
Not sure how to tackle the problem yet…
Recovering data from a corrupted, encrypted drive is way trickier than from a simply corrupted drive, I imagine.
Some European ones because the domains have European TLDs. .eu
for example is only available by EU registrars IINM. But also, I do my best to keep the money local where I can.
There’s also rootless docker. There shouldn’t be any more firewall shenanigans.
What’s up with Owncloud? Why did devs leave for Nextcloud? And what happened to prevent that from happening again?
I too dislike that Nextcloud is in PHP, but if Owncloud went closed-source, then opened it up again (not saying that’s the story here), who’s to say it won’t happen again? Putting my eggs in that basket might seem quite dangerous as I don’t want my server to suddenly stop working and sit behind a paywall or something because management decided they want to make a quick euro.
Anti Commercial-AI license