![](https://lemmy.world/pictrs/image/d7844a9a-d604-4147-8867-fb19b2db2b70.jpeg)
![](https://lemmy.world/pictrs/image/4271bdc6-5114-4749-a5a9-afbc82a99c78.png)
I was in the same place as you a few years ago - I liked swarm, and was a bit intimidated by kubernetes - so I’d encourage you to take a stab at kubernetes. Everything you like about swam kubernetes does better, and tools like k3s make it super simple to get set up. There _is& a learning curve, but I’d say it’s worth it. Swarm is more or less a dead end tech at this point, and there are a lot more resources about kubernetes out there.
They are, but I think the question was more “does the increased speed of an SSD make a practical difference in user experience for immich specifically”
I suspect that the biggest difference would be running the Postgres DB on an SSD where the fast random access is going to make queries significantly faster (unless you have enough ram that Postgres can keep the entire DB in memory where it makes less of a difference).
Putting the actual image storage on SSD might improve latency slightly, but your hard drive is probably already faster than your internet connection so unless you’ve got lots of concurrent users or other things accessing the hard drive a bunch it’ll probably be fast enough.
These are all Reckons without data to back it up, so maybe do some testing
Sounds like a great idea - I suspect the biggest obstacle will be finding someone at the home who is confident enough in what to do with it to be willing to accept it.
I’ve run into similar issues with schools where they are hesitant to accept donations of things like that because they don’t want to be saddled with equipment they don’t know how to use and maintain. Maybe worth seeing if you can raise a bit of money for a second hand Xbox or something?
If you figure it out, I know several companies that would be more than willing to drop 7 figures a year to license the tech from you
git-annex maybe?
Yeah, they are mostly designed for classification and inference tasks; given a piece of input data, decide which of these categories it belongs to - the sort of things you are going to want to do in near real time, where it isn’t really practical to ship off to a data centre somewhere for processing.
Seems pretty reasonable. At the end of the day people have to eat, so projects like this either trundle on as hobby-and-spare-time projects for a few years until people get bored and burnt out, or you find a way to make working on the project a paid gig for the core people
“is not exactly tailored to my specific requirements, aesthetic preferences and built using technology I’m familiar with” = “sucks” apparently
This is an “x-y question” - what are you actually trying to achieve?
Clearly you are concerned about… someone… knowing your home IP address - who, and why?
Removed by mod
Definitely interested - is the mainline situation any better than with ARM?
I’ve been bitten before with a device that “supports” a major distribution, but only if you install our custom pre-built image (good luck auditing what we’ve tweaked) and only with our special pre-built kernel that isn’t even an LTS version, and has a bunch of patches applied to support whatever weird peripherals we decided to throw on the board, and will get exactly 0 updates after the initial release.
Raspberry Pi gets around this by being big enough to get buy in from vendors (Ubuntu distributes a special kernel + firmware bundle), but support for all the other smaller knock offs seem shaky at best
I have a machine at work (no screenshots sorry) that is using ~200GB of RAM as disk cache and still has over 100GB of free RAM - not “used for cache but can be freed if an application needs it”, actually genuinely unallocated.
As in, hardware RAID is a terrible idea and should never be used. Ever.
With hardware RAID, you are moving your single point of failure from your drive to your RAID controller - when the controller fails, and they fail more often then you would expect - you are fucked, your data is gone, nice try, play again some time. In theory you could swap the controller out, but in practice it’s a coin flip if that will actually work unless you can find exactly the same model controller with exactly the same firmware manufactured in the same production line while the moon was in the same phase and even then your odds are still only 2 in 3.
Do yourself a favour, look at an external disk shelf/DAS/drive enclosure that connects over SAS and do RAID in software. Hardware RAID made sense when CPUs were hewn from granite and had clock rates measures in tens of megahertz so offloading things to dedicated silicon made things faster, but that’s not been the case this century.
Por qué no los dos?
I’d be super surprised if this was western intelligence. Stuxnet escaping Natanz was an accident, and there is no way that an operation like this would get approved by the NSAs Vulnerabilities Equities Process.
My money would be MSS or GRU. Outside chance this is North Korean, but doesn’t really feel like their MO
It’s a really wicked problem to be sure. There is work underway in a bunch of places around different approaches to this; take a look at SBoM (software bill-of-materials) and reproducible builds. Doesn’t totally address the trust issue (the malicious xz releases had good gpg signatures from a trusted contributor), but makes it easier to spot binary tampering.
The reason openssh links liblzma in the first place is to enable a systemd feature, so naturally “systemd bad, it’s proximity to a security issue is yet more proof that a pile of shell scripts in a trenchcoat is a superior init system” etc
Also,
Arch is the most stable
Are you high?