Funny thing is that the only reason I’ve found *arrs a few years ago was Netflix deciding to be stupid, making me look at how I can manage my local library better nowadays.
Funny thing is that the only reason I’ve found *arrs a few years ago was Netflix deciding to be stupid, making me look at how I can manage my local library better nowadays.
Performance of the snapdragons is roughly that of an i7 from a decade ago - so yes, it’s a good machine for office tasks and light development, but in no way suitable for gaming. That’s not a Windows problem, though, just the hardware is not suitable for that.
I’ve been using an Arm notebook with Windows for over a year now (not as main system, but development system for a customer project). I’m running a lot of x86 software (like Emacs) as a gcc port for Windows/Arm is being developed only now - with no problems. It integrates nicely into the native stuff - which is one area where you run into issues on the Mac: If you start a shell in rosetta it’s annoying to make calls to native arm binaries.
The only issue I ran into were some drivers not available for Arm - emulation layer (unsurprisingly) just is for userland, not kernel drivers. Also x86 emulation isn’t working well if Windows is running in a virtual machine on MacOS - but supposedly that’ll be fixed in the upcoming Windows release.
All of this only applies to Windows 11 - if for some reason you decide to run Windows 10 on Arm you’re in a world of pain.
Windows 11 has pretty good x86 emulation, both 32 and 64bit - imo better than what macos does with rosetta. Windows 10 for arm is just a pretty broken tech preview, though.
Pretty much same here - I kept an x230 alive until I had to accept earlier this year that it just is bad for overall productivity, and ended up getting a macbook. None of the newer thinkpads are good - and they’re still one of the less bad manufacturers.
There’s also enough stuff I don’t like about the mac - but the current keyboard is one of the better notebook keyboards available right now, and if you want long battery life, lots of RAM and a lot of CPU power available in a compact device they’re the only manufacturer currently offering that.
It helps not having a computer with specs from a decade ago.
It should work - possible that it won’t let you create a one disk raid 0, but creating a one disk raid 1 and then converting it to a two disk raid 0 should word. It’s been years since I played with a pure raid 0 (don’t see much sense in them), but managed conversion back then.
If your install is using LVM (which anything installed over a bit more than a decade should be) you can set up the new second drive as a RAID with a missing device, add it as additional PV, use pvmove to move all PEs to the RAID, remove the old PV, and now add that disk to the RAID.
The annoying aspect from somebody with decades of IT experience is - what should happen is that crowdstrike gets sued into oblivion, and people responsible for buying that shit should have an epihpany and properly look at how they are doing their infra.
But will happen is that they’ll just buy a new crwodstrike product that promises to mitigate the fallout of them fucking up again.
It’s already in the name - XDG stands for X Desktop Group (nowadays freedesktop), which works on interoperability for desktop environments. In a pure shell environment (or even if you’re not running a full desktop) none of the XDG variables are defined, and especially in shell environments the default fallbacks specified by XDG are not necessarily what the operator would expect.
Probably half the entries in that list are not GUI apps, and XDG doesn’t apply (though some still support it). For some others there (like emacs) XDG is used if it exists.
I’m fine with that. I don’t want to talk with people - I just want an email address to write to.
Don’t have links anymore, but few months ago I came across some startup trying to sell AI that watches your production environment and automatically optimizes queries for you.
It is just a matter of time until we see first AI induced large data loss.
browsers did vertical tabs in the 90s and it flopped
There are extensions for that. Which are worse than they used to be because they didn’t provide APIs enabling to do that properly, about 10 fucking years after they dropped the old APIs. There are a lot of other feature requests from back then open, often even filed years before they went through with dropping the old APIs. The best way of doing custom keyboard shortcuts in Firefox is still injecting Javascript into each page, with all the shortcomings this has. Usability of Firefox is way worse nowadays than what it was 10 years ago - and I do understand (and agree) with the decision to dump the legacy APIs, but you can’t just break functionality lots of people use, and not provide APIs in over a decade to fix that.
I’m trying other browsers now and then, but every single one is a dumpster fire. At least the Firefox dumpster fire is a bit less out of control - but that’s the most positive thing I can say about it nowadays.
Did pretty much the same with a new server recently - spent ages debugging why it didn’t find the SAS disks. Turns out, disks like to have power connected, and no amount of debugging on software level will help you.
I was referring to work setups with the overengineering - if I had a cent for every time I had to argue with somebody at work to not make things more complex than we actually need I’d have retired a long time ago.
Unless you are gunning for a job in infrastructure you don’t need to go into kubernetes or terraform or anything like that,
Even then knowing when not to use k8s or similar things is often more valuable than having deep knowledge of those - a lot of stuff where I see k8s or similar stuff used doesn’t have the uptime requirements to warrant the complexity. If I have something that just should be up during working hours, and have reliable monitoring plus the ability to re-deploy it via ansible within 10 minutes if it goes poof maybe putting a few additional layers that can blow up in between isn’t the best idea.
Everything is deployed via ansible - including nameservices. So I already have the description of my infra in ansible, and rest is just a matter of writing scripts to pull it in a more readable form, and maybe add a few comment labels that also get extracted for easily forgettable admin URLs.
Shitty companies did it like that back then - and shitty companies still don’t properly utilize what easy tools they have available for controlled deployment nowayads. So nothing really changed, just that the amount of people (and with that, amount of morons) skyrocketed.
I had automated builds out of CVS with deployment to staging, and option to deploy to production after tests over 15 years ago.
Have been using it since late 90s, stopped using it with the shutdown of SixXs as there still were no viable native options in pretty all my infra locations. Recently started using it again as I finally have an ISP providing proper v6.