If it’s for more than a minute I’ll screw in VGA and DVI cables
If it’s for more than a minute I’ll screw in VGA and DVI cables
20 years ago there were 2000000000 fewer people in the world.
BTRFS deciding it’s corrupt and refusing even read only access.
Edit: You beat it by trashing the disk, using any other file system, restoring from backup and accepting any losses.
Careful. You might get hit by a bus.
It stole all my data. It’s a bit of a clusterfuck of a file system, especially one so old. This article gives a good overview: https://arstechnica.com/gadgets/2021/09/examining-btrfs-linuxs-perpetually-half-finished-filesystem/ It managed to get into a state where it wouldn’t even let me mount it readonly. I even resorted to running commands of which the documentation just said “only run this if you know what you’re doing”, but actually gave no guidance to understand - it was basically a command for the developer to use and noone else. It ddn’t work anyway. Every other system that was using the same disks but with ext4 on their filesystems came back and I was able to fsck them and continue on. I think they’re all still running without issue 6 years later.
For such an old file system, it has a lot of braindead design choices and a huge amount of unreliability.
I gave up trying to setup a Mastodon server in docker. Lemmy was pretty tricky at the time as the docs were wrong. My email server was a bit tricky, but I’ve not really done much to tinker with it in the proceeding 6 years, so was worth it.
Why fake serial numbers?
I used btrfs once. Never again!
Are you saying SSDs are faster than HDDs?
I was thinking Proxmox would add a layer between the raw disks and the VM that might interfere with ZFS, in a similar way how a non IT more HBA does. From what I understand now, the passthrough should be fine.
The server runs Proxmox and one of the VMs runs as a fileserver. Other VMs and containers do other things.
I won’t be running ZFS on any solid state media, I’m using spinning rust disks meant for NAS use.
My desire to move to ZFS is bitrot prevention and as a result of this:
Good point. Having a small VM that just needs the HBA passed through sounds like the best idea so far. More portable and less dependencies.
I’m starting to think this is the way to do it because it loses the dependency on Proxmox to a large degree.
Could this because it’s a RAIDZ-2/3? They will be writing parity as well as data and the usual ZFS checksums. I am running RAID5 at the moment on my HBA card and my limit is definitely the 1Gbit network for file transfers, not the disks. And it’s only me that uses this thing, it sits totally idle 90+% of the time.
Did you have atime
on?
What I have now is one VM that has the array volume passed through and the VM exports certain folders for various purposes to other VMs. So for example, my application server VM has read access to the music folder so I can run Emby. Similar thing for photos and shares out to my other PCs etc. This way I can centrally manage permissions, users etc from that one file server VM. I don’t fancy managing all that in Proxmox itself. So maybe I just create the zpool in Proxmox, pass that through to the file server VM and keep the management centralised there.
I’m not intending to run Proxmox on it. I have that running on an SSD, or maybe it’s an NVME, I forget. This will just be for data storage mainly of photos that one VM will manage and NFS share out to other machines.
Q
Quake Quake II
But maybe U
Unreal Tournament 2004
That was my point