• 0 Posts
  • 16 Comments
Joined 2 years ago
cake
Cake day: July 30th, 2023

help-circle
  • Ah kay, definitely not a RAM size problem then.

    iostat -x 5 Will print out per drive stats every 5 seconds. The first output is an average since boot. Check all of the drives have similar values while performing a write. Might be one drive is having problems and slows everything down, hopefully unlikely if they are brand new drives.

    zpool iostat -w Will print out a latency histogram. Check if any have a lot above 1s and if it’s in the disk or sync queues. Here’s mine with 4 HDDs in z1 working fairly happily for comparison:

    Here's mine with 4 HDDs  in z1 working fairly happily for comparison

    The init_on_alloc=0 kernel flag I mentioned below might still be worth trying.



  • After some googling:

    Some Linux distributions (at least Debian, Ubuntu) enable init_on_alloc option as security precaution by default. This option can help to prevent possible information leaks and make control-flow bugs that depend on uninitialized values more deterministic.

    Unfortunately, it can lower ARC throughput considerably (see bug).

    If you’re ready to cope with these security risks 6, you may disable it by setting init_on_alloc=0 in the GRUB kernel boot parameters.

    I think it’s set to 1 on Raspberry Pi OS, you set it in /boot/cmdline.txtI think.

    Exhaustive ZFS performance tuning guide


  • sync=disabled will make ZFS write to disk every 5 seconds instead of when software demands it, which maybe explains your LED behavior.

    Jeff Geerling found that writes with Z1 was 74 MB/sec using the Radxa Penta SATA HAT with SSDs. Any HDD should be that fast, the SATA hat is likely the bottleneck.

    Are you performing writes locally, or over smb?

    Can try iostat or zpool iostat to monitor drive writes and latencies, might give a clue.

    How much RAM does the Pi 5 have?


  • Personally I think it’s fallen out of fashion. For my blog I’d either use a meme or other dump picture for each post. When generated images first came out I used a few for blog posts, it was new and interesting and said “I’m interested in technology and like playing around with new things”.

    Nowadays I’m back on the meme pics. I feel now it’s so much easier to generate images, it more says “I want to look professional but also spend no money and have no standards”.




  • Like Hawke said it seems like the graphics card or driver crashing. Very hard to troubleshoot, especially when it’s random. Bazzite probably already has very recent drivers, there’s this post on the bad website listing some things to try. This stuff can lead to superstitious thinking, with people changing something, rebooting to have it work fine for a while then they post that change as if it fixed it.

    God speed.


  • Your motherboard wouldn’t happen to be an AsRock? There’s been reports of ASRock mobos in particular causing problems with 9000 series AMD chips, especially the X3D. Mate of mine running windows has been having it crash especially when idle at desktop.

    I’m not familiar with a green Linux equivalent to the BSOD. Is it completely green? In that case it may be a graphics problem…


  • My understanding is that it’s technically against their TOS but loosely enforced. They don’t specify precise limits since they probably change over time and region. Once you get noticed, they’ll block your traffic until you pay. Hence you can find people online that have been using it for years no problem, while other folks have been less lucky.

    Basically their business strategy is to offer too-good-to-be-true free services that people start using and relying on, then charging once the bandwidth gets bigger.

    It used to be worse, and all of cloudflare’s services were technically limited to HTML files, but selectively enforced. They’ve since changed and clarified their policy a bit. As far as I’ve ever heard, they don’t give a toss about the legality of your content, unless you’re a neo Nazi.


  • I’m guessing the cloudflared daemon isn’t connecting to jellyfin. You want to use http://. Also is jellyfin the hostname of the VM? Using localhost or 127.0.0.1 might be better ways to specify the same VM without relying on DNS for anything.

    Personal opinion, but I wouldn’t bother with fail2ban, it’s a bit of effort to get it to work with cloudflare tunnel and easy to lock yourself out. Cloudflare’s own zero trust feature would be more secure and only need fiddling around cloudflare’s dashboard.




  • I’ve been using pcloud. They do one time upfront payments for ‘lifetime’ cloud storage. Catch a sale and it’s ~$160/TB. For something long term like backups it seems unbeatable. To the point I sort of don’t expect them to actually last forever, but if they last 2-3 years it’s a decent deal still.

    Use rclone to upload my files, honestly not ideal though since it’s meant for file synchronisation not backups. Also they are dog slow. Downloading my 4TBs takes ~10 days.




  • ZFS doesn’t have fsck because it already does the equivalent during import, reads and scrubs. Since it’s CoW and transaction based, it can rollback to a good state after power loss. So not only does it automatically check and fix things, it’s less likely to have a problem from power loss in the first place. I’ve used it on a home NAS for 10 years, survived many power outages without a UPS. Of course things can go terribly wrong and you end up with an unrecoverable dataset, and a UPS isn’t a bad idea for any computer if you want reliability.

    Totally agree about mainline kernel inclusion, just makes everything easier and ZFS will always be a weird add-on in Linux.