• 0 Posts
  • 78 Comments
Joined 10 months ago
cake
Cake day: March 1st, 2024

help-circle

  • L2ARC is not a read cache in the conventional sense, but something closer to swap for disks only. It is only effective if your ARC hit rate is really low from memory constraints, although I’m not sure how things stack up now with persistent L2ARC. ZFS does have special allocation devices, though, where metadata and optionally small blocks of data (which HDDs struggle with) can go, but you can lose data if these devices fail. There’s also the SLOG, where sync writes can go. It’s often useful to use something like optane drives for it.

    Personally, I’d just keep separate drives. A lot of caching methods are afterthoughts (bcache is not really maintained as Kent is now working on bcachefs) or, like ZFS, are really complex are not true readback/writeback caches. In particular, LVM cache can, depending on its configuration, lead to data loss if a cache device is lost, and LVM itself can occur some overhead.

    Flash is cheap. A 2TB NVMe drive is now roughly the cost of 2 AAA games (which is sad, really). OP should just buy a new drive.




  • then IBM/Red Hat steal it lol.

    Not really. RH provides all the hosting for the Fedora project, pays multiple people to work on it full time, and on top of that, the RPM specs (which are used to actually build packages) are all MIT licensed. It’d be like complaining bluehat steals the Linux kernel by cloning it from a git repo and making/distributing their own version of it, which is exactly what they do.



  • They didn’t murder centos

    They murdered it, hollowed it out, then re-used the name for something completely new. Granted, what’s new is far from a bad thing, and despite having half the support cycle, the cycle itself is way more consistent and constant because there is no lag time between minor updates (because there are none). Releases are still apparently checked by RH QA, and bug fixes now come a little faster, too.

    Most people have bought into FUD, and spout off the same BS points, and were never centos users to begin with.

    I’ll do you one better: the centos users got exactly what they paid for, and were able to step in at any time to keep centos from turning into centos stream by making their own supported distro. Nobody did until centos original was gone, and were somehow surprised that a distro with a fixed 10 year support cycle takes a nontrivial amount of resources to run. I guess Oracle kind of tried to make their own version of centos with OL before the advent of CentOS Stream, though it was far from being “by the community, for the community”.


  • biribiri11@lemmy.mltoLinux@lemmy.mlDeduplication tool
    link
    fedilink
    arrow-up
    3
    ·
    edit-2
    6 months ago

    Tbf you did start your post with

    I’m in the process of starting a proper backup

    So you’re going to end up with at least a few people talking about how to onboard your existing backups into a proper backup solution (like borg). Your bullet points can certainly probably be organized into a shell script with sync, but why? A proper backup solution with a full backup history is going to be way more useful than dumping all your files into a directory and renaming in case something clobbers. I don’t see the point in doing anything other than tarring your old backups and using borg import-tar (docs). It feels like you’re trying to go from one half-baked, odd backup solution to another, instead of just going with a full, complete solution.


  • biribiri11@lemmy.mltoLinux@lemmy.mlDeduplication tool
    link
    fedilink
    arrow-up
    3
    arrow-down
    1
    ·
    6 months ago

    As said previously, Borg is a full dedplicating incremental archiver complete with compression. You can use relative paths temporarily to build up your backups and a full backup history, then use something like pika to browse the archives to ensure a complete history.


  • I think it’d also be good to document:

    Alpine and NixOS: both 6 months

    Minor releases of RHEL: 6 months

    Non LTS Ubuntu: 6 months

    The question also brings up Fedora rawhide. Fedora rawhide never has releases, though its version is always the current latest branched release (not necessarily stable/beta/alpha) + 1.

    Since the pace of development was also brought up:

    Fedora Rawhide and ELN (same package set) -> Fedora Stable after ~2-3 months of being “stabilized” (this stabilization period has periodic “freezes” which is why bad versions of XZ never made it into Fedora 40’s beta)

    Fedora Rawhide and ELN (same package set) -> CentOS Stream (currently unclear how long it takes to go from branched to full release, though it was branched months ago from ELN) -> RHEL every 6 months

    AlmaLinux releases are tagged from CentOS stream every 6 months, and patched with security updates. When a new version releases, the current minor release is immediately EOL’d, unlike RHEL. Rocky is the same. Both have extra support services from third parties. RHEL offers EUS releases for every other minor release (as of RHEL 9).

    It’s a common misconception that Fedora stable releases become CentOS Stream releases. This pattern was true pre-CentOS stream, but now, for the most part, CentOS Stream and Fedora stable might share a few patches at most, but their development timelines are different. They branch directly off the rolling Fedora Rawhide/ELN trunk.

    Debian unstable -> Debian testing (auto-promoted after 2 weeks iirc) -> Ubuntu stable or Debian stable




  • It’s not about funding. Many prefer mirrors because the main instance isn’t globally available (the GitHub issue I linked, for example, is all about people trying and failing to access flathub in China) or because they can’t for compliance reasons (many businesses already mirror stuff like epel, too, which is what throws off Rocky’s stat counters). Neither of those issues can be assessed by throwing more money at a CDN.




  • Am I wrong in assuming that both OS’s should be sharing the EFI and /boot partitions?

    Shared ESP is fine as long as you don’t run out of space. Nothing in /boot should conflict but that’s not guaranteed, although having 2 potential boot partitions means having 2 potential grub configs. I’d make sda3 a ~2GB ext4 boot partition just for Fedora (mounted at /boot), and an sda5 with btrfs with a home subvolume mounted at /home, and a root subvolume mounted at /, then mount sda1 at /boot/efi (this is the default layout iirc, albeit with different partitions, ofc). This might be easier to do in the advanced blivet gui.

    And yes, Linux’s boot process is a convoluted, fragile mess and there are currently multiple ongoing discussions on how to improve it.


  • So you want to build something like apko (alpine packages/repos, used in chainguard’s images) or rules_oci (used in google’s Debian-based distroless images) but for portage?

    I think it’d be cool. Just keep in mind:

    1. Container scanning tools (like trivy), afaik, tend to look for a package db. Going distroless breaks them. I believe this is why chainguard generates a SBOM (software bill of materials).
    2. Container images are already de-duplicated, and often, the gains in pull times aren’t worth the additional debugging effort (for example, you won’t be able to have dig/curl installed without rebuilding and deploying the whole image, or even a bash prompt in most cases). They’re even more not worth it because lazily pulling OCI images is now a thing, though it’s in its infancy. See: eStargz and I believe dragonfly which uses nydus. More generally though, zstd:chunked will probably eventually become mainstream and default, which will all but eliminate the need for “minimal” starting images.
    3. If you wanted to go really small, there’s stuff like slim which makes tailor made images.
    4. Gentoo, afaik, doesn’t really do LTS releases, making it undesirable for server use, which is the main place containers are.
    5. Distroless containers don’t share common base images because they are normally scratch-built. This breaks image deduplication, leading to increased disk usage instead of decreased disk usage, and why I personally swapped off chainguard’s images.



  • I sometimes hear that it is a different story on servers.

    Wonder what their usages are, especially in a container-focused world, where most containers simply don’t have an init, and the base system just needs, at most, to have a container runtime (+/- a few other things, see: talos linux and their 130MB bare-metal ISOs).