• 1 Post
  • 214 Comments
Joined 1 year ago
cake
Cake day: January 17th, 2024

help-circle
  • Nvidia rightfully earned their bad reputation on linux,

    Really? IMO not with GPUs. They have released linux drivers for decades, and always in time for new kernel versions. ATI was typically way behind and buggy as hell. I would likely not have switched to Linux on the desktop in 2006 if it wasn’t for my GPU “just working”, without any fiddling. Performance was always equal to Windows and stuff like multimonitoring just worked. They even had their nice setup utility to configure Xorg for you.

    Could they have handled the transition to Wayland better? Maybe. But claiming they earned a bad reputation in regards to GPU when they are the one big vendor that had extremely active linux support for ages is dishonest and unwarranted, IMO.




  • My impression of Starfield (after release, at least) was, that it was a bunch of pretty well intended and implemented subsystems (as is, to my knowledge quite common in game development; each team works on a different one), but they just don’t fit really well together. All the subsystems are good parts of a theoretically good overall big picture, but the complexity seemed too high for them to actually flesh out the big picture.

    Technically it all works, but IMO you feel the conceptual gaps whenever you transition (UX wise) from one gameplay mechanic to the next. It just doesn’t (or didn’t) feel like a cohesive game.



  • aksdb@lemmy.worldtoLinux@lemmy.mlVirus
    link
    fedilink
    arrow-up
    2
    ·
    28 days ago

    We recently had a funny problem. Our service ran fine, but a postgres upgrade failed because some pg internals were broken (broken ref ids). Dumping the DB also failed for the same error. Reading and writing was still fine, though. So we restored backup after backup… no dice. They all had the same issue: it was working for the service but we couldn’t perform any maintenance. Ultimately we had to “manually” dump the data of the service and replay it into a fresh db. That took quite long. But that was interesting, since even the verification of the backups didn’t help us notice that kind of corruption.












  • btrfs because it was simple

    Personally I found ZFS far more simple. The userspace tools make more sense to me. Also I like, that volumes can have a default (relative) mount point attached. So in a recovery scenario, I simply have to open the zpool with a relative base path, and then have all my volumes ready to go. If I want to recover a btrfs system with multiple subvolumes, I typically need to know exactly which ones and where to I have to mount them (each individually).

    Also I go really used to zfsbootmenu.




  • I can understand Hellwig’s fear, though.

    From what I gather as a bystander, it’s apparently common that a refactoring in your module that breaks its API will involve fixing all the call-sites to keep the effort on the person responsible for the change. Now the Rust maintainers say “it’s fine; if it breaks, we’ll deal with it” which is theoretically takes away the cross-language issue for the C-maintainer. Practically I can very well see, that this will still cause friction in the future.

    Let’s say such a change happens and at that time there’s a bit of time pressure and the capacity on the rust maintainers is thing for whatever reasons. Will they still happily swallow that change or will they start to discuss if it’s really necessary to do that change? And suddenly, the C-maintainer has a political discussion on top of the technical issue they wanted to solve.

    As someone who just wants to get shit done, I would definitely have that fear.

    (That doesn’t mean it’s still a bullet not worth swallowing. The change overall can still be worth the friction. I am just saying that I think it’s not totally unwarranted that a maintainer feels affected by this even though current pledges from the other parties promise otherwise; this stance can change or at least be challenged over and over.)