A while back there was some debate about the Linux kernel dropping support for some very old GPUs. (I can’t remember the exact models, but they were roughly from the late 90’s)

It spurred a lot of discussion on how many years of hardware support is reasonable to expect.

I would like to hear y’alls views on this. What do you think is reasonable?

The fact that some people were mad that their 25 year old GPU wouldn’t be officially supported by the latest Linux kernel seemed pretty silly to me. At that point, the machine is a vintage piece of tech history. Valuable in its own right, and very cool to keep alive, but I don’t think it’s unreasonable for the devs to drop it after two and a half decades.

I think for me, a 10 year minimum seems reasonable.

And obviously, much of this work is for little to no pay, so love and gratitude to all the devs that help keep this incredible community and ecosystem alive!

And don’t forget to Pay for your free software!!!

  • Dariusmiles2123@sh.itjust.works
    link
    fedilink
    arrow-up
    8
    ·
    6 months ago

    10 years is clearly not enough. I’d say 20 years but I clearly don’t know how much work is involved.

    I also clearly think that preserving the history of technology isn’t given enough importance with games disappearing, OS’s being not useable anymore and stuff like this.

    But Linux is clearly the good student here.

    • Count Regal Inkwell
      link
      fedilink
      arrow-up
      8
      ·
      6 months ago

      Linux is absolutely the gold standard when it comes to supporting legacy stuff.

      With Windows trailing behind. At least Microsoft tries to support stuff from older versions of Windows, whereas Apple just says “**** you” every few years.

    • Telorand@reddthat.com
      link
      fedilink
      arrow-up
      3
      arrow-down
      1
      ·
      5 months ago

      Devil’s Advocate: just because something can be preserved doesn’t mean it’s worth preserving. For all the golden games of the 80s and 90s, there were even more turds, and the same goes for other software.

      Really though, the issue comes down to kernel bloat and volunteer support. Imagine having a kernel that’s bigger than Ubuntu, simply because it supports decades of hardware that only a shrinking, tiny minority uses, and an equally shrinking number of people who care to try to patch it so it stays up to date. It’s untenable.

      • Hawke@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        5 months ago

        I think you might have a different understanding of support than most. Nobody’s saying that the code to run this 30-year-old hardware should be enabled by default nor that distros should have them included by default.

        That’s very different from whether the code is in the kernel in case someone wants to compile a custom kernel that does support it. Source code that’s disabled doesn’t add bloat to running systems.