Haha, there’s still things embedded deep in code and in CPUs that go way back to the 80s. If only y’all knew. It’s all shit built on top of older shit, built on top of even older shit with kludges and hacks to glue it all together. Know why Windows has five different ways to access the same setting? Because if they get rid of the older methods, they break a ton of other shit that depends on it too. A house of cards or a Jenga tower.
A modern PC can STILL natively boot a DOS floppy from 1986 in legacy BIOS mode because of this.
Theres also examples in the corporate world where some companies are STILL running 70s mainframes, and use shiny new PCs as front end terminals that just connect to the same old backend.
Seriously, each new windows update just adds a fresh new coat of paint on top, as if to make finding the actually useful win 7 and xp menus, that are still there, harder.
Linux Mint feels to me like what windows 10/11 should’ve been
CPU architecture wise, you can see the difference between cluttered, old x86 and ARM or even RISC-V chips. They just run so much more efficient, as you can tell with your phone lasting a day or two, or apple silicon consuming a fraction for the same performance.
An example for the ancient backend would be the flight pathing system DAL. (Wendover video)
Yeah I can’t even think of any recent CPUs that aren’t based on previous designs. Even Apple’s new M1 is an ARM derivative, which itself is based on an ancient computer from the 80s known as the “Acorn”.
It’s a bit poetic. They were directly competing with Apple at the time, and Acorn named themselves such so that they would appear in front of Apple in the phone book. Of course, they haven’t existed in a long time, but 35-40 years later, Apple decides to use the great-grandson of Acorn’s CPU in their new products.
Haha, there’s still things embedded deep in code and in CPUs that go way back to the 80s. If only y’all knew. It’s all shit built on top of older shit, built on top of even older shit with kludges and hacks to glue it all together. Know why Windows has five different ways to access the same setting? Because if they get rid of the older methods, they break a ton of other shit that depends on it too. A house of cards or a Jenga tower.
A modern PC can STILL natively boot a DOS floppy from 1986 in legacy BIOS mode because of this.
Theres also examples in the corporate world where some companies are STILL running 70s mainframes, and use shiny new PCs as front end terminals that just connect to the same old backend.
ISA still exists on new, modern hardware.
https://www.tomshardware.com/news/isa-slot-tpm-soundblaster-header-pc
The control panel peaked at windows 7 though
everything Windows peaked with Windows 7.
Imma gonna stop you there and say the peak was Windows 2000.
Windows 2000 was definitely peak, for its time.
but Windows 7 eclipsed it, and remains the best microsoft OS… I will fight and die on this hill <3
If we ignore the spontaneous BSOD, Windows 98 had the best performance among all systems lol
Seriously, each new windows update just adds a fresh new coat of paint on top, as if to make finding the actually useful win 7 and xp menus, that are still there, harder.
Linux Mint feels to me like what windows 10/11 should’ve been
This is super interesting and I had NO IDEA! Makes me very curious how much more efficient an entire fresh start might be with new tech.
CPU architecture wise, you can see the difference between cluttered, old x86 and ARM or even RISC-V chips. They just run so much more efficient, as you can tell with your phone lasting a day or two, or apple silicon consuming a fraction for the same performance.
An example for the ancient backend would be the flight pathing system DAL. (Wendover video)
Yeah I can’t even think of any recent CPUs that aren’t based on previous designs. Even Apple’s new M1 is an ARM derivative, which itself is based on an ancient computer from the 80s known as the “Acorn”.
It’s a bit poetic. They were directly competing with Apple at the time, and Acorn named themselves such so that they would appear in front of Apple in the phone book. Of course, they haven’t existed in a long time, but 35-40 years later, Apple decides to use the great-grandson of Acorn’s CPU in their new products.