this post was submitted on 15 Jul 2025
279 points (93.7% liked)
Technology
73567 readers
3282 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The author's take is detached from reality, filled with hypocrisy and gatekeeping.
It absolutely is nostalgia talking. Yes your TI-99 fires up immediately when plugged in, and its old. However my Commodore 64 of the same era risk being fried because the 5v regulator doesn't age well and when fails dumps higher voltage right into the RAM and CPU. Oh, and c64 machines were never built with overvoltage protection because of cost savings. So don't confuse age with some idea of golden era reliability. RAM ICs were also regularly failed in those age of computers. This is why you had RAM testing programs and socketed ICs. When was the last time, Mr author, you had to replace a failed DIMM in your modern computer?
By that logic, even the TI-99 he's loving on is just a fancier ENIAC or UNIVAC. All technology is built upon the era before it. If there was no technological or production cost improvement, we'd just use the old version. Yes, there is a regular shift in computing philosophy, but this is driving by new technologies and usually computing performance descending to be accessibly at commodity pricing. The Raspberry Pi wasn't a revolutionary fast computer, but it changed the world because it was enough computing power and it was dirt cheap.
I agree, there is something appealing about it to you and me, but most people don't care....and thats okay! To them its a tool to get something done. They are not in love with the tool, nor do they need to be. There were absolutely users of TI-99 and C64 computers in the 80s that didn't give two shits about the shift register ICs or the UART that made the modem work, but they loved that they could get invoices from their loading dock sent electronically instead of a piece of paper carried (and lost!) through multiple hands.
Mr. author, no one is stopping you from using your TI-99 today, but in fact you didn't use it to write your article either. Why is that? Because the TI-99 is a tiny fraction of the function and complexity of a modern computer. Creating something close to a modern computer from discrete components with "part numbers you can look up" would be massively expensive, incredibly slow, and comparatively consume massive amounts of electricity vs today's modern computers.
Ugh, this is frustrating. Do you think a surgeon understands how a CCD electronic camera works that is attached to their laparoscope? Is the surgeon un-educated that they aren't fluent in circuit theory that allows the camera to display the guts of the patient they're operating on? No, of course not. We want that surgeon to keep studying new surgical technics, not trying to use Ohm's Law to calculate the current draw of the device he's using. Mr author, you and I hobby at electronics (and vintage computing) but just because its an interest of ours, doesn't mean it has to be of everyone.
Such gatekeeping! So unless you know the actual engineering principles behind a device you're using, you shouldn't be allowed to use it?
Innovation isn't just creating new features or functionality. In fact, most I'd argue is taking existing features or functions and delivering them for substantially less cost/effort.
As I'm reading this article, I am thinking about a farmer watching Mr. author eat a sandwich made with bread. Does the Mr author know when to till soil or plant seed? How about the amount of irrigation Durum wheat needs during the hot season? How about when to harvest? What moisture level should the resulting harvest have before being taking to market or put in long term storage? Yet there he sits, eating the sandwich blissfully unaware of all the steps and effort needed to just make the wheat that goes into the bread. The farmer sits and wonders if Mr author's next article will be deriding the public on just eating bread and how we've forgotten how to grow wheat. Will Mr Author say we need fewer people ordering sandwiches and more people consulting US GIS maps for rainfall statistics and studying nitrogen fixing techniques for soil health? No, probably not.
Perhaps, but these simple solutions also can frequently only offer simple functionality. Additionally, "the best engineering solutions" are often some of the most expensive. You don't always need the best, and if best is the only option, then that may mean going without, which is worst than a mediocre solution and what we frequently had in the past.
The reason your TI-99 and my c64 don't require constant updates is because they were born before the concept of cybersecurity existed. If you're going to have internet connected devices they its a near requirement to receive updates for security.
If you don't want internet connected devices, you can get those too, but they may be extremely expensive, so pony up the cash and put your money where your mouth is.
It is a machine of extremely limited functionality with a comparably simple design and construction. Don't think even a DEC PDP 11 mainframe sold in the same era was entirely known by a handful of people, and even that is a tiny fraction of functionality of today's cheap commodity PCs.
Take off the rose colored glasses. It was made as a consumer electronics product with the least cost they thought they could get away with and have it still sell. Sales of it absolutely served quarterly revenue numbers even back in the 1980s.
We don't need most of these consumer electronics to last. Proof positive is the computer Mr. author is writing his article on is unlikely to be an Intel based 486 running at 33Mhz from the mid 90s (or a 68030 Mac). If it still works, why isn't he using one? Could it be he wants the new features and functionality like the rest of us? Over-engineering is a thing, and it sounds like what the author is preaching.
Apologies if my post turned into a rant.
I came here to post a screed a bit like this, but you did it so eloquently I don't have to, so thanks! A perfect take, imo.