this post was submitted on 04 Feb 2025
78 points (98.8% liked)
PC Gaming
9158 readers
727 users here now
For PC gaming news and discussion. PCGamingWiki
Rules:
- Be Respectful.
- No Spam or Porn.
- No Advertising.
- No Memes.
- No Tech Support.
- No questions about buying/building computers.
- No game suggestions, friend requests, surveys, or begging.
- No Let's Plays, streams, highlight reels/montages, random videos or shorts.
- No off-topic posts/comments, within reason.
- Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I don't have any stats to back this up, but I wouldn't be surprised if failure rates were higher back in the 90s and 2000s.
We have much more sophisticated validation technologies and the benefit of industry, process and operational maturity.
Would be interesting to actually analyze the real world dynamics around this.
Not very many people had a dedicated GPU in the 90s and 2000s. And there's no way the failure rate was higher, not even Limewire could melt down the family PC back then. It sure gave it the college try, but it was usually fixable. The biggest failures, bar none, were HD or media drives.
We all did they used to cost like 60 bucks
I am going to guess the amount made is also much higher than 90s and 2000s since hardware tech is way more popular and used in way more places in the world. So maybe a lower percent but just a high total amount.
But I have no idea..