This article needs a clearer title. I agree that upgrading from a 6000 or 3000 series card right now is almost completely pointless, and even going back another generation it's still not a great proposition. But I know people with "gaming PCs" rocking 1650s or even 1050s. Lots of folks with medium or low end several generations old hardware out there, for whom great upgrade options exist.
I still use a 1070, so the GPU comparisons here aren't relevant.
The main issue I hit was deciding between DDR4 and DDR5 RAM since we're in an awkward transition phase - and that affects motherboard and so CPU choices too.
Well, I've had the same CPU/Mobo/RAM for over ten years and only upgraded my GPU once from a GTX660 to a 5700xt at the start of the pandemic. I'm finally seeing some issues with some modern AAA content. Hogwarts legacy won't really run at all, for example.
I also haven't wiped my system in the same amount of time, so that may be more the culprit than the system itself. Still going strong!
The CPU becomes the real issue though - which then means changing motherboard, which means changing RAM, etc. and then you might as well get an NVMe too etc.
I used to upgrade every generation, and yeah, it was stupidly expensive. But it was my only hobby, and you could actually seen performance increases each time.
But for the last 10 years or so, there's much less point. Sometimes there are major advances (Cuda, RTX) that make it worthwhile for a single generation upgrade, but mostly it's just a few FPS at highest settings. So now I just upgrade every few years.
Back in the 90s and early 00s, frequent upgrades were kind of required to stay up to date with new games. The last 10-15 years have been muuuuch slower in that regard, thanks to consoles I guess. I'm not complaining, but I miss the sense of developers really pushing boundaries like they did in the old days.
The only reason I upgraded my 10 series to a 30 series is because I'm a dummy and bought a monitor with only HDMI 2.1 and no display port, so I needed to upgrade my GPU or I would have no gsync. Otherwise, I probably would have waited at least 2 more generations to upgrade.
Well, like some, I am still on the 10xx series (1060 3gb 🤣🥲) and starting to look to the futur full system upgrade for a Rx79xx or 78xx when out.
Targeting Black Friday sale jump
I would be curious to know if many others are on a refresh cycle up to 4-5years
I have recently ordered a 7900xtx to replace my 1080ti. It was a good companion but just doesn't cut it anymore.
Originally wanted to upgrade with the last gen but scalpers made that impossible. And the used market is still fucked in my region.
I upgraded last summer to a 6700XT from a 1070ti. I didn't need the upgrade, since the 1070ti is still a solid performer even now. There's not much that it wouldn't still run now and reasonable settings. I really only upgraded because there was a decent sale, and I had some money burning a hole in my pocket. I could have easily waited another year, and gone with a 6800XT or better for a similar price.
I guess I'm on the 7-10 year cycle. I just upgraded from a GTX 1050 (non-Ti) and .an i5-4570. Played almost everything I wanted to play just fine at 1080p and some at 1440p. I tend to be a patient gamer and play mostly indies, so it was great.
This article feels like it was written in a language I don't understand. I understand that other people are more into the hobby than I am (which is fine, no judgement, good for them), but its just so far outside what I would consider normal for me that it took me off guard.
Exactly. I'm on a 1050Ti and I'm not sure Starfield will be happy with that. Cyberpunk wasn't too happy. And of course if I get a new card I will need new MB/CPU/RAM/etc.
1070ti here. I think the fact that the needle hasn't moved significantly forward, as the article puts it, has decreased prices to the point where an upgrade to a more updated setup makes sense now for me personally.
I agree with the article if we're talking an upgrade from 30' series gpus, but things seem great for all other cases.
The starting premise of the article is based on upgrading from a previous generation. Which sane mind does that? Aside from the one time I got a freebie, all my upgrades were at least two generations apart.
Edit: Also, coming with certain prices on the RX 6000 series, as long as you're from three generations behind you'll get a good upgrade. I went from a GTX 1070 to a RX 6700 XT. Felt a big improvement there.
They do make the point in the article that even upgrading from two generations back is a waste, as you're getting basically getting no real benefit to having waited two generations instead of one. You may as well upgrade to last generation instead of this one and save yourself some money.
If you're three generations behind, no matter what your upgrade path is, you're getting a significant upgrade, but it's still not worth upgrading to the current gen when last gen is much better value for a marginal performance difference.
The exception to all this is buying the absolute top-of-the-line, which is never good value, but is again significantly inflated in price from the previous gen.
I'm halfway through a 10 year cycle, with a 1060 3Gb on a 7th Gen i5. It's mostly Civ6, Stellaris, and Rocksmith 2014 @ 1080p so it's fine. The main problem is end-of-life for Windows 10 without support in the current hardware, and Rocksmith doesn't work well on Linux. I'll probably keep it as-is and start from scratch... when I see a title that I want to play enough to drop big cash on hardware.