It’s weird - when I played at launch, I had precisely one bug that impacted my gameplay. Other than that, the game ran pretty smooth and was a joy to play.
Now mind you, I was playing on a PC with a Xeon, 64GB of RAM, and an RTX 2080ti. Nothing ram badly on that system three years ago. Nowadays the older CPU, slower RAM and admittedly older GPU without all the newest bells and whistles (DLSS Framegen I’m looking at you) can’t quite measure up to the latest titles.
Cyberpunk, at launch, was great. For me. Specifically for me. I loved it and still do. But this article hits a point for me that I’ve been struggling to find reason to write about without feeling like I’m ignoring people who primarily play on consoles or can’t afford a nice PC. Regardless…
Man it fuckin’ sucks how you can spend a huge amount of money on a new GPU and then four months later a new one comes out that blows it out of the water. New hardware is so much better and - because all the game devs are using that hardware to design their games both on and for - systems like mine that are still fairly new can’t run the latest games at high settings anymore.
It used to be that if you ponied up the money for a high-end rig, you could expect decent performance for years to come. But I guess blowing a grand on a GPU these days just means you’ll be doing it again in a year or something, instead of the decade or so before.
I’m not saying my PC is bad. Most of what I play runs excellently. But when I spend a grand on just a GPU I expect that GPU to run the newest games at high settings for a long time. Jedi Survivor, Starfield, both run like crap on my system. Never mind the 2TB NVMe drive everything’s installed on.
I had a similar “it’s great I don’t get what people are talking about” experience, only I was running it on lesser hardware than yours: 4GB GTX 970, 12-core i7-5820k, 32GB ram. I ran into a handful of bugs that were funny but not really disruptive (e.g. some dude’s corpse floating behind a car as I was on the highway) and otherwise had a blast.
Nonetheless, it didn’t really feel finished, y’know? That part wore on me, and I think is what undermined my enjoyment the most. It really was released too early.
Ran it on a 980ti, i7 4770k, 32GB ram at release. It certainly struggled at parts, but overall decent experience. And that was a pretty outdated rig at that point.
People just threw a tantrum. There were fewer serious bugs than Skyrim, which got all around glowing reviews. People have claimed the hype was why their expectations were so high, but as someone who wasn’t even planning on playing it for a couple years until it was gifted to me, it was a decent game that had some areas of obvious improvement. Definitely a worthy first attempt at a GTA kind of game, and its a damn shame the gaming community chose it to be the meme pinata for the year.
It’s weird - when I played at launch, I had precisely one bug that impacted my gameplay. Other than that, the game ran pretty smooth and was a joy to play.
Now mind you, I was playing on a PC with a Xeon, 64GB of RAM, and an RTX 2080ti. Nothing ram badly on that system three years ago. Nowadays the older CPU, slower RAM and admittedly older GPU without all the newest bells and whistles (DLSS Framegen I’m looking at you) can’t quite measure up to the latest titles.
Cyberpunk, at launch, was great. For me. Specifically for me. I loved it and still do. But this article hits a point for me that I’ve been struggling to find reason to write about without feeling like I’m ignoring people who primarily play on consoles or can’t afford a nice PC. Regardless…
Man it fuckin’ sucks how you can spend a huge amount of money on a new GPU and then four months later a new one comes out that blows it out of the water. New hardware is so much better and - because all the game devs are using that hardware to design their games both on and for - systems like mine that are still fairly new can’t run the latest games at high settings anymore.
It used to be that if you ponied up the money for a high-end rig, you could expect decent performance for years to come. But I guess blowing a grand on a GPU these days just means you’ll be doing it again in a year or something, instead of the decade or so before.
I’m not saying my PC is bad. Most of what I play runs excellently. But when I spend a grand on just a GPU I expect that GPU to run the newest games at high settings for a long time. Jedi Survivor, Starfield, both run like crap on my system. Never mind the 2TB NVMe drive everything’s installed on.
But I’m just bitching to bitch. Ignore me.
I had a similar “it’s great I don’t get what people are talking about” experience, only I was running it on lesser hardware than yours: 4GB GTX 970, 12-core i7-5820k, 32GB ram. I ran into a handful of bugs that were funny but not really disruptive (e.g. some dude’s corpse floating behind a car as I was on the highway) and otherwise had a blast.
Nonetheless, it didn’t really feel finished, y’know? That part wore on me, and I think is what undermined my enjoyment the most. It really was released too early.
Ran it on a 980ti, i7 4770k, 32GB ram at release. It certainly struggled at parts, but overall decent experience. And that was a pretty outdated rig at that point.
People just threw a tantrum. There were fewer serious bugs than Skyrim, which got all around glowing reviews. People have claimed the hype was why their expectations were so high, but as someone who wasn’t even planning on playing it for a couple years until it was gifted to me, it was a decent game that had some areas of obvious improvement. Definitely a worthy first attempt at a GTA kind of game, and its a damn shame the gaming community chose it to be the meme pinata for the year.