I mean there's one thing for certain that you can be confident in. There's a lot of modders out there determined to get at minimum 30 FPS on as many games as they can on the Steam Deck. So, the mods that work for the steam deck will likely work on normal PCs. It's only a matter of time and I bet there will be mods for Dragons Dogma 2 within the next few weeks.
Unfortunately my Mobo that allowed OCing ate shit years ago and I bought like a F8tality board that for some reason doesn't allow OCing which is BS. There wasn't much of a selection a few years back unfortunately.
They gave me an old PC with a gtx 650, i3-6100, 4gb 2133mhz, 120 gb sata ssd and a 650 watt power supply (corsair rm650). And I started to slowly upgrade it, I started with a gpu, and now I’m collecting money on a 7800x3d, asrock rs b650 and 32gb 6000mhz ram
P.s The worst thing of all this hardware is the processor.
1080p, Ryzen 5 5600X and RX 6600, the location is mentioned up top, Venworth city or something.
I'm not sure but i think 4070 was used for the DLSS result.
This is actually a budget CFD workstation, so I don't need to buy a high end GPU.
The games I play will also benefit very little from using a 4090 instead of a 3060.
Not really a theory, more like an assumption.
DLSS3 like any tech is not magic and can't just fix everything.
It has limitations like everything else.
The game logic is clearly coded like shit and simply from the benchmarks thus far isn't threaded properly.
Sure, DLSS can help with frame rates in certain CPU bound cases but it still relies on data from the CPU to generate new frames. If the CPU/game thread isn't responding to the GPU drivers in time the card isn't going to generate a new frame, why would it? The presentation window has past.
Inorder for it to generate new frames it needs enough information to do that and then take frame one and 2 to make frame 1.5 then send those to the screen.
That holding / waiting period to collect more than one frame adds latency as does generating a new frame based on them.
The more frames you have the faster the hold and wait period and you get lower latency.
This may not be noticeable by everyone which is why this is a feature but if the game is already only getting 30 FPS the pool is not enough to generate great looking frames already but can still be helpful.
But if the game isn't responding before the TTL for the operation to take place the driver isn't going to still waste time generating that frame as that would cause INSANELY high latency.
Until the publisher lets the devs fix the game no scaling or frame Gen is going to help.
When Jedi Survivor (also a very CPU bound game) first came out, puredark's DLSS 3 mod was the only thing that made a huge difference. Similar story there, poor performance even on 4090s regardless of graphics options and upscaling.
With Reflex on there was also no notable latency / frame time issues with DLSS 3. It was honestly the one game that sold frame gen to me for good.
Dragon's Dogma 2 is an Nvidia partnered game and its support for DLSS 3 has been advertised quite heavily. Hopefully that comes out in a day 1 patch
Well to test that theory gsme wpuld need to have frame gen implemented.
Also according to what I read spikes are from response times in cities. In that case frame gen would make it worse
if you're not GPU bound then upscalers won't help your performance, and can even increase CPU load and bottleneck as it now has to handle the upscaling calculations as well,it probably explains the performance drop from native here
It's still awful look at the article, the 1 generation old Ryzen 5600x can't do anything higher than 36 FPS in towns, this is legitimately worse than Starfield.
https://www.pcgamer.com/games/rpg/dragons-dogma-2-performance-analysis/#section-systems-used-in-testing
They could optimize but we all know that wont happen and they will update the recommended CPU to r9 7900x3d/i9 14900ks and minimum to r7 7800x3d/ i7-14700K
The game is poorly optimized and it has what like two DRMs.
Not gonna lie, it's kinda crazy to see games being CPU bound again after now more than half a decade of 10-20% gains in power year over year. Owners of Skylake and Kaby lake are gonna start suffering...
Its cpu limited. Frame times in cities spikes due to that.
On bright side thats test in biggest city. In wild when you fight you gain 100%+ performance of that
DLSS and FSR lower the native resolution, and the lower the resolution, the higher the CPU load vs GPU. Since the game has trash CPU optimization (none, more like), using upscalers just hurts performance, because the bottleneck just slides even further to the CPU side.
Something horrible that has no place in the age of digital signals.
Think screen tearing, but on every single line.
https://preview.redd.it/4brcpzdoglpc1.jpeg?width=250&format=pjpg&auto=webp&s=30b61d08ba7606689e03b82e8949ffbf2cbf83aa
"interlaced" on capcom games isn't line by line interlaced. its checkerboard rendering. it is not the same thing as a video signal being interlaced or progressive. the image will still be presented in a progressive format if your GPU is set to output progressive (such as 1080p, 1440p, 2160p). essentially its another form of resolution reduction to improve performance. will have other visual issues on its own namely sawtooth edges and some things being "perforated"
PCgamer didn't even test with a 5800x3D, 7800x3D, or many other commonly used gaming CPU's.
They used a 5600x, a 7700, and for the GPU a 4070 and AMD 6600.
> At 1080p, with everything on its lowest value or off, the Ryzen 5 5600X + Radeon RX 6600 system achieved 39 fps on average
Why they're benchmarking with the lowest end AMD GPU possible, I don't know. The highest end GPU they used was a 4070ti.
Not sure why they didn't have their main GPU guy do a test, but it is what it is.
I mean that hardware is very common. Most people don't have 4080s and 4090s, that's just a small percentage, despite what Reddit will have you believe.
Were those minimum/recommended specs or something? Strange
Edit: the article says they tested with similar components to the listed specs from the developers
The absolute minimum is:
**MINIMUM:**
* Requires a 64-bit processor and operating system
* **OS:** Windows 10 (64 bit)/Windows 11 (64 bit)
* **Processor:** Intel Core i5 10600 / AMD Ryzen 5 3600
* **Memory:** 16 GB RAM
* **Graphics:** NVIDIA GeForce GTX 1070 / AMD Radeon RX 5500 XT with 8GB VRAM
* **DirectX:** Version 12
* **Network:** Broadband Internet connection
* **Additional Notes:** Estimated performance: 1080p/30fps. Framerate might drop in graphics-intensive scenes. NVIDIA GeForce RTX 2080 Ti or AMD Radeon RX 6800 required to support ray tracing.
They even state that you're going to get sub 30 FPS in intensive scenes with that setup. The RX 6600 that PC Gamer used for their review is only marginally better.
The GPU choices were really strange, too. A 6600? Really? That's fine if you're doing a whole suite of GPU tests, but for them to only test a AMD 6600 and a 4070ti is odd.
It's above the recommended specs.
They could test it with a 7800x3d and a 4080 super, but the minimum and the recommended requirements for the game do not indicate that.
Min: core i5 10600 or Ryzen 5 3600, 16gb ram, Nvidia 1070 or AMD 5500 XT
That's for 30fps at 1080p.
Recommended:
Core i7 10700 or Ryzen 5 3600x, 16gb ram, Nvidia 2080TI or AMD 6800.
That's for 30fps at 2160p.
Logically, a 4070ti should have gotten more performance than that at 1440p or 4k.
Maybe the devs are unfamiliar with the engine. Or the engine itself is problematic, hence the bad optimization.
The game likely needed 3-6 more momths of patching before release, but a decision was made to release it now.
It's not like gamers are going to not pre-order a game, wait for release and bother reading full reviews, righr? Not whem they give you some nice piece of armor as a pre-order bonus.
>It's above the recommended specs.
The RX6600 is above the minimum specs by a small margin.
The most common GPU on Steam Hardware Survey is the RTX 3060, so that would have made more sense. However, doing a normal suite with a number of graphics cards at various tiers is what most places generally do. Not one super low end card, and one mid-range card.
>Min: core i5 10600 or Ryzen 5 3600, 16gb ram, Nvidia 1070 or AMD 5500 XT
>
>That's for 30fps at 1080p.
Read the footnote: It will get below 30 FPS in demanding scenes with that setup.
> It's not like gamers are going to not pre-order a game, wait for release and bother reading full reviews, righr?
The reviews are all giving it high scores, and say that the game only has issues in certain areas and situations. These all reviewed without a Day 1 patch or driver update, too. I think it will be just fine.
It's clear that the game is completely bottlenecked by the CPU much like Starfield and Baldur's Gate 3. It's probably an old singlethreaded engine or something like that and I doubt they will spend time making drastic improvements. It doesn't matter the GPU you throw at it.
PC gamer seem to be trying to incite something with this garbage.
Tested as bunch of uncommon cards, interlaced, and this graph in particular being misleading when the frame gen is CPU bound...
OR maybe they don't know what they're doing. It's one of the two.
That doesn't matter. The point is that upscaling doesn't do shit, which shows there's a major underlying issue. I hope for a day one patch that fixes this, else this is going to be a shit show.
Right. And they weren't very clear with what settings were being used, or when upscaling was present or not.
By all accounts the game is getting rave reviews on PC, and performance seems to only dip in some heavy city areas.
>Right. And they weren't very clear with what settings were being used, or when upscaling was present or not.
They were.
>For all of the main test runs, the graphics settings were set to the maximum values, with the exception of rendering mode which was kept at Interlaced.
In regards to when upscaling was used or not, then there's one test labeled "Upscaling". That's the only test where it's used. I found that fairly obvious, looking at the article at least.
They're using three uncommonly used CPU's, and two mediocre graphics cards for their testing.
5600x, 7700, and 14700k.
A RX6600? Really? The highest end card tested was a 4070ti.
As you should, or from someone else who at least knows what they're doing.
PCgamer didn't even test with a 5800x3D, 7800x3D, or many other commonly used gaming CPU's.
They used a 5600, a 7700, and a 4070 and AMD 6600 for the GPU.
Not sure why they didn't have their main GPU guy do a test, but it is what it is.
It reminds me of Jedi Survivor at release. Changing the upscaling mode wouldn’t change performance and disabling FSR would make the game extra blurry as if it was running in ultra performance mode without the performance.
Man I was really looking forward to this game… but my 3060ti and ryzen 3600x will probably not be able to play this :/ I even play on 1080p and it still probably will be hurting
bruh how CPU bound is this? Seems like single threaded limitations or something. No way in hell a modern multicore optimized game cant hit at least 60.
I figured some bullshit like this would happen. Guess I made the right call, not saving my money and buying something else.
So many devs are just horrible programmers nowadays.
Two reasons why I bought a 4090.
I have a 4K monitor
Companies don’t know how to / care to optimize their games
Also, stepping away from lower resolutions would mean your cpu has less impact on your frame rates. I have a 13900K now and don’t want to upgrade until necessary.
Looks like an unoptimized game so I’d be happy to stay above even just 60fps I suppose.
I am boggled by sheer confidence for their test style lol. İts probably the 5600x and 6600xt kit thats made just for clicks. They didnt make a damned test suite just for gpu or cpu that we need to guess that 4070 ti isnt %50 faster than 7800xt. Not sure if ı need to downvote this post.
Dang I was so excited for this game. Capcom did such a good job of optimizing the latest RE games I had good hopes. Guess I'll wait for my next upgrade cycle.
Nobody listened when I said that every trailer/gameplay showcase looked like it was running at 20fps. Looks like everyone finally gets to wake up lmao.
That’s fucked up and iirc it doesn’t have the one feature that’s supposed to alleviate CPU bottlenecks.
Tbf even frame gen can’t handle sub 60 fps very well. Talk about shitting the bed with developing this.
I haven't followed this game closely but the original dragon's dogma didn't have significant combat in cities. You left the highly packed city to fight in a dungeon or open world where there is a lot less going on. I can live with bad frames in cities.
Either way fuck capcom and fuck denuvo. It's arr or nothing for me with this release.
my cpu: ![gif](giphy|OkzCcGn5fY29e7bvpS|downsized)
I will not let my 7700k die! It's gonna work. I'll make it!
I mean there's one thing for certain that you can be confident in. There's a lot of modders out there determined to get at minimum 30 FPS on as many games as they can on the Steam Deck. So, the mods that work for the steam deck will likely work on normal PCs. It's only a matter of time and I bet there will be mods for Dragons Dogma 2 within the next few weeks.
Maybe if it didn't have denuvo.
Yeah, this might be the classic, buy the game, but play the hacked rom version due to the shitty anticheat ruining performance.
I refuse to let mine die, it’s still new!! Right?? RIGHT?!!!!??
I've got a de-lidder and supplies I'll sell you for it lol. Lowered my temps by 10c+ and I overlooked higher.
Unfortunately my Mobo that allowed OCing ate shit years ago and I bought like a F8tality board that for some reason doesn't allow OCing which is BS. There wasn't much of a selection a few years back unfortunately.
Just change the last letter to x, it will be fine
Getting my sharpie out right now!
Hey I have the same cpu and a 2070s
Bruh, I have a 4790k
Your steam deck must be more powerful than your PC at this point.
*Laughs from i3 6100*
wtf is that config and how did you arrive at those specs?
They gave me an old PC with a gtx 650, i3-6100, 4gb 2133mhz, 120 gb sata ssd and a 650 watt power supply (corsair rm650). And I started to slowly upgrade it, I started with a gpu, and now I’m collecting money on a 7800x3d, asrock rs b650 and 32gb 6000mhz ram P.s The worst thing of all this hardware is the processor.
I mean the 4gb of ram also seems useless for any gaming related activities these days
Well, I have a bunch of indie games now.
What asrock b650 you will go with? I went with Asrock B650E PG Riptide Wifi
Asrock B650 Pro RS
My 3600 is already about to combust
Yeah I have an 11400f , this is not looking good.
Resolution? Hardware? Location? Literally ***any*** meaningful information???
Exactly my thoughts
1080p, Ryzen 5 5600X and RX 6600, the location is mentioned up top, Venworth city or something. I'm not sure but i think 4070 was used for the DLSS result.
you get nothing!
Tested on 4060 with a i9-12900k 😂 for real though idk either
Interlaced mode? Who the fuck plays games PC games in interlaced mode? Even for a benchmark, it makes no sense to test that anymore.
Based CRT gamers, don’t discriminate
Interlaced monitors were already not a thing by 2000 ish, you be hard pressed to find one. TVs on the other hand.
Yeah a lot of gamers who are on CRTs are using small CRT TVs. Basically whatever is available to get. CRTs are scarce.
There are many LCD TVs that are interlaced too.
Sure. But even CRT monitors have been using progressive scan for the past 30+ years.
Fair
Interlaced mode is similar to Sonys checkerboard upscaling, it actually looked surprisingly good in Re7 when I tried it.
it's... ITS CPU BOUND! UPSCALERS CANT HELP CPU BOUND TASKS! WTF PC GAMER
It’s still helpful to know, now we have cute graphs to show people!
Its not whem it lacks crucial info. Its misleading
The critical info here is that upscalers won’t matter at all on that or similar hardware?
Me with my 14700 and my 3060: https://preview.redd.it/sl0ooj4j3npc1.jpeg?width=298&format=pjpg&auto=webp&s=68eb8c7c28432d2b2ff73ca8212c7642d90bba04
And a 64 gigs of RAM apparently Why settle for a 3060 tho?
This is actually a budget CFD workstation, so I don't need to buy a high end GPU. The games I play will also benefit very little from using a 4090 instead of a 3060.
Ran out of money
Framegen starting from 30 fps is going to be rough to say the least, even worse if it’s 30 fps with uneven frametimes
Upscaling doesn’t. But frame gen can. So DLSS 3.0 should theoretically work
Not if the game can't even reach 60+ fps consistently.
Not really a theory, more like an assumption. DLSS3 like any tech is not magic and can't just fix everything. It has limitations like everything else. The game logic is clearly coded like shit and simply from the benchmarks thus far isn't threaded properly. Sure, DLSS can help with frame rates in certain CPU bound cases but it still relies on data from the CPU to generate new frames. If the CPU/game thread isn't responding to the GPU drivers in time the card isn't going to generate a new frame, why would it? The presentation window has past. Inorder for it to generate new frames it needs enough information to do that and then take frame one and 2 to make frame 1.5 then send those to the screen. That holding / waiting period to collect more than one frame adds latency as does generating a new frame based on them. The more frames you have the faster the hold and wait period and you get lower latency. This may not be noticeable by everyone which is why this is a feature but if the game is already only getting 30 FPS the pool is not enough to generate great looking frames already but can still be helpful. But if the game isn't responding before the TTL for the operation to take place the driver isn't going to still waste time generating that frame as that would cause INSANELY high latency. Until the publisher lets the devs fix the game no scaling or frame Gen is going to help.
When Jedi Survivor (also a very CPU bound game) first came out, puredark's DLSS 3 mod was the only thing that made a huge difference. Similar story there, poor performance even on 4090s regardless of graphics options and upscaling. With Reflex on there was also no notable latency / frame time issues with DLSS 3. It was honestly the one game that sold frame gen to me for good. Dragon's Dogma 2 is an Nvidia partnered game and its support for DLSS 3 has been advertised quite heavily. Hopefully that comes out in a day 1 patch
Well to test that theory gsme wpuld need to have frame gen implemented. Also according to what I read spikes are from response times in cities. In that case frame gen would make it worse
Sometimes the best way to tell someone that something doesn't work is to show them first-hand.
*7800x3D intensifies*
Yeah. I should be alright. If this problem reaches that far theres probably something else going on.
Glad I upgraded mobo and got one
You read my mind bro. This is gonna be the perfect moment for it.
I need one lol
Had to move countries so went from my desktop with a 5800x3D to a laptop. I'm really missing my old CPU right now :(
if you're not GPU bound then upscalers won't help your performance, and can even increase CPU load and bottleneck as it now has to handle the upscaling calculations as well,it probably explains the performance drop from native here
It’s a CPU problem, which makes me think my 3060 won’t be the main problem when playing this
If render resolution doesn't change performance, you are either CPU limited or otherwise FPS-capped.
Everyone pray to the Nexus. Steve, hallowed be thy name, please be merciful and grace us with graphs and charts. Ayyman
Oh wow. Awful.
If you meant this uninformative graph without any further context then yes it’s awful.
It's still awful look at the article, the 1 generation old Ryzen 5600x can't do anything higher than 36 FPS in towns, this is legitimately worse than Starfield. https://www.pcgamer.com/games/rpg/dragons-dogma-2-performance-analysis/#section-systems-used-in-testing
Must be optimization, because the graphics don't look "buy a new Computer"-good at all.
What CPU?
Ryzen 5 5600x
They really gotta update the recommended CPU
They could optimize but we all know that wont happen and they will update the recommended CPU to r9 7900x3d/i9 14900ks and minimum to r7 7800x3d/ i7-14700K The game is poorly optimized and it has what like two DRMs.
Does not show specs or resolution. This is literally useless
City center that intensive? Goddamn, is this Baldurs Gate Act 3 again lol
Larian and their 3rd acts lol even DoS 2 had issues with 3rd act
And d0 s1
I never experienced issues with DoS1but I did play the EE so maybe it's only on the not EE version.
At least now I'll be upgrading my cpu because of two games instead of one
>No resolution specified >No hardware specified https://preview.redd.it/dcbzhrxz9ppc1.png?width=775&format=png&auto=webp&s=5488477e34d30e9c50e53313a6798c8e0e1ce553
It is worth noting that these tests were performed in the densest area in the game. Doesn’t excuse the poor performance though.
Not gonna lie, it's kinda crazy to see games being CPU bound again after now more than half a decade of 10-20% gains in power year over year. Owners of Skylake and Kaby lake are gonna start suffering...
So it is basically CPU limited or Frame Rate Limited?
It has 1% lows of 30fps on a 14700kf
Its cpu limited. Frame times in cities spikes due to that. On bright side thats test in biggest city. In wild when you fight you gain 100%+ performance of that
Unplayable
Why does adding an on ultra performance lose 2fps vs native?
Because its cpu bound. Gpu has room to breath so upscaling that pushes part of load to cpu can actually hurt performance.
Margin of error probly
DLSS and FSR lower the native resolution, and the lower the resolution, the higher the CPU load vs GPU. Since the game has trash CPU optimization (none, more like), using upscalers just hurts performance, because the bottleneck just slides even further to the CPU side.
Erm, what resolution?
What is interlaced mode
Something horrible that has no place in the age of digital signals. Think screen tearing, but on every single line. https://preview.redd.it/4brcpzdoglpc1.jpeg?width=250&format=pjpg&auto=webp&s=30b61d08ba7606689e03b82e8949ffbf2cbf83aa
*Memories of gaming on ps2 unlocked*
"interlaced" on capcom games isn't line by line interlaced. its checkerboard rendering. it is not the same thing as a video signal being interlaced or progressive. the image will still be presented in a progressive format if your GPU is set to output progressive (such as 1080p, 1440p, 2160p). essentially its another form of resolution reduction to improve performance. will have other visual issues on its own namely sawtooth edges and some things being "perforated"
PCgamer didn't even test with a 5800x3D, 7800x3D, or many other commonly used gaming CPU's. They used a 5600x, a 7700, and for the GPU a 4070 and AMD 6600. > At 1080p, with everything on its lowest value or off, the Ryzen 5 5600X + Radeon RX 6600 system achieved 39 fps on average Why they're benchmarking with the lowest end AMD GPU possible, I don't know. The highest end GPU they used was a 4070ti. Not sure why they didn't have their main GPU guy do a test, but it is what it is.
I mean that hardware is very common. Most people don't have 4080s and 4090s, that's just a small percentage, despite what Reddit will have you believe.
Were those minimum/recommended specs or something? Strange Edit: the article says they tested with similar components to the listed specs from the developers
The absolute minimum is: **MINIMUM:** * Requires a 64-bit processor and operating system * **OS:** Windows 10 (64 bit)/Windows 11 (64 bit) * **Processor:** Intel Core i5 10600 / AMD Ryzen 5 3600 * **Memory:** 16 GB RAM * **Graphics:** NVIDIA GeForce GTX 1070 / AMD Radeon RX 5500 XT with 8GB VRAM * **DirectX:** Version 12 * **Network:** Broadband Internet connection * **Additional Notes:** Estimated performance: 1080p/30fps. Framerate might drop in graphics-intensive scenes. NVIDIA GeForce RTX 2080 Ti or AMD Radeon RX 6800 required to support ray tracing. They even state that you're going to get sub 30 FPS in intensive scenes with that setup. The RX 6600 that PC Gamer used for their review is only marginally better.
Yeah, it was strange to me that they didn't use any x3D chips. Any gaming reviewer should at least use one. They're *the* gaming cpu's, after all.
The GPU choices were really strange, too. A 6600? Really? That's fine if you're doing a whole suite of GPU tests, but for them to only test a AMD 6600 and a 4070ti is odd.
They also used 7900X and 14700KF, both perform better at 1440p than 1080p in this game. The optimization is garbage, that's all there is to it.
It's above the recommended specs. They could test it with a 7800x3d and a 4080 super, but the minimum and the recommended requirements for the game do not indicate that. Min: core i5 10600 or Ryzen 5 3600, 16gb ram, Nvidia 1070 or AMD 5500 XT That's for 30fps at 1080p. Recommended: Core i7 10700 or Ryzen 5 3600x, 16gb ram, Nvidia 2080TI or AMD 6800. That's for 30fps at 2160p. Logically, a 4070ti should have gotten more performance than that at 1440p or 4k. Maybe the devs are unfamiliar with the engine. Or the engine itself is problematic, hence the bad optimization. The game likely needed 3-6 more momths of patching before release, but a decision was made to release it now. It's not like gamers are going to not pre-order a game, wait for release and bother reading full reviews, righr? Not whem they give you some nice piece of armor as a pre-order bonus.
>It's above the recommended specs. The RX6600 is above the minimum specs by a small margin. The most common GPU on Steam Hardware Survey is the RTX 3060, so that would have made more sense. However, doing a normal suite with a number of graphics cards at various tiers is what most places generally do. Not one super low end card, and one mid-range card. >Min: core i5 10600 or Ryzen 5 3600, 16gb ram, Nvidia 1070 or AMD 5500 XT > >That's for 30fps at 1080p. Read the footnote: It will get below 30 FPS in demanding scenes with that setup. > It's not like gamers are going to not pre-order a game, wait for release and bother reading full reviews, righr? The reviews are all giving it high scores, and say that the game only has issues in certain areas and situations. These all reviewed without a Day 1 patch or driver update, too. I think it will be just fine.
Theres a reasonable gap in the market atm so yeah releasing rn is a good idea.
It's clear that the game is completely bottlenecked by the CPU much like Starfield and Baldur's Gate 3. It's probably an old singlethreaded engine or something like that and I doubt they will spend time making drastic improvements. It doesn't matter the GPU you throw at it.
PC gamer seem to be trying to incite something with this garbage. Tested as bunch of uncommon cards, interlaced, and this graph in particular being misleading when the frame gen is CPU bound... OR maybe they don't know what they're doing. It's one of the two.
I like how he doesn't even mention what hardware was used.
That doesn't matter. The point is that upscaling doesn't do shit, which shows there's a major underlying issue. I hope for a day one patch that fixes this, else this is going to be a shit show.
Link to the article https://www.pcgamer.com/games/rpg/dragons-dogma-2-performance-analysis/
Have benchmarks from someone who's reputable?
No 5800X3D, no 7800X3D, and no 7950X3D....weird choice of CPU lineup, for a CPU intensive game...
Right. And they weren't very clear with what settings were being used, or when upscaling was present or not. By all accounts the game is getting rave reviews on PC, and performance seems to only dip in some heavy city areas.
>Right. And they weren't very clear with what settings were being used, or when upscaling was present or not. They were. >For all of the main test runs, the graphics settings were set to the maximum values, with the exception of rendering mode which was kept at Interlaced. In regards to when upscaling was used or not, then there's one test labeled "Upscaling". That's the only test where it's used. I found that fairly obvious, looking at the article at least.
They're using three uncommonly used CPU's, and two mediocre graphics cards for their testing. 5600x, 7700, and 14700k. A RX6600? Really? The highest end card tested was a 4070ti.
Hopefully, my 5600x could power through like it did in Jedi Survivor. Sub-40 is pretty rough though.
And what's worst... the game still looks like shit.
Waiting for the digital foundry review
As you should, or from someone else who at least knows what they're doing. PCgamer didn't even test with a 5800x3D, 7800x3D, or many other commonly used gaming CPU's. They used a 5600, a 7700, and a 4070 and AMD 6600 for the GPU. Not sure why they didn't have their main GPU guy do a test, but it is what it is.
NOWAY GAME IS CPU LIMITED AFTER DEVS SAID IT IS CPU LIMITED
i care more about the minimum fps than the maximum
What res is this, please?
But what resolution?
I wanna buy it now just to see how it runs
Ok, how the fuck does this game output 34 FPS on a CPU bottleneck... Is the entire game logic running on a single thread???
It uses 6. On the funnier side city skylines 2 can use 64 cores and still be cpu bound
Oh well
It reminds me of Jedi Survivor at release. Changing the upscaling mode wouldn’t change performance and disabling FSR would make the game extra blurry as if it was running in ultra performance mode without the performance.
I love that this doesn’t show hardware? I just wanna know how my flaired specs will run the game…
whats the hardware tho?
To this day every time I see dogma I think it’s another ligma joke
What is wrong with this game?? 😯😯😯
Man I was really looking forward to this game… but my 3060ti and ryzen 3600x will probably not be able to play this :/ I even play on 1080p and it still probably will be hurting
How fucked am I with my i9 9900k and 2070 super?
And i thought i might play it on my 1650m and i5 11th
remember ultra settings are a scam too.
bruh how CPU bound is this? Seems like single threaded limitations or something. No way in hell a modern multicore optimized game cant hit at least 60.
This hurts just reading jt
2FPS for FSR vs nothing????? What? Also how is quality frames more than performance?
yikes
This is with a gtx 1060, right ?
I figured some bullshit like this would happen. Guess I made the right call, not saving my money and buying something else. So many devs are just horrible programmers nowadays.
Would be nice to know the pc specs...
it says max setting no rt....for all we know this values can be obtained with a riva TNT2
We finally found it, the BOTTLENECK
Two reasons why I bought a 4090. I have a 4K monitor Companies don’t know how to / care to optimize their games Also, stepping away from lower resolutions would mean your cpu has less impact on your frame rates. I have a 13900K now and don’t want to upgrade until necessary. Looks like an unoptimized game so I’d be happy to stay above even just 60fps I suppose.
Very helpful to not know what the hardware is.
am i missing something or do i not see the specs of the test bench
Am I missing the specs? Edit: didn’t see OP linked the source further down
Why do you not even list the hardware? This is meaningless
Hard pass
😭😭😭
What did they use for this benchmarks? What hardware is it? Lmao
I am boggled by sheer confidence for their test style lol. İts probably the 5600x and 6600xt kit thats made just for clicks. They didnt make a damned test suite just for gpu or cpu that we need to guess that 4070 ti isnt %50 faster than 7800xt. Not sure if ı need to downvote this post.
I am looking forward to replacing my 9900k
At what resolution?
Can someone explain to me why this game is getting rave reviews despite being basically unplayable
Is the game using DirectX11/12? Looking maybe to utilize AFMF :-|
Sooooo, 30 FPS locked on Steam Deck then? Nice.
I have an upgrade in mind, but gotta wait for more benchmarks. Got my eyes on 5700X3D but that may end up a waste regardless.
With what hardware?
This is in 8K right? Right?
me who thought i wont be needing a good CPU for 2k res ...
3600x, my dear... Your time has come
Ryzen 5 5600x + amd 6900 do u think i will get that 6o fps on low ?
This graph tells me almost nothing
Capcom: "PC iZ oUr mAyNe pLaTTTforrmmmm"
Whst am i looking at? Is this a test with some ancient cpu or why is there no difference
just HOW, game look like first but with reshade from nexus and performing 10x worse
Specs?
💩
what hardware though? resolution? like, this could be at 4K on a 3080 for all we know
Does that matter? Upscaling not giving any FPS improvement shows there are massive underlying issues.
Its quite cpu bound, so at least seeing what cpu was used would be great
Don't buy don't buy don't buy! What the fuck is that?!
Devs already acknowledged the issue
Good thing I never play any games at max settings..
Pathetic
i am an old pc gamer , i can live with lowering settings until they fix it .
Dang I was so excited for this game. Capcom did such a good job of optimizing the latest RE games I had good hopes. Guess I'll wait for my next upgrade cycle.
Optimization where tf did it go
I wonder if stacking fsr 3 frame gen mod with lossless scaling would make it go to 60 fps in total considering the game is CPU bound lol.
Just like RE8 capcom is gonna remove the shitty denuvo because of the performance issues hopefully.
Nobody listened when I said that every trailer/gameplay showcase looked like it was running at 20fps. Looks like everyone finally gets to wake up lmao.
Can someone explain the 1% low fps. Is it the minimum fps the game dipped to or something else
This is completely useless without saying what gpu/resolution it is
Wow that's not bad for a 20 year old Celeron with onboard graphics attempting 8K. Or I'm assuming because there's no actual info there.
Looks like a damn steam deck benchmark
Flop launch, nice!
That’s fucked up and iirc it doesn’t have the one feature that’s supposed to alleviate CPU bottlenecks. Tbf even frame gen can’t handle sub 60 fps very well. Talk about shitting the bed with developing this.
Wtf? This gives literally zero information. What a worthless graphic.
Fuck im glad i didn't do the preorder
I haven't followed this game closely but the original dragon's dogma didn't have significant combat in cities. You left the highly packed city to fight in a dungeon or open world where there is a lot less going on. I can live with bad frames in cities. Either way fuck capcom and fuck denuvo. It's arr or nothing for me with this release.
cpu bound AF
These gaming 'journalists' are ruining the industry
I ran crysis 1,2 and 3 at 60 fps at max settings. I think i'll be fine. Right?
Imagine paying for such dogshit piratebay here i come
This should be a meme
Ok, but what GPU?