So a user on Reddit (ed: u/Yoraxx ) posted on the Starfield subreddit that there was a problem in Starfield when running the game on an AMD Radeon GPU. The issue is very simple, the game just won’t render a star in any solar system when you are at the dayside of a moon or even any planetary object. The issue only occurs on AMD Radeon GPUs and users with the Radeon RX 7000 & RX 6000 GPUs are reporting the same thing.
The issue is that the dayside of any planetary body or moon needs a source of light that gets it all lit up. That source is the star and on any non-AMD GPU, you will see the star/sun in the sky box which will be illuminating light on the surface. But with AMD cards, the star/sun just isn’t there while the planet/moon remains illuminated without any light source.
I’ve got a 7900XTX Ultra, and FSR2 does literally nothing, which is hilarious.
100% resolution scale, 128 FPS.
75% resolution scale … 128 FPS.
50% resolution scale, looking like underwater potatoes … 128 FPS.
I don’t know how it’s possible to make an engine this way, it seems CPU-bound and I’m lucky that I upgraded my CPU not too long ago, I’m outperforming my friend who has an RTX 4090 in literally all scenes, indoor, ship, and outdoor/planet.
He struggles to break 70 FPS on 1080p Ultra, meanwhile I’m doing 4K Ultra.
Some of the benchmarks definitely pointed out that it was CPU bound in many areas (eg. the cities).
I think the HUB one mentioned that some of the forested planets were much more GPU bound and better for testing.
I’m on a tv so capped at 60fps, but I do see a power usage difference with FSR - 75% vs FSR- 100% that’s pretty substantial on my 7900xt.