Low fps on AMD gpu
I'm using a 5600XT and this is vanilla with some performance mods added
Still I cannot figure out why my gpu is 99% utilized but the fps remain so low
90 Replies
I also tried Optifine as a last resort but the result is basically the same, so I think it's a hardware problem
52 fps?
What shader are you using?
that FPS is hella respectable
Complementary Reimagined
16 chunks
Yeah I think that's normal
what quality preset?
High (Default)
I just thought it would be higher
Yeah I get around the same FPS on my GTX 1660 ti
I don't know what the nvidia equivalent is to your gpu
Yours
Ah
a GTX 1660 ti?
Huh
yeah so the FPS you're getting is perfectly reasonable, in fact, a bit better than what I get
Yeah they are similar although mine is a bit better
It's now 70 for no reason
Jungles are laggy as heck
It feels all over the place, sometimes it's 50 sometimes it's 90 in the same spot
I saw youtube and like people getting 200~ with bsl and stuffs
That why I wonder if there was anything wrong
that's with crazy powerful GPUs and CPUs
Nah I searched for the same spec
if you don't have a 4090 or the equivalent, you're gonna be sitting around the 60 fps mark
Either they faking it or something is really wrong with mine
yeah that would be odd
Profile: HIGH:linus:
:ioa:
Your shaders setting profile is a little high for that card. Lower it to medium or tune settings manually.
Also my fps plummet whenever I look down on the ground, is the game rendering all the caves even when I can't see it?
For reference, my RX 580 gets a solid 60 fps on medium. Your getting high 50 on high goes to show how expensive that profile is.
First off, you're in an jungle. The ground have a lot of leaves in that biome.
High is default but it generally works well for me
so long as I'm not in a jungle
Leaves are like the laggiest block in the game.
Rendering wise.
yup
Sad
Guess my next gpu will be back to green
Suggestion, install More Culling. It helps reducing the rendering work on leaves blocks by cull (not rendering) unseen leaves.
Sodium does culling out of the box, however, it doesn't do it as aggressive as More Culling does.
I got it to 60 - 90 16 chunks Complementary High by installing cull leaves and setting the javaw priority to high
Thanks guy
The jungle still gonna kill me but normally this is good enough
:haha_yes:
is supposed to run bout 160 fps but ok
They're running in the worst place possible.
:harold:
60 fps on gtx 1660 ti bro I get that on my rtx 3080 with all the optimization mods known to man
It was a normal forest
You also have a 5600xt?
So I just tested on my friend pc which also run a 5600xt, the fps is better and much more stable, even in jungle it doesnt drop to <70
Something is either really wrong with my machine or software
:Render_Thinking:
You must be killing your GPU with something else
Don’t think so Minecraft is fully utilizing my gpu according to f3
that's really weird
even the 2080 ti could get 60 fps with most shaders to my knowledge and that thing is still a real beast of a card
I mean I can get 80 fps if I don’t move around too much and stay in a house my fps is ranges between 50-80 it’s very inconsistent and annoying
cap the FPS to a lower value
Try using GraalVM 21 (or 22) ZGC. I gained 30-50 fps with DH and Complementary on extreme profile with ACL over using G1GC flags.
na
what's GraalVM bruh
Another flavor of Java.
should I install it
nvm that shi too complicated
Unless you want to maximize the server performance, to the average user, the difference is negligible.
For user you can try Shenandoah GC
Have you try running only Iris + Sodium with nothing else?
For some odd reason I decided to remove all other optimization mods and my performance sky rocketed
yes but it's been awhile so I might try again and see
I'm gonna try 1 by 1 later to find out which one is causing the drop
that is the default tbf
Not that powerful my gtx 1070 reaches 200fps with just iris and dh one time it even reached 800fps with no mods
With shaders?
What shaders and what settings? You must've dumped quality because 1070's are good but not that good
No bliss shaders runs really smoothly with reasonable quality and realism
but I am always at 100% gpu usage
same here
da heck, with bliss shader?
Yea
damn
me over here with a 4070 having the same
i wonder if i can improve performance with it too
.... wtf
yeah I dunno man
that is a bit weird
WTH
Yeah it caps at 130 sometimes with vsync turned off, DH, Iris, and shaders enabled.
Try having a look at your nvidia settings
How do I do that? You guys act like I should be running much higher frames
Which I probably should be
I just have no idea how to fix it
Current settings
not really you get like 20-30 more fps than me who has an rtx 3080 so it sounds reasonable
hmm
I now face a new random problem
Sometimes my fps is just low, frequently dropping to 40s even with no shader
A restart or sometimes it just fixed itself
Anyone has an idea why?
world generation?
possible memory leak
I would install FerriteCore and MemoryLeakFix
Nothing works, but non-shader performance is normal now, I can run modded at 200-300 fps. Shader is 40-50fps
My friend got the opposite problem as me (He also use the same gpu). His shader performance is better but non-shader is worse than shader
On plain vanilla with just sodium + iris I can get 120-160 shader fps but it's not stable and drop a lot
I guess it's just something to do with the gpu, gonna cope with no shader until I can afford a better one
same resolution and refresh rate?
same jvm? same flags?
same cpu?
GPU isn't the only factor at play.
I have a RTX 4070TI S and an intel core i9-14900KF and I get around 150 frames with no vsync after playing for over like ten minutes with bliss shaders. If I turn vsync on, it sets it to 60 frames, but can dip and stutter. My main theory is overheating but I’m not sure what’s going on.
You can check if your gpu is overheating with your gpu driver/overclocking software
something like that happens
with the time u will get 12 fps or something
only restarting the game fixes for me
Good news on this, I upgraded to window 11 and it fixes the problem, I have no idea how it worked as I've ruled the possibility that it's window 10 fault at first since I literally just reinstall window 10 a few days ago
Heard my specific software is basically bloatware so I never installed it. Will probably check out fan control though instead
Nevermind got 1 good test then it's all shit again
use hwmonitor or smth to check the temp out
Okay
What about hwinfo?
Oh yeah i meant that one sorry
Do you have it currently?
Yeah
To check your gpu temperature use speccy 64
You will need to install it online
Would you recommend fancontrol?
I personally don't use it but I just did a quick bit of research and it sounds pretty good
This might be the final update, I changed my mainboard because the audio jack is borked and somehow the fps problem is mostly gone
Some session has lower fps than others but never so low that it's unplayable
Final final update: Turn out it was the Resizable RAM feature that was causing all this. My new mainboard has it disabled by default so I thought it was the old mb's problem
I tested a few times and am pretty sure it was the cause
Leaving this here for anyone who also having trouble
My spec:
I5-11400F, H510M-HVD, Sapphire 5600XT, 16Gb Ram
Great if you wouldn't mind marking this post as solved or adding the solved tag if you haven't already
Do you mean AMD’s smart access memory for their GPUs?
So in short, ReBAR doesn't play nice with Sodium/Iris?
:ioa:
Well my mobo is intel so it's named CAM (Clever Access Memory) but yeah on amd it's that
Maybe, I don't know if it's just my machine specifically so I can't tell for sure