this post was submitted on 04 Jul 2025
381 points (95.2% liked)

Technology

72452 readers
2552 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] just_another_person@lemmy.world 91 points 2 days ago (67 children)

My mind is still blown on why people are so interested in spending 2x the cost of the entire machine they are playing on AND a hefty power utility bill to run these awful products from Nvidia. Generational improvements are minor on the performance side, and fucking AWFUL on the product and efficiency side. You'd think people would have learned their lessons a decade ago.

[–] RazgrizOne@piefed.zip 17 points 2 days ago (23 children)

Once the 9070 dropped all arguments for Nvidia stopped being worthy of consideration outside of very niche/fringe needs.

[–] CheeseNoodle@lemmy.world 5 points 2 days ago (1 children)

Got my 9070XT at retail (well retail + VAT but thats retail for my country) and my entire PC costs less than a 5090.

[–] RazgrizOne@piefed.zip 2 points 2 days ago* (last edited 2 days ago) (1 children)

Yeah I got a 9070 + 9800x3d for around $1100 all-in. Couldn’t be happier with the performance. Expedition 33 running max settings at 3440x1440 and 80-90fps

[–] FreedomAdvocate@lemmy.net.au -1 points 1 day ago* (last edited 1 day ago) (1 children)

But your performance isn’t even close to that of a 5090…….

80-90 fps @ 1440 isn’t great. That’s like last gen mid tier nvidia gpu performance.

[–] RazgrizOne@piefed.zip 4 points 1 day ago* (last edited 1 day ago) (2 children)

Not 1440 like you’re thinking. 3440x1440 is 20% more pixel to render than standard 2560x1440’s. It’s a WS. And yes at max settings 80-90fps is pretty damn good. It regularly goes over 100 in less busy environments.

And yeah it’s not matching a 5090, a graphics card that costs more than 3x mine and sure as hell isn’t giving 3x the performance.

You’re moving the goalposts. My point is for 1/4th the cost you’re getting 60-80% of the performance of overpriced, massive, power hungry Nvidia cards (depending on what model you want to compare to). Bang for buck, AMD smokes Nvidia. It’s not even close.

Unless cost isn’t a barrier to you or you have very specific needs they make no sense to buy. If you’ve got disposable income for days then fuck it buy away.

[–] FreedomAdvocate@lemmy.net.au 1 points 1 day ago (1 children)

I assume people mean 3440x1440 when they say 1440 as it’s way more common than 2560x1440.

Your card is comparable to a 5070, which is basically the same price as yours. There’s no doubt the 5080 and 5090 are disappointing in their performance compared to these mid-high cards, but your card can’t compete with them and nvidia offer a comparable card at the same price point as AMDs best card.

Also the AMD card uses more power than the nvidia equivalent (9700xt vs 5070).

[–] RazgrizOne@piefed.zip 2 points 1 day ago* (last edited 1 day ago) (1 children)

I assume people mean 3440x1440 when they say 1440 as it’s way more common than 2560x1440.

Most people do not use WS as evidenced by the mixed bag support it gets. 1440 monitors are by default understood to be 2560x1440p as it’s 16:9 which is still considered the “default” by the vast majority of businesses and people alike. You may operate as if most people using 1440+ are on WS but that’s a very atypical assumption.

Raytracing sure but otherwise the 4090 is actually better than the 5070 in many respects. So you’re paying a comparable price for Raytracing and windows dependency, which if that is important to you then go right ahead. Ultimately though my point is that there is no point in buying the insanely overpriced Nvidia offerings when you have excellent AMD offerings for a fraction of the price that don’t have all sorts of little pitfalls/compromises. The Nvidia headaches are worth it for performance, which unless you 3-4x your investment you’re not getting more of. So the 5070 is moot.

I’m not sure what you’re comparing at the end unless you meant a 9070XT which I don’t use/have and wasn’t comparing.

[–] FreedomAdvocate@lemmy.net.au 1 points 20 hours ago

I’m not sure what you’re comparing at the end unless you meant a 9070XT which I don’t use/have and wasn’t comparing.

Sorry I thought I read you had the 9070 XT, which is better than the 9070 that you have. The 9070 and the 5070 are the same price, and are neck and neck in performance , so the nvidia card isn't "insanely overpriced" compared to AMDs offerings, is it? The 9070 isn't a "fraction of the price" of the equivalent nvidia card, it's the same price.

As you said, there are 40 series cards that are better than the 50 series cards apart from probably the 5090, and the prices on those is cheaper than the 9070.

[–] FreedomAdvocate@lemmy.net.au 0 points 1 day ago (1 children)

Better bang for your buck, but way less bang and not as impressive of a bang.

[–] RazgrizOne@piefed.zip 1 points 1 day ago* (last edited 1 day ago) (1 children)

Not all of us can afford to spend $3000 for a noticeable but still not massive performance bump over a $700 option. I don’t really understand how this is so difficult to understand lol. You also have to increase the rest of your machine cost for things like your PSU, because the draw on the 5xxx series is cracked out. Motherboard, CPU, all of that has to be cranked up unless you want bottlenecks. Don’t forget your high end 165hz monitor unless you want to waste frames/colors. And are we really going to pretend after 100fps the difference is that big of a deal?

Going Nvidia also means unless you want to be fighting your machine all the time, you need to keep a Windows partition on your computer. Have fun with that.

At the end of the day buy what you want dude, but I’m pulling down what I said above on a machine that cost about $1700. Do with that what you will

[–] glog78@digitalcourage.social 1 points 1 day ago (2 children)

@RazgrizOne @FreedomAdvocate the reason why i decided for AMD after being nearly all my life team green ( aka >20 years ) , i feel like AI Frame Generation and Upscalling are anti consumer cause the hide the real performance behind none reproducable image generation. And if you look correctly ... this is how nvidia has a performance lead over AMD.

[–] FreedomAdvocate@lemmy.net.au 1 points 20 hours ago (1 children)

Calling DLSS "anti consumer" is one of the dumbest things I've read about PC gaming in a long time.

[–] glog78@digitalcourage.social -1 points 18 hours ago* (last edited 18 hours ago) (1 children)

@FreedomAdvocate you remember the time when AMD was called out for even the smallest of difference from a default render ? Now since nvidia basically use some kind of statistic guessing method -> Noone is allowed to call them out ?
I call them out cause basically they removed the possibility for any consumer to compare other graphics card with themself. Or did i miss nvidia making dlss / frametime generation and all the features available on other gpu brands ?
Do you know AI Models behind all this and how they would perform on other hardware ? Do we want to talk about how they try to force media to have access to tests ? Yes imho there is alot anti consumer here ...

[–] FreedomAdvocate@lemmy.net.au 1 points 18 hours ago* (last edited 18 hours ago) (2 children)

No, I don’t remember that. What are you talking about?

Why would Nvidia make DLSS work on other brands hardware? It’s hardware dependant btw - it needs their cuda cores.

[–] glog78@digitalcourage.social 0 points 16 hours ago (1 children)

@FreedomAdvocate ... this question is totally unimportant for the fact that their current behaivior is not very consumer friendly or harder expressed anti consumer.

Second cuda is not hardware dependend ;) https://github.com/vosen/ZLUDA/tree/master | https://www.xda-developers.com/nvidia-cuda-amd-zluda/

"Imagine a world where noone needed a brand specific addition to have modern features" ... oh those ideas exist since centuries ( DX / OpenGL / Vulkan .... ) ... now ask yourself why nvidia always tries to operate outside of those api's ?

....

[–] FreedomAdvocate@lemmy.net.au 1 points 15 hours ago (1 children)

Second cuda is not hardware dependend

That's essentially an emulation layer. Nvidia make DLSS specifically for their GPUs, which have CUDA cores on them. It's the reason why DLSS doesn't work on their pre-CUDA core hardware.

Could they make DLSS work on AMDs hardware? Sure, they could - but it would not be DLSS as we know it, and again - why would they? They are allowed to make stuff exclusively for their hardware.

[–] glog78@digitalcourage.social -1 points 10 hours ago (1 children)

@FreedomAdvocate zuda is an reimplementation of an api not a emulation.

[–] FreedomAdvocate@lemmy.net.au 1 points 2 hours ago

I said it’s essentially emulation, which it is. Its like WINE, which is also essentially emulation but isn’t emulation.

[–] glog78@digitalcourage.social -1 points 18 hours ago* (last edited 18 hours ago) (1 children)
[–] FreedomAdvocate@lemmy.net.au 1 points 15 hours ago* (last edited 15 hours ago) (1 children)

But nvidia got dragged across the coals for using frame-gen in their performance benchmarks too. Did you miss that?

Also ATI wasn't owned by AMD then.....AMD aquired ATI in 2006. Your link is from 2001.

Also no one should be listening to official GPU manufacturer benchmark results. No one. Review companies do their own benchmarking, and you do know that you can turn off DLSS and DLSS Frame-Gen, don't you? I haven't seen any reviewers only compare DLSS+Frame-Gen on an nvidia card to native-with-no-frame-gen on AMD cards. You must have, so can you link to any?

[–] glog78@digitalcourage.social 0 points 10 hours ago (1 children)

@FreedomAdvocate so you didn't read the heise link which showed you that pre release tests had strict rules on how to test including framegen settings ...

[–] FreedomAdvocate@lemmy.net.au 1 points 2 hours ago* (last edited 2 hours ago)

Nvidia can say what they want, but reviewers didn’t follow those.

Sounds like you need to find better GPU review sites.

[–] RazgrizOne@piefed.zip 1 points 1 day ago* (last edited 1 day ago) (1 children)

I’m not even against tricks like upscaling and such to be honest. If it looks good I’ll take it lol. But I do agree they don’t feel like long-term, hardened solutions vs something more like “raw performance.” And there’s no doubt There is a certain elegance to AMD’s cards

[–] FreedomAdvocate@lemmy.net.au 1 points 20 hours ago

And there’s no doubt There is a certain elegance to AMD’s cards

What exactly do you mean by this?

load more comments (21 replies)
load more comments (64 replies)