this post was submitted on 22 Nov 2024
38 points (100.0% liked)

PCGaming

6525 readers
108 users here now

Rule 0: Be civil

Rule #1: No spam, porn, or facilitating piracy

Rule #2: No advertisements

Rule #3: No memes, PCMR language, or low-effort posts/comments

Rule #4: No tech support or game help questions

Rule #5: No questions about building/buying computers, hardware, peripherals, furniture, etc.

Rule #6: No game suggestions, friend requests, surveys, or begging.

Rule #7: No Let's Plays, streams, highlight reels/montages, random videos or shorts

Rule #8: No off-topic posts/comments

Rule #9: Use the original source, no editorialized titles, no duplicates

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] Ptsf@lemmy.world 3 points 4 hours ago* (last edited 4 hours ago)

Not to defend nvidia entirely, but there are physical cost savings that used to occur with actual die wafer shrinkage back in the day since process node improvements allowed such a substantial increase in transistor density. Improvements in recent years have been lesser and now they have to use larger and larger dies to increase performance despite the process improvements. This leads to things like the 400w 4090 despite it being significantly more efficient per watt and causes them to get less gpus per silicon wafer since the dies are all industry standardized for the extremely specialized chip manufacturering equipment. Less dies per wafer means higher chip costs by a pretty big factor. That being said they're certainly... "Proud of their work".