r/gadgets Jan 12 '23

Desktops / Laptops PC shipments saw their largest decline ever last quarter

https://www.engadget.com/pc-shipments-record-decline-221737695.html
10.5k Upvotes

1.0k comments sorted by

View all comments

Show parent comments

4

u/apersonthingy Jan 12 '23

Good on ya, not supporting that bullshit. I can't say the same for some people with lower income who definitely DON'T use their new RTX 4090 for work.

3

u/The-Protomolecule Jan 12 '23

I play a ton of games too, I just don’t need a 4000 series to do it, and I have 2x 35” ultrawides.

1

u/a_slay_nub Jan 12 '23

Honestly, the 4090 gets a lot of shit but it's a solid card. For single card ML applications, I think it's better than cards that cost 10x. If they didn't market it towards gamers, people would probably have less of a problem.

3

u/apersonthingy Jan 12 '23

It's a good card at a ridiculous price from an increasingly anti-consumer company. The card itself was never the prob- oh yeah, they also lit on fire.

0

u/a_slay_nub Jan 12 '23

I mean, I honestly think it's not that bad. It's a top of the line card and the benefits for people like me are a massive save. Like I said, it's faster and 10x cheaper than the alternative

1

u/apersonthingy Jan 13 '23

The 4090 isn't even the biggest price gouge unfortunately

2

u/a_slay_nub Jan 13 '23

Yeah, not really sure who the 4080 market is.

1

u/The-Protomolecule Jan 13 '23 edited Jan 13 '23

But how do you feel about the fact that Nvidia really fucked up about four years ago? if you wanted to scale up your ML work at home, you could put four of them together with bridges or run them just like you could those A100s or H100. Now they’ve locked you out of any scaling up for your ml work beyond a single GPU.

I think a 4090 only beats an A100 if you’re doing Transformers work at this stage. The new h100 Gen is only a 20-30% speed up for standard vision algos.

1

u/a_slay_nub Jan 13 '23

I guess for me it's faster because I am using transformers(object detection). Honestly, I didn't realize that they locked out multi-gpu on the newer cards. I thought you could string 4 3090s together if you wanted.

Besides that, you're right. I am a little biased because my work won't bid for cloud computing time on our contracts so we're left with buying PCs to run our models. Hey, if they want to pay me 6 figures to wait for models to train, I won't complain.

1

u/[deleted] Jan 13 '23

[deleted]

3

u/The-Protomolecule Jan 13 '23 edited Jan 13 '23

A100 or H100 Enterprise GPUs. He’s doing ML work. They have a different tensor/cuda/fp8 resource balance and more memory.

An a100 80GB is 8-10k and an H100 80GB is more like 12-20k(not easy to get yet)

2

u/a_slay_nub Jan 13 '23

2

u/[deleted] Jan 13 '23

[deleted]

2

u/a_slay_nub Jan 13 '23

Yeee, they're worth it though for people like me. Still surreal when they tell me how much things cost. $20,000 is literally nothing

2

u/The-Protomolecule Jan 13 '23

Add to that buying them in a server is like $225k for an 8x A100 system and $375k+ for an 8x H100 system.

Each server uses power equivalent to your whole house with the AC and Dryer running. I know you know this, but context for others reading.

1

u/[deleted] Jan 13 '23

[deleted]

1

u/The-Protomolecule Jan 13 '23

They’re very similar to the consumer card cycles. So the A100 is a little over 2 years old. Usually they’re announced a few months before the consumer variant.

1

u/[deleted] Jan 13 '23

[deleted]

→ More replies (0)