Originally posted by coder
View Post
Announcement
Collapse
No announcement yet.
AMD Announces Radeon RX 7900 XTX / RX 7900 XT Graphics Cards - Linux Driver Support Expectations
Collapse
X
-
Originally posted by coder View PostIt's a distinction commonly used to describe deep learning ASICs.
For instance: https://www.anandtech.com/show/14187...ators-for-2020
Originally posted by coder View PostFirst off, Nvidia doesn't permit gaming cards to be used in data centers. So, they wouldn't even market the RTX 3090 for deep learning.
Notice how it says "TensorFlow/Resnet50 Training"? And these are both less powerful and less power hungry GPUs than a top-line desktop card, so by the logic in your next line the latter should be even less of a good fit for inference:
Originally posted by coder View PostSecond, you should be looking at whether it's more cost-effective to use the A40 or the A100 for inference, and then tell me using the A40 for inference is a waste.
If you're a deep learning practitioner/researcher have just a couple of cards available in a local workstation, it makes sense to make the most use of the resources you have available so that the fixed cost of the card is amortized. Because local machines are used primarily for prototyping, this means that the vast majority of your workload will be training. Outside of production, it is diminishingly rare to have local inference requirements so onerous that they require you to get a top-of-the-line consumer card just to keep up.
Originally posted by coder View PostBecause you're probably a student or hobbyist, and that's the best thing you can afford to train on. Moreover, a researcher is primarily focused on model development, not deployment at scale. When a model has been developed for commercial purposes, it needs to be deployed to achieve a return on the investment of developing it. That means putting a lot more data through it than would typically be used to train it. And that means you want hardware that's not overkill for the purpose, since you're probably using many instances and tying them up for long periods of time.
Originally posted by coder View PostThe word "oriented" is key. Nobody is saying you couldn't use an A100 for inference, just that it's generally overkill for that task.
And it's not like we don't have a barometer on how people see these cards in the context of deep learning work either. For example, a ton of people in this space look at the benchmarks Lambda Labs does when new GPUs are released, because getting people access to hardware for deep learning is their MO. Guess what they decide to focus on benchmarking when new 80/90 series cards drop? Not inference, that's what.
For anyone still unsure what the right answer to this discussion is, here's a quick way to get yourself some closure. Find a handful of ML researchers/practitioners/engineers and ask them the following questions:- What do you use for training your deep learning models?
- Would you consider the RTX 3090 a training oriented or inference oriented GPU?
- Local workstations with GTX/Quadro cards for prototyping + clusters/cloud (which use data centre cards).
- Blank stares and confused expressions
Edit: well, I'm not sure what I expected from a response. For those unfortunate folks who come across this waste of bandwidth discussion in the future, I hope it was at least a nice showcase of how one can carry on an internet argument indefinitely with just unsubstantiated claims, general statements, no concrete evidence and without ever consulting people who might actually understand something about the topic they're writing about. If you or your company find yourself in the market for ML hardware, don't consult Phoronix forumsLast edited by StillStuckOnSI; 15 November 2022, 10:00 PM.
Comment
-
Originally posted by StillStuckOnSI View PostAnd not everyone is working out of a machine in a data centre. Also, I'd be careful with the claim that Nvidia wouldn't market consumer cards for deep learning. See this official page on RTX in gaming/productivity laptops: https://www.nvidia.com/en-us/geforce...s/stem-majors/.
If you look at the URI, they're obviously targeting college kids who are going to buy a gaming GPU no matter what.
Originally posted by StillStuckOnSI View PostNotice how it says "TensorFlow/Resnet50 Training"? And these are both less powerful and less power hungry GPUs than a top-line desktop card, so by the logic in your next line the latter should be even less of a good fit for inference:
Originally posted by StillStuckOnSI View PostWhat we're arguing here is not that, but whether a 3090-level card makes more sense for inference than it does for training.
Originally posted by StillStuckOnSI View PostGuess what they decide to focus on benchmarking when new 80/90 series cards drop? Not inference, that's what.
Originally posted by StillStuckOnSI View PostFor anyone still unsure what the right answer to this discussion is, here's a quick way to get yourself some closure. Find a handful of ML researchers/practitioners/engineers and ask them the following questions:- What do you use for training your deep learning models?
- Would you consider the RTX 3090 a training oriented or inference oriented GPU?
When you need a dump truck, it's usually the only viable option. And by the time you reach the point of doing construction jobs that require a dump truck, you typically know enough to figure out when one is needed and what size/type is required.
Comment
Comment