Originally posted by ultimA
View Post
Rusticl OpenCL Driver Nearing Cross-Vendor Shared Virtual Memory Support
Collapse
X
-
-
-
Originally posted by coder View PostI recall reading about someone getting SVM working on Nvidia GPUs without going through CUDA, but I don't remember the details. I wouldn't presume it's impossible for Rusticle to do, but perhaps either a lower priority or maybe takes more effort, due to the state of their Mesa driver.
Intel supports multi-GPU Ponte Vecchio configurations. Indeed, this is what they have in the Aurora supercomputer.Last edited by ultimA; 06 January 2025, 04:53 AM.
Comment
-
-
Originally posted by coder View PostIf that's your answer, then it was a bad decision by your logic. So, you're saying Nvidia made a mistake by enabling a shared memory model with NVLink because you'd rather the host try to orchestrate all the data movement in the system, instead of letting the nodes pull what they need from wherever they need it, on-the-fly.Last edited by ultimA; 06 January 2025, 04:52 AM.
Comment
-
-
Originally posted by ultimA View PostIntel itself of course could contribute to open-source efforts, but there's no technical or commercial reason to support making it cross-vendor for a data-center-only solution.
Comment
-
-
Originally posted by coder View PostThat's just hilarious. You should walk up to their engineers and try saying that. I'd recommend wearing an extra cushion on your ass, if you ever do.Last edited by ultimA; 06 January 2025, 05:32 AM.
Comment
-
-
Originally posted by ultimA View PostI said, instead they are instead interested in being the most wide-spread and outperforming others. Are you saying Nvidia cares more about performance than dominating the market?
For a different perspective on how Nvidia sees AI hardware, check out their purpose-built DLA engines. Also, those are inference-oriented, so that's another big difference.
Originally posted by ultimA View PostSo stop mentioning asses and maybe pay closer attention to what you are answering to.
Comment
-
-
Originally posted by ultimA View PostIndeed it does. I said in this case there's no reason to make it cross-vendor, not that there's no reason for SVM at all in this case.
Yes the opencl performance depends on the opencl implementation and the limitation of the hardware. There will be particular cases where the p2p memory transfer is cost problem because of the difference in operational speeds..
ultmA think prime where you are rendering on one GPU and outputing on another this is a p2p memory setup that avoids going by the CPU memory. Lot of ways since we have had DMABUF for this it make sense to allow compute workloads running on different gpu to share data the same way and in a GPU neutral way.
Comment
-
Comment