Originally posted by DeepDayze
View Post
Announcement
Collapse
No announcement yet.
The Student Working On "Soft" FP64 Support Is Good News For Older GPUs
Collapse
X
-
Originally posted by eydee View PostGood news. While forcing OGL version works most of the time, if you run into an unrelated bug and want to take a closer look, a forced context can be annoying, as things like apitrace don't work properly with them. Having that library just for this reason is a good enough reason in my opinion. Speed doesn't matter, there are still 0 games using fp64.
But still I ran into an annoying bug in Ubuntu 3-4 months ago which made the override option not to work anymore (and side effects). Neither Dirt Showdown nor any other game requiring the override to 4.1 would work. A Steam or xserver-xorg-video-ati update eventually solved it 2 months later.
If this work is integrated to mesa, then this kind of bug can be avoided without having to manually override the GL and GLSL versions. This has been awaited for some time, hence I'm just glad it's on its way.
If everybody gets to thank him in french:
Bon boulot! Ça faisait un moment que les détenteurs de cartes un peu plus anciennes (mais toujours en bon état) attendaient ça! Merci.
Comment
-
Hi,
can someone explain me about FP64 performance. My collages are researchers and they develop algorithm that need doubles. Single floats are not good enough (the data would corrupt too much)...
Anyway we NEED doubles
There are lists of AMD and NVIDIA gpus on Wikipedia. The gpus have Single percision and Double percision score in GFLOPS. All nvidias with Maxwell and Pascal has BAD double performance - 1/32 (3%) of single (except 5000$ "P100"). Even freaking Nvidia Titan X has 317GFLOPS in FP64 (vs RX 480: 323)
The AMD has bit better ratio, mostly 1/16 (6%), or even better in the top class.
Does this mean that 99% of GPUs have "soft FP64"? Or simply very small HW part for double arithmetics?
Also why doesn't lower GPUS (R5, HD x6xx or less, ...) mentioned GFLOPS for doubles?
- Likes 1
Comment
-
Originally posted by gsedej View PostDoes this mean that 99% of GPUs have "soft FP64"? Or simply very small HW part for double arithmetics?
Originally posted by gsedej View PostAlso why doesn't lower GPUS (R5, HD x6xx or less, ...) mentioned GFLOPS for doubles?
Comment
-
Originally posted by gsedej View PostHi,
can someone explain me about FP64 performance. My collages are researchers and they develop algorithm that need doubles. Single floats are not good enough (the data would corrupt too much)...
Anyway we NEED doubles
There are lists of AMD and NVIDIA gpus on Wikipedia. The gpus have Single percision and Double percision score in GFLOPS. All nvidias with Maxwell and Pascal has BAD double performance - 1/32 (3%) of single (except 5000$ "P100"). Even freaking Nvidia Titan X has 317GFLOPS in FP64 (vs RX 480: 323)
The AMD has bit better ratio, mostly 1/16 (6%), or even better in the top class.
Does this mean that 99% of GPUs have "soft FP64"? Or simply very small HW part for double arithmetics?
Also why doesn't lower GPUS (R5, HD x6xx or less, ...) mentioned GFLOPS for doubles?
So, NVidia and AMD have worked to segment their hardware sales, to make lots more money by selling "professional" cards like Tesla that provide good fp64 support. Then they provide the slow 1/32 performance in their normal line of cards primarily so that anyone can test and develop fp64 there, even if they don't have great performance.
In a few cases, this limitation was done artificially, and the underlying hardware is actually there and just limited by software (BIOS or drivers). In other cases, the hardware is actually missing the extra fp64 hardware (an optimization which lets them pack in extra fp32 hardware for the same cost).
The older cards which don't list a speed are the ones without hardware support. I'd expect the soft float performance would be quite slow though. Presumably it could vary depending on the exact workload and mix of operations you were using.Last edited by smitty3268; 02 October 2016, 05:37 PM.
Comment
Comment