BTW, I'm not really sure what technology they used, but I'm reminded of a reconfigurable computing company called Stretch, Inc. One thing I've heard is that they're so fast at re-configuring their on-chip logic that they'd do it up to several times, in the course of compressing a single H.264 video frame. It seems they've been acquired.
Announcement
Collapse
No announcement yet.
Tachyum Gets FreeBSD Running On Their Prodigy ISA Emulation Platform For AI / HPC
Collapse
X
-
Originally posted by coder View PostI have no idea what they're doing, but I'm imagining you could implement a variable-ISA CPU by having a sufficiently general backend, sitting behind a frontend that's implemented using FPGA-like configurability (although more specialized to the task of instruction decoding than a normal FPGA would be).
Of course, there's no way that'd be faster or more efficient than the fully hard-wired implementations it'd be competing with, so I doubt that's what it really is.
(Not that I would actually expect that magnetic semiconductors will work at room temperature anytime soon or ever, but there might be other materials I am not aware of that might offer similar potential.)
- Likes 2
Leave a comment:
-
Originally posted by brucethemoose View PostNo shuffling data around over slow, power-inefficient buses like PCIe. And it eliminates redundancy, as all the accelerators in a core can share the same decoders, the same cache and so on. If the design uses heterogenous cores (that all still share an ISA), then it still makes sharing LLC and other resources easier.
And when you're running a deep learning workload, it would thrash the hell out of your system-level cache, making it next to useless for CPU jobs. IMO, you really don't want these things in the same package. It's just a matter of cost/size/convenience for mobile platforms that they are.
Originally posted by brucethemoose View PostOn the software side, no more split code between different devices, and less software infrastructure to maintain. And I assume compiler optimization and language support would be easier/better when a whole app is targeting a single ISA.
Now, someone is probably going to point to Fujitsu's A64FX, which is a legit point. However, it's not great at truly general-purpose compute. It tries to be a hybrid between a CPU and GPU, but is as good at neither workload as the corresponding processor. Its main benefits are that you only need a single toolchain and it can tackle a broader scope of problems than a traditional GPU.
- Likes 3
Leave a comment:
-
Originally posted by timofonic View Postcurrent architecture mess is a total disaster,
Originally posted by timofonic View Postprices skyrocketting,
- Likes 1
Leave a comment:
-
Originally posted by soulsource View PostWhat I'm missing here are tech details on the hardware.
Of course, there's no way that'd be faster or more efficient than the fully hard-wired implementations it'd be competing with, so I doubt that's what it really is.
- Likes 1
Leave a comment:
-
Originally posted by sinepgib View Post
Why, exactly?
On the software side, no more split code between different devices, and less software infrastructure to maintain. And I assume compiler optimization and language support would be easier/better when a whole app is targeting a single ISA.
- Likes 1
Leave a comment:
-
It's too bad that they're so full of shit, because the technology is actually cool.
Leave a comment:
-
Originally posted by sinepgib View Post
It's dubious that you can have all three. Look at general purpose CPUs. You know why they don't excel at AI, gfx, HPC, etc, all of their uses? Because specialization leads to better thermal, silicon use, power consumption and ISA properties, that's why the current trend is adding more units. GPUs excel at what they do, which is transformation of floating point values in a highly parallel, dependency free (or close to) setting. The price is that they're completely useless to drive a system. You can't have your cake and eat it too, whoever says you do is either delusional or a liar.
Now, you can make things cheaper. If you assume you have specialized hardware you can simplify your general purpose CPUs, for example. For the tasks that do remain in your CPU, you could also probably make it somewhat faster. But 10x lower power consumption and 1/3 of the price while being faster? Nope, really unlikely to become true unless we talk about a maaaajor breakthrough. I guess they play with not needing as many chips for the 1/3 cost claim, but the scale only will probably make that assertion false.
Right now you shuffle data to the CPU, or GPU, or tensor processor or whatever with its own instructions... but it would be advantageous to have those various components as part of the same core, using the same ISA.
AMD pushed this but kinda fumbled it, particularly in the AI space. Nvidia and Intel are going there as well (with Tegra/AMX), albeit very slowly, as they are heavily invested in their own discrete niches.
Leave a comment:
Leave a comment: