All you have to look at is ASIC miners. Once they had them, they were 10x faster than GPUs easily and made GPUs useless for those algos. Something very similar can happen soon.
The fundamentals are different. Bitcoin mining is not intrinsically suited to acceleration on a GPU. It is a not-very-wide serial integer operation.
AI inference on the other hand is basically just very large floating point tensor matrix multiplication. What does an ASIC for matmul look like? A GPU.
This was sponsored by AMD for quite a while and dropped. There had to be some talks behind closed doors that resulted in the current situation. I mean, we all know they should've and they're not dumb either.
All you have to look at is ASIC miners. Once they had them, they were 10x faster than GPUs easily and made GPUs useless for those algos. Something very similar can happen soon.
The fundamentals are different. Bitcoin mining is not intrinsically suited to acceleration on a GPU. It is a not-very-wide serial integer operation.
AI inference on the other hand is basically just very large floating point tensor matrix multiplication. What does an ASIC for matmul look like? A GPU.
Sorta? If that was the full story, TPU would not be a thing.
Asic for matmul is systolic array more or less
They don't even need to invent anything. Just put in some work to polish what already exists: https://github.com/vosen/ZLUDA
IMHO AMD should have done this a decade ago.
This was sponsored by AMD for quite a while and dropped. There had to be some talks behind closed doors that resulted in the current situation. I mean, we all know they should've and they're not dumb either.