Tesla DOJO Exa-Scale Lossy AI Network using the Tesla Transport Protocol over Ethernet TTPoE
Tesla DOJO Exa-Scale Lossy AI Network using the Tesla Transport Protocol over Ethernet TTPoE
Tesla-Answer is a dedicated community forum for Tesla owners and enthusiasts. Our platform connects experienced Tesla drivers who share practical knowledge, troubleshooting tips, and maintenance insights to help you maintain and upgrade your Tesla.
Tesla DOJO Exa-Scale Lossy AI Network using the Tesla Transport Protocol over Ethernet TTPoE
Tesla's wafer-sized Dojo processor is in production
What Is Going On With Tesla's Dojo? - CleanTechnica
Tesla's head of Dojo supercomputer is out, possibly over issues with next-gen (in addition to earlier Dojo delays)
Did some ballpark rough math, and while I probably got some things wrong, Dojo D1 looks super uncompetitive.
It's got a ton of cache with a ton of local bandwidth which is nice and a ton of interconnect bandwidth... But it has no dram, so it has to access dram over the interconnect bandwidth with horrible latency.
Cost isn't too bad at first glance, but performance / $ is worse than the most expensive Nvidia GPUs while being far less capable. Performance / watt is worse than just buying consumer cards like the 4090 or 7900xtx.
If you want the best possible performance or the most ram, amd mi300x or Nvidia h100 and h200 are the way to go, with the most capable hardware and software.
If you want the cheapest hardware per int16 performance, 4090 or 7900xtx.
If you want the most cost effective compute gpu, it's the Intel 1550. If you want the most power efficient compute gpu, it's the H100 or H200. If you want the most cost effective and power efficient ai accelerator, it's the Intel Gaudi 2 and probably the Intel Gaudi 3 soon. If you want the most cost effective and scalable cloud Ai accelerator, it's some form of Google TPU.
But just based on publisher specs, Dojo gives you the performance / dollar of the most expensive Nvidia GPUs with the performance / watt of consumer gpus that are handicapped for machine learning. And you don't have any DRAM, the programming model is an AI accelerator (possibly worse than Intel Gaudi or Google TPU) and not a GPU, and it's functionality is that of an AI accelerator and not a GPU.
Just completely uncompetitive with a weird architecture.