George Hotz recently stepped down from comma.ai and announced he may devote more effort to his Tinygrad package:
I’m considering another company, the Tiny Corporation. Under 1000 lines, under 3 people, 3x faster than PyTorch? For smaller models, there’s so much left on the table. And if you step away from the well-tread ground of x86 and CUDA, there’s 10x+ performance to gain. Several very simple abstractions cover all modern deep learning, today’s libraries are way too complex.
Superficially, this sounds a bit like the goals of SimpleChains.jl. Question is how SimpleChains.jl differs in approach and goals from his Tinygrad. The speed improvements vs. PyTorch sound pretty comparable. I think both are currently CPU only, and in short term Tinygrad may support Apple Silicon and Google TPU, and long term they want to do their own hardware. What do people think, will SimpleChains exceed it?