a16z General Partner Anjney Midha shares his perspective on how hardware for artificial intelligence will improve —especially at the inference layer — and what that means for how we'll build, train, and interact with AI models.
a16z General Partner Anjney Midha joins the podcast to discuss what's happening with hardware for artificial intelligence. Nvidia might have cornered the market on training workloads for now, but he believes there's a big opportunity at the inference layer — especially for wearable or similar devices that can become a natural part of our everyday interactions.
Here's one small passage that speaks to his larger thesis on where we're heading:
"I think why we're seeing so many developers flock to Ollama is because there is a lot of demand from consumers to interact with language models in private ways. And that means that they're going to have to figure out how to get the models to run locally without ever leaving without ever the user's context, and data leaving the user's device. And that's going to result, I think, in a renaissance of new kinds of chips that are capable of handling massive workloads of inference on device.
"We are yet to see those unlocked, but the good news is that open source models are phenomenal at unlocking efficiency. The open source language model ecosystem is just so ravenous."
More from Anjney:
The Quest for AGI: Q*, Self-Play, and Synthetic Data
Making the Most of Open Source AI
Safety in Numbers: Keeping AI Open
Follow everyone on X: