← October 3, 2025 edition

tensorpool

Vercel For GPUs

TensorPool Wants to Be the Vercel of GPUs, and the Pricing Math Actually Works

AIOpsDeveloper ToolsSaaSDevOpsCloud Computing

The Macro: GPU Compute Is Expensive and Getting Worse

Training ML models requires GPUs. GPUs are expensive. This is not news, but the scale of the problem keeps getting worse. An H100 on AWS runs roughly $30 per hour on-demand. Training a serious model can take days or weeks. The math gets ugly fast, and it’s the reason most ML teams spend a disturbing percentage of their engineering time on infrastructure instead of actual model work.

The GPU cloud market has fragmented in response. Lambda Labs built a business on cheaper GPU instances. RunPod offers on-demand and spot GPUs at lower prices. CoreWeave has raised billions positioning itself as the GPU cloud alternative to the big three. Modal took a serverless approach to compute. AWS SageMaker exists and does approximately everything, which is both its strength and its weakness.

The common thread across all of these is that they require you to think about infrastructure. You’re still choosing instance types, managing environments, handling failures, and paying for idle time. For an ML engineer who just wants to train a model, this is overhead. Necessary overhead, but overhead.

The “Vercel for X” framing gets overused in startup pitches, but for GPU compute it actually maps cleanly. Vercel’s insight was that frontend developers don’t want to think about servers. TensorPool’s bet is that ML engineers don’t want to think about GPU orchestration. The question is whether the abstraction can be high enough to be useful without being so high that it loses flexibility.

The Micro: Stanford, NVIDIA, and a CLI That Does the Boring Parts

TensorPool is a command-line tool. You describe your training job, and it handles GPU selection, orchestration, and execution. The pricing claim is half the cost of major cloud providers. If that holds up at scale, it’s a compelling pitch for any team currently wincing at their AWS bill.

The team is three people, all from Stanford, and their collective resume reads like someone designed it in a lab to build exactly this product. CEO Tycho Svoboda came from Blackstone, which is unusual for a dev tools founder but means he probably understands pricing and unit economics at a cellular level. CTO Joshua Martinez worked at NVIDIA, Apple, and Chan Zuckerberg Initiative, with Stanford CS and Stats degrees. He’s literally been on the GPU hardware side. CPO Hlumelo Notshe comes from DeepMind and NextDoor, Stanford CS and Math. That’s a team where each person has direct experience with a different piece of the problem. They came through YC’s Winter 2025 batch.

The product is open source on GitHub, which is a smart move for developer trust. The CLI-first approach means they’re targeting ML engineers directly, not going through procurement departments. That’s a bottom-up adoption play, similar to how Vercel and Railway grew.

The Verdict

TensorPool is solving a real problem for a clearly defined audience. ML engineers who are tired of being part-time infrastructure managers will try a CLI that promises to handle the GPU stuff at half the cost. The pitch is clean and the team has the exact right background for it.

The risk is margin. If they’re offering half-price GPU compute, they need to be buying or accessing GPUs at a price that leaves room for a real business. GPU arbitrage is a thin-margin game, and the providers they’re competing with have significant scale advantages. Lambda Labs has been at this for years. CoreWeave has raised $12 billion. RunPod has aggressive pricing already.

The other risk is the abstraction layer itself. ML training jobs vary enormously in their requirements. A simple fine-tuning run is different from distributed training across multiple nodes is different from reinforcement learning workloads with complex dependencies. If TensorPool handles the simple cases well but falls apart on complex multi-GPU training, power users will hit the wall fast and go back to managing their own infrastructure.

At 30 days, I’d look at GitHub stars and CLI downloads as a proxy for developer interest. At 60 days, the question is whether teams are using it for production training runs or just experiments. By 90 days, retention will tell the story. If people try it, save money, and keep using it, the business works. If they try it, hit a limitation, and go back to Lambda Labs, it doesn’t. I think the team gives them a better shot than most. The Stanford-NVIDIA-DeepMind combination is hard to beat for credibility in this specific space.