← August 28, 2026 edition

arcten

Infrastructure for deploying AI agents into production

Arcten Is Building the Production Runtime That AI Agents Still Do Not Have

The Macro: Everyone Is Building Agents, Nobody Knows How to Ship Them

I have a theory about where the AI agent market is right now. We are in the “easy demo, hard production” phase. Building a prototype agent that works in a notebook takes an afternoon. Getting that same agent to run reliably in production, with conversation persistence, error handling, scaling, safety guardrails, and monitoring, takes weeks of engineering work that has nothing to do with AI and everything to do with infrastructure.

This is a familiar pattern in technology. The gap between prototype and production is where most software lives and most startups die. Web apps had this problem before Heroku and then AWS Lambda made deployment trivial. Mobile apps had it before Firebase handled the backend. Every new computing paradigm goes through a phase where building the thing is easy and shipping the thing is unreasonably hard. AI agents are in that phase right now.

The market for agent infrastructure is getting crowded. LangChain and LlamaIndex handle the orchestration layer. Modal and Replicate handle model serving. Vercel’s AI SDK handles the frontend integration. But none of these solve the full production stack. You still need to stitch together conversation state management, data source connectors, safety layers, scaling logic, and monitoring dashboards. Each of those is its own small engineering project, and the integration surface between them is where bugs live.

The companies that will win this market are the ones that collapse the entire production stack into a single platform. Not a framework you assemble yourself, but a runtime you deploy to. That is a meaningfully different product from what exists today.

The Micro: Two Caltech AI Researchers Who Published at NeurIPS and CVPR

Arcten provides a runtime and SDK for deploying AI agents into production. The platform handles conversation persistence, data source integration, scaling, safety guardrails, and monitoring. Instead of spending weeks wiring together infrastructure, developers integrate Arcten’s SDK and deploy agents that are production-ready from the start.

Armeet Singh Jatyani is a founder. He conducted AI research at Caltech under the former director of AI at NVIDIA. His work on neural operator architectures was published at CVPR and NeurIPS. Michael Nguyen Jr is the other founder. He collaborated with Caltech and DeepMind researchers, and his research on agentic behavior and novel AI architectures was also presented at NeurIPS. Both founders come from the research side of AI, which is unusual for an infrastructure company and worth paying attention to.

They are a two-person team in San Francisco, Y Combinator Fall 2025 batch. The company was founded in 2024, which means they were building before the current agent infrastructure wave hit peak hype. That is a good sign. It suggests they identified the production gap from their own research experience rather than chasing a trend.

The product includes hosted RAG for connecting agents to data sources, pre-built data connectors, embedded UI components for integrating agents into existing applications, and safety guardrails for controlling agent behavior in production. The embedded UI piece is interesting because it addresses a problem I see constantly. Companies build agents that work great in a standalone chat interface but have no way to integrate them into their actual product. Arcten’s embeddable components mean the agent can live inside an existing app rather than requiring users to switch to a separate interface.

The competitive landscape here is real. Steamship was early to the agent hosting space. Fixie built an agent platform. Langbase is doing agent deployment. Helicone handles observability. But most of these tools solve one slice of the production problem. Arcten is trying to be the complete runtime, which is both its value proposition and its risk. A complete runtime is more useful than individual components, but it is also harder to build, harder to maintain, and harder to sell because the buyer needs to trust you with their entire agent stack.

The research backgrounds of the founders cut both ways. On one hand, they understand AI systems at a depth that most infrastructure founders do not. They know what agents need because they have built them. On the other hand, the transition from research to product company is notoriously difficult. Research rewards novelty. Infrastructure rewards reliability. Those are different muscles, and the question is whether Armeet and Michael can make the switch.

The Verdict

The production infrastructure gap for AI agents is real and getting more urgent by the month. Every company I talk to that has built an agent prototype is running into the same set of problems when they try to ship it. Arcten is going after the full stack, which is the right approach but the hardest one to execute.

At 30 days, I want to see how many agents are running on their platform in production, not in development, and what the uptime looks like. Infrastructure companies live and die on reliability metrics. At 60 days, I want to understand the developer experience. Is the SDK intuitive enough that a developer can go from prototype to production in a day, or does it require significant ramp-up time? At 90 days, the question is positioning. The agent infrastructure market is getting crowded, and Arcten needs to articulate clearly why their runtime is better than assembling individual best-of-breed components. The research pedigree gives them credibility. Turning that credibility into market share is the hard part.