← June 4, 2027 edition

replicas

Use background coding agents from anywhere

Replicas Gives Your Coding Agents Their Own Sandboxed Development Environments So They Stop Trashing Your Branch

Developer ToolsAICoding AgentsB2B

The Macro: Coding Agents Need Their Own Space to Work

The first generation of AI coding assistants lived inside your IDE. Copilot suggested code as you typed. Cursor helped you edit files. These tools are useful, but they are fundamentally human-in-the-loop. The developer is always present, always reviewing, always directing.

The next generation of coding tools operates differently. Agents like Claude Code and Codex can take a task description, work on it independently, and produce a complete pull request. They do not need someone watching over their shoulder. They can work in the background while the developer does something else.

But running these agents on your local machine is messy. They modify files. They install dependencies. They run builds. If something goes wrong, you are stuck cleaning up the mess in your own development environment. And you cannot run multiple agents on different tasks because they will conflict with each other.

The solution is sandboxed environments. Each agent gets its own isolated VM with a full development environment, including databases, package managers, and all the tools it needs. It works in its sandbox. When it finishes, it produces a clean pull request. If something goes wrong, you just discard the sandbox. Your local environment is untouched.

Replicas, backed by Y Combinator, provides exactly this: sandboxed development environments for background coding agents, accessible from anywhere.

The Micro: Delegate From GitHub, Slack, or Linear

The multi-platform access is clever. You can delegate tasks to Replicas agents from GitHub issues, Slack messages, Linear tickets, or a web dashboard. This means you can assign work to coding agents from wherever you are already tracking tasks. No context switching to a separate tool.

Each agent runs in its own isolated VM with development tools, databases, and package managers pre-configured. The agents support Claude Code and Codex today, with OpenCode coming soon. Security is a priority: code stays in your connected repositories and is not stored or used for training by Replicas.

The traction numbers are interesting. Engineering teams reportedly ship over 30% of their pull requests through Replicas. That is not a marginal contribution. That is a significant share of engineering output being produced by background agents. Customers include Helicone, Bluma, and Sorce.

The founder, Connor Loi, is a Waterloo CS student building out of the gate with YC’s P26 batch. The lean team and early traction suggest focused execution.

The competitive space includes Devin from Cognition Labs, which positions itself as a fully autonomous software engineer. Codex from an LLM provider runs background coding tasks. Superset orchestrates multiple agents in parallel on your local machine. Replicas differentiates by focusing on the sandboxed environment and multi-platform integration rather than the agent itself. It is agent-agnostic infrastructure.

The distinction between Replicas and Superset is worth noting. Superset runs agents on your local machine in Git worktrees. Replicas runs agents in remote VMs. Both solve the isolation problem but with different tradeoffs: Superset is free and local, Replicas is remote and managed. For teams that want to offload agent compute entirely, Replicas makes more sense.

The Verdict

Background coding agents are becoming a standard part of engineering workflows, and they need infrastructure. Replicas is building that infrastructure with a focus on isolation, security, and multi-platform accessibility.

At 30 days: what is the average quality of agent-produced pull requests? Merge rate without modifications is the clearest quality signal.

At 60 days: how many agents are teams running simultaneously through Replicas? Concurrency shows whether teams are using this for one-off tasks or as a core part of their development process.

At 90 days: are engineering leaders tracking productivity gains from Replicas at the team level? If managers are measuring and seeing improvement, budget allocation follows.

I think the infrastructure approach is the right bet. Agents will evolve and improve. The underlying need for sandboxed development environments will persist regardless of which agent is best. Replicas is building the container, not the contents, and that is a durable position.