← February 5, 2026 edition

github-agent-hq

Run Claude, Codex & Copilot directly in GitHub & VS Code

GitHub Stops Picking Sides in the AI Coding War

GitHub Stops Picking Sides in the AI Coding War

The Macro: Everyone’s Building the Same Editor, Nobody’s Building the Layer Above It

The AI coding assistant market has spent two years in a very productive argument with itself. Cursor built a standalone editor that ate VS Code’s lunch by wrapping it in smarter autocomplete and agentic features. Anthropic shipped Claude Code. OpenAI shipped Codex. Each positioned as the thing you should probably just switch to. The implicit assumption: one model, one workflow, one winner.

That assumption is getting complicated. According to the Stanford HAI 2025 AI Index Report, 78% of organizations reported using AI in 2024, up from 55% the year prior. Anyone working inside a mid-sized engineering org can confirm that number anecdotally. The pressure isn’t whether to use AI coding tools anymore. It’s which ones, when, and how to stop them from living in twelve different browser tabs.

The field is genuinely crowded. Cursor has momentum and a dedicated user base that treats it like a religion. Claude Code has strong reviews for complex, multi-file reasoning tasks. Codex CLI exists for people who prefer their terminal. Copilot has the distribution advantage of being already inside the tools most enterprise developers have open all day. No one has decisively won.

That’s not a sign the market is broken. It’s a sign it’s still early.

The real question isn’t which model wins. It’s whether the fight is between models or between platforms. GitHub, which sits at the center of where code actually lives rather than just where it gets written, is making a quiet bet that the platform layer is the durable position. Agent HQ is the product expression of that bet.

The Micro: The Depot Where Your Agents Report For Duty

GitHub Agent HQ, launched into public preview in early 2026, does something structurally different from what most AI coding tools do. Instead of asking you to choose a model, it lets you run Claude, OpenAI Codex, and GitHub Copilot simultaneously, inside GitHub itself and inside VS Code, with Copilot CLI support listed as coming soon.

Access is gated: you need a Copilot Pro+ or Copilot Enterprise subscription. Not a casual price point. But it’s also not aimed at the solo developer who’s happy with a free tier. The target is the team or enterprise already paying for Copilot that wants to stop context-switching between tools every time a different model might do the job better.

What Agent HQ is actually solving for is orchestration friction.

According to the GitHub blog post announcing the feature, the promise is that context, history, and code review stay attached to the work. Running an agent doesn’t mean leaving GitHub’s native flow to poke at something in a separate interface. The agents communicate status back to wherever the workflow started. That’s a real problem for anyone who’s tried to coordinate an agentic coding task across three windows and a browser tab.

Coverage in The New Stack and Pragmatic Engineer’s newsletter notes that Cursor is increasingly resembling a GitHub competitor rather than just an IDE replacement. That gives GitHub competitive reason to accelerate here beyond simple feature completeness.

It got solid traction on launch day on Product Hunt, though enterprise-facing infrastructure products don’t typically dominate those leaderboards. Public preview means it’s real, not vaporware. It also means it’s not finished.

The Verdict

The structural logic here is sound. GitHub has what no standalone AI coding tool has: the repo. The pull request. The issue. The place where code actually goes after someone writes it. Building an agent orchestration layer on top of that is a more defensible position than building better autocomplete, because the platform lock-in is real and the workflow integration is genuine rather than cosmetic.

What I’d want to know before getting too enthusiastic is whether the multi-agent comparison actually surfaces meaningfully different outputs in practice, or mostly just adds steps. Letting three models run at your problem sounds powerful. Making that useful requires real UX work that a public preview may not have finished yet.

The 30-day test is whether the Copilot Pro+ subscriber who tries this once actually builds it into their daily loop. The 90-day test is whether enterprises already comfortable paying for Copilot Enterprise see this as a reason to deepen that relationship, rather than just hand developers a Cursor license and call it done.

I think this is probably the right product for teams already embedded in GitHub who want orchestration without adding another tool to manage. I don’t think it moves the needle much for developers who’ve already built their workflow around Cursor or Claude Code and see GitHub as storage, not home base. The bet GitHub is making is that developers want their agents to live where their code lives. That’s not a crazy bet. It’s just not proven yet.