Engineering Notes

Engineering Notes

Thoughts and Ideas on AI by Muthukrishnan

What If GitHub Was Built for AI Agents

17 Feb 2026

GitHub was built for humans reading diffs. But what happens when most of the code is written, reviewed, and merged by AI agents?

You do not just add an AI layer on top of Git. You rethink the entire system around how agents operate: intent, prediction, simulation, reputation, and policy.

Here is what an agent native source code management system could look like.

Intent Based Commits

Agents do not just commit code. They commit intent.

Instead of a message like fix bug in auth, you get structured metadata:

The system stores outcome versus intent. Over time it learns which agents are good at which types of changes. Commit history stops being a log of what happened and starts becoming a training signal for what works.

Continuous Branching

No long lived branches.

Every agent runs in a short lived micro branch created per task. Branches auto expire. Merged branches become immutable snapshots.

Think of it as event sourcing for code, optimized for machines. There is no stale branch problem because branches are disposable by design. The unit of work is not a branch. It is a task.

Simulation Before Merge

Before merging, the system spins up:

Agents review structured metrics, not UI diffs. Merge happens only if model predicted stability crosses a threshold.

This is the fundamental shift. Humans review code by reading it. Agents review code by running it against reality.

Agent Reputation Layer

Each agent has a live reliability score based on real metrics:

A routing engine assigns critical tasks to high trust agents. New or low trust agents get sandboxed work. Reputation is earned, not configured.

Auto Rollback Contracts

Instead of humans deciding when to rollback, you define contracts upfront:

Rollback is automatic. No meeting. No Slack thread. No on-call engineer waking up at 3 AM. The contract executes and the system restores the last known good state.

Prompt and Model Version Graph

In an agent driven system, prompt changes are first class citizens. You can diff:

The graph shows how agent behavior evolved over time. When a regression appears, you do not just ask “what code changed.” You ask “what prompt changed, what model version was swapped, and what tools were reconfigured.”

Multi Agent Conflict Resolution

When two agents propose different solutions to the same problem, the system does not block and wait for a human. It can:

Humans only intervene if confidence drops below a threshold. Most conflicts resolve themselves through measurement, not opinion.

Deterministic Replay Mode

Every change can be replayed with:

This makes debugging agent behavior far more precise than reading PR comments. You do not ask “why did the agent do this.” You replay the exact conditions and observe.

Policy as Code for Agents

Organization rules become machine readable constraints:

Agents are evaluated against policies before merge, not after. Compliance is not an audit trail. It is a gate.

Fleet Dashboard Instead of Repo View

The main UI is not files and folders. It is:

Humans supervise. Agents operate. The interface reflects that shift.

The Center of Gravity Moves

If GitHub was built for humans reading diffs, an agent native system would be built around:

The question is not whether agents will change how we manage code. The question is whether we will keep forcing them into systems designed for humans, or build something native to how they actually work.

The center of gravity shifts from reviewing code to managing autonomous change.

● Intelligence at Every Action

AI Native
Project Management

Stop using tools that bolt on AI as an afterthought. Jovis is built AI-first — smart routing, proactive monitoring, and intelligent workflows from the ground up.