Small group
A handful of teams. Direct line to the people building it.
The PR shows the result. It doesn’t show how the code was produced. We’re working with a small number of engineering teams to test whether session-based context solves that problem in practice — before building further.
The problem you’re already in
Most of the reasoning behind an agent-authored change — the intent, the tool calls, the attempts, the rejected paths — lived in a session that closed the moment the PR opened. The diff shipped. The context didn’t. And the gap widens as agent usage grows, not the other way round.
The proposed answer
This is the kind of artifact we believe solves the problem. It already exists as a product direction — the partnership is how we find out whether it holds up against your real workflows.
About the partnership
Not a pilot of a finished product. A small-group engagement to test a specific hypothesis — that session-based context meaningfully changes how agent-written code gets reviewed, debugged, and audited — before we build further. You get early say in how the thing is shaped. We get real ground truth.
A handful of teams. Direct line to the people building it.
Against your repos, your agents, your review process.
No commitment beyond the engagement. You keep what you learn.
How it works
A deliberate, focused sequence. Weeks, not quarters. Designed so both sides learn something real before either of us commits further.
A short working session with your team. We map out where agent-written code is showing up, where review is struggling, and what “good enough context” would actually look like for you.
We walk through real agent-authored changes in your environment — PRs where reviewers got stuck, incidents where provenance mattered. No abstractions, no slideware.
We capture sessions against a slice of your workflow and see where the extra context actually changes a review, a debug path, or an audit conversation.
We use what we learn together to validate product direction. If it’s working, we keep going. If it’s not, we learn why — and you keep the notes.
A concrete example
A reviewer questions a rate-limit value on an agent-authored PR. Because the session is linked to the commit, the reviewer, the author, and the agent itself can all speak to the same record — not a transcript that vanished into Slack.
Maya
reviewer · 14:14Are we sure 50 rps is the right ceiling? Our analytics showed 38 rps p95 last week but there are bursts.
Claude Code
agent · 14:15The intent said “cap around 50 rps”, so I kept headroom above p95. I tried 40 first and backed out — not enough slack for the burst window. Happy to tighten to 45 if you want a follow-up session.
Abigail
author · 14:16Good — that’s the reasoning I needed to see. Leaving it at 50 and watching for a week. Merging.
Who this is for
Not everyone is ready for this conversation yet. The partnership works best with teams already living the pain.
Next step
A 30-minute discussion about what your team is seeing, what we’re building, and whether a design partnership makes sense. No pitch deck.