Now in development

Write the spec.
Walk away.
Come back to results.

Pilot is the infrastructure for autonomous AI development — async execution, enforced pipelines, a fleet of agents running on a server while you focus on something else entirely.

Join the waitlist No spam · Just a heads-up when it ships
pilot — workspace
Pilot
Hello world test
code
blocked
Shutdown test
review
blocked
Refactor auth service
code
done
Payment webhook handler
plan
running
Write tests for user service
spec
draft
Update API docs
spec
draft
Payment webhook handler
planning
14m 32s · 8,241 tokens · $1.84
spec
plan
test
code
review
verify
artifacts
Payment Webhook Handler
Implement a Stripe webhook handler that verifies signatures, processes payment events, updates order status in the database, and sends confirmation emails using the existing mail utility.
1. src/webhooks/stripe.ts
2. src/orders/service.ts
3. src/mail/templates/payment.ts
[ ]Signature verified on every incoming request
[ ]Order status updated on payment.succeeded
[ ]Confirmation email sent with order details
[ ]Idempotent — safe to replay events
specspec
planplan
Plan: Payment Webhook Handler
Summary
Implement a Stripe webhook handler that verifies signatures, processes payment events, and sends confirmation emails.
Files to Modify
1. src/webhooks/stripe.ts
2. src/orders/service.ts
3. src/mail/templates/payment.ts
Acceptance Criteria
[ ]Signature verified on all requests
[ ]Order status updated on success
[ ]Confirmation email sent
[ ]Idempotent replay-safe

The journey to autonomous AI coding.

Every developer using AI today has gone through some version of this progression. Each stage unlocks new leverage — and reveals the next ceiling.

Stage 01
Prompting
Where most of us started. Describe something in a chat window, get code back, copy it into the editor. You're the pipeline — fast for small things, exhausting for anything bigger.
Stage 02
Pair programming
AI moves into the editor. Inline completions, suggestions on every line. Real speed — but you're still driving. Every keystroke still needs you at the wheel.
Stage 03
Delegating
Describe a whole task, walk away, come back to working code. Extraordinary leverage — until you notice you're re-explaining the rules every session and the agent stops the moment you close your laptop.
Stage 04
Orchestrating
Specs queue up. Agents run in parallel on a server, following an enforced pipeline, whether you're watching or not. You check results when you're ready — after the meeting, in the morning, on your phone.
← this is Pilot

How orchestration actually works.

Pilot is built on five infrastructure primitives — the ones that turn AI coding from a productivity tool into a development platform.

01 — Async

Works while you sleep.

Queue a spec, close your laptop, and come back to results. When agents run on a server — not your machine — they report back when they're done. Your schedule and the agents' schedule are fully decoupled.

02 — Fleet

Many agents, one queue.

Multiple specs execute in parallel. One task or ten — the queue handles it. Each agent works in an isolated environment, so they never interfere with each other. Submit from your phone, check results over coffee.

03 — Enforced

The pipeline runs itself.

Every task follows the same path: spec → plan → code → review → verify. The pipeline is fully configurable — define the steps that fit your workflow. The system enforces them, not the LLM's memory. Proof is collected at each step, every time, regardless of what the agent decides. Each agent runs in a sandboxed container, isolated from your environment and from each other.

04 — Infrastructure

Server-based, always on.

Runs on infrastructure you control — a server, or your own laptop while you're getting started. Remote-accessible from anywhere. No terminal babysitting, no dropped sessions. The work continues whether your lid is open or closed.

05 — Models

The right model for the right job.

Pilot is model-agnostic by design. Use a powerful model for complex reasoning, a faster one for planning, a local model for anything that can't leave the building. Mix and match across the same pipeline — each task can use a different provider if that's what it calls for. Swap models anytime without changing your workflows. Bring Claude, GPT-4, Gemini, Mistral, or run fully local via Ollama. The orchestration layer doesn't care — it just runs the right tool for the job.

Run it yourself or let us run it.

Open source · Free

Framework

Self-host the full orchestration stack on your own infrastructure. MIT licensed, no lock-in, full control.

  • Full source, MIT licensed
  • Async agent queue with enforced pipelines
  • Sandboxed execution in Docker
  • GitHub integration built in
  • Runs on your server, or your laptop — your rules
  • For developers who want full control
Hosted · Paid · Coming soon

Team Edition

We host the infrastructure. Your whole team gets orchestration without spending a sprint on DevOps.

  • Zero infrastructure to manage
  • Shared queue for your whole team
  • Agent fleet scaling included
  • Usage metering and cost controls
  • Managed review pipelines
  • For teams who want it running today

Get it first.

The Framework ships when it can build itself. The hosted Team Edition follows. Tell us which you're after — it shapes what we prioritise.

No marketing sequences. Just a heads-up when it ships.

Andrew Crookston

Built in public by Andrew Crookston — engineering leader writing about AI coding at andrewcrookston.com.