M

How I Build

Context Engineering + Agentic Execution

How a solo founder operates at the output of a full engineering team. Not by working harder — by structuring information so AI agents can execute complex work with precision and minimal intervention.

60+
missions shipped
S10.5
current sprint
v27
schema version
L3.8
automation level
6 Levels Below

The Automation Levels

Level 1 — Manual

AI writes code, human integrates everything.

GitHub Copilot autocomplete, manual file management, no structure. Each file change required human context-switching between editor, terminal, and documentation. Useful for line-level suggestions, but no architectural awareness.

Level 2 — Assisted

AI edits files, human runs commands.

Windsurf for intelligent editing. VS Code Copilot for inline suggestions. Google’s Project Antigravity for multi-agent experiments. Each tool taught something — Antigravity showed the promise of agent orchestration, Windsurf proved that AI could manage multi-file edits. None delivered a structured, mission-driven pipeline for production code.

Key learning: The bottleneck was never code generation — it was context management.

Level 3 — Supervised Agentic

Claude Code executes missions. Human plans, approves, pushes.

The breakthrough. Claude Code shifted the paradigm from “AI assists a coder” to “architect directs an engineering team.” The core workflow crystallised: /start → /architect → /engineer → /summarise. Mission files became contracts — self-contained specs that agents execute line by line. The Golden Path rules emerged: package imports only, no Navigator.push, no setState, surgical migrations, dart analyze zero errors.

TrekMeet went from concept to Google Play Beta in this phase. 60+ missions shipped across 10 sprints.

Level 3.8 — Current

Hooks, state machine, sprint planning, multi-persona agents.

The system now has memory, governance, and specialisation. A tasks.json state machine tracks mission lifecycle. /sprint and /phase commands run full audits at boundaries — retrospectives, dependency checks, backlog cross-references. Four MCP servers give agents direct backend access to GitHub, Supabase, Firebase, and PowerSync. The agent team has expanded beyond a single engineer:

Opus

Orchestrator & AI Co-founder. Senior Solutions Architect who owns all tech decisions, writes mission files, runs sprint planning, and coordinates the full agent team.

Claude Code

Engineer. Implements missions autonomously, runs build gates, commits and reports.

UX Partner

Produces detailed UX specs. Reviewed by Opus, stored in docs, integrated into backlog.

Co-Pilot QA

Ingests screenshots and debug logs. Prepares structured reports for Opus triage.

GitHub
Supabase
Firebase
PowerSync
Level 4 — Sprint 11–12

Agents trigger each other. Human approves at gates only.

Agent SDK harness for chained execution. Co-Pilot QA agent detects a bug → Opus triages and writes a mission → Engineer agent implements → QA validates. The human reviews output at approval gates, not process. Supabase message queue connects agent sessions.

Level 5 — Sprint 13+

Full autonomous loop: plan, build, test, deploy.

The engineering pipeline runs end-to-end without human message bussing. Sprint planning, mission generation, implementation, testing, and deployment happen as a continuous loop. The founder reviews results and makes product decisions — the agents handle everything else. The trekmeet-agent-framework is extracted as a reusable toolkit — the methodology becomes transferable.

Level 6 — Solo Founder Operating System

Multi-department agent orchestration. One founder, zero staff, two hours per day.

The endgame is not just automated engineering — it is an automated organisation. Engineering, QA, UX, community moderation, support, and growth are each managed by specialised agents. The founder starts each day with a Command Centre briefing: overnight metrics, pending actions with confidence levels, agent recommendations. Review each department (10–15 minutes each), approve or adjust, and the agents execute.

Virtual Departments

Engineering

Backlog management, bug triage, sprint execution, build monitoring.

Product & UX

Feature proposals from user feedback, UX specs, A/B test analysis.

Community

Content moderation, group approvals, dispute escalation.

Growth & Support

User acquisition, App Store optimisation, auto-categorised enquiries.

Self-healing agents monitor production health. Hotfix proposals auto-queue. UX validation agents compare deployed screens against specs. Sprint retrospectives auto-generate from mission metrics. The framework itself becomes extractable and reusable — a repeatable system for any solo founder building at scale with AI agents.

Live Session

trekmeet — sprint 10.5
$ /start
Reading CLAUDE.md… Schema v27 confirmed
Reading PROJECT_STATE_SNAPSHOT.md… Sprint 10.5
Reading BACKLOG.md… 18 items queued
MCP: GitHub ✓ Supabase ✓ Firebase ✓ PowerSync ✓
✓ All caught up. Ready for mission.
$ /architect MISSION_046
Opus reads context → validates arch → scopes 3 files
# HARD STOP — awaiting CEO approval
$ /engineer –approved
Sonnet executes → hooks enforce → dart analyze 0 errors
flutter build apk –debug –flavor dev → BUILD SUCCESSFUL
✓ Commit ready. SHA: 995ffef
# Push review: git diff HEAD~1 → awaiting approval

Anatomy of a Mission

Every feature is a contract.

A mission file is the only context an executing agent receives. If something is ambiguous in the mission, the agent will guess wrong. Precision is the methodology.

Objective

One paragraph — what and why. No implementation detail here.

Architecture Decision

Pattern choices locked before execution begins. Agents must not freelance.

Infrastructure: CEO Parallel Work

Supabase SQL, dashboard config, RLS policies — runs while the agent builds.

Agent Steps (Numbered)

File path, exact changes, imports, before/after code blocks. Max 30KB or split.

Do Not Touch

Explicit list of files and patterns the agent must not modify. Prevents scope creep.

Verification Checklist

dart analyze zero errors, flutter build passes, manual test steps on device.

Risk Register

What could go wrong + mitigation. The agent knows its failure modes in advance.

This methodology shipped TrekMeet from zero to Google Play in 8 weeks.
The product is the proof. The framework is what scales.

© 2026 Manish Kumar. Cultivating focus, shipping software.  ·  hi@manish.pm