$ cat ./field-notes/wedge.mdx

Where the wedge is

field-notes · chapter 06 · market gaps and what a new product could own

What the research tells us about where the market is going, what the real opportunities are, and where Atelier could create something that doesn't exist yet. Synthesized from problem-space.md and competitive-landscape.md.

Last updated: 2026-04-13


## The Six Market Shifts

Every trend converges on the same set of shifts. The research is consistent across Anthropic's report, Addy Osmani's writing, RedMonk's analysis, and community patterns.

### Shift 1: Single Agent → Multi-Agent Teams

What's happening: The era of one human + one agent is ending. The new paradigm is one human orchestrating 3-5+ specialized agents working in parallel.

Evidence:

  • Claude Code Agent Teams (Feb 2026): shared task lists, peer messaging, dependency resolution
  • OpenAI Codex subagents (March 2026): manager agent + specialized workers in cloud sandboxes
  • Antigravity Mission Control: up to 5 parallel agents
  • Cline Kanban: autonomous multi-agent workflows from a single prompt
  • Jules: 60 concurrent tasks
  • Zapier: 89% AI adoption across entire org, 800+ agents deployed internally
  • TELUS: 13,000+ custom AI solutions, 30% faster code shipping, 500,000+ hours saved
  • Rakuten: Claude Code completed autonomous 7-hour task in 12.5M-line codebase with 99.9% accuracy
  • Anthropic data: engineers use AI in ~60% of work but fully delegate only 0-20% of tasks

What this means for Atelier: The orchestration layer is where the value is. Not writing code — coordinating agents that write code. Atelier's board is an orchestration surface.


### Shift 2: Chat Interface → Structured Workflows

What's happening: "Vibe coding" (prompt, iterate, hope) is hitting documented walls. The industry is moving toward structured approaches where specs and plans precede code.

Evidence:

  • Kiro: spec-driven development (requirements → design → tasks). AWS bet an entire product on this.
  • GitHub Spec Kit: 72,000+ GitHub stars. Open-source spec-driven toolkit.
  • Tessl Framework: another spec-driven toolchain.
  • Anthropic's data: 89% acceptance rate with diff summaries vs 62% raw output. Structure wins.
  • The emerging consensus: "Use specs for major changes, vibe coding for iteration."

What this means for Atelier: Atelier's card-type system IS structured workflow. Research cards structure the research phase. Backend cards structure the API design phase. This isn't just a UI gimmick — it's the spec-driven movement applied to the full development cycle, not just code.


### Shift 3: Session Amnesia → Persistent Memory

What's happening: The fundamental limitation of all AI tools — starting fresh every session — is being attacked from multiple angles.

Evidence:

  • Claude Code Auto Dream (March 2026): automated memory consolidation modeled after REM sleep
  • Windsurf Memories: implicit learning over 48 hours
  • agentmemory (open source): full memory pipeline with vector + knowledge graph indexing
  • Mem0, Letta/MemGPT: dedicated memory APIs
  • ETH Zurich study: generic context files hurt more than they help. Minimal, focused memory wins.
  • "Persistent memory is the fundamental infrastructure challenge separating AI tools from AI collaborators."

What this means for Atelier: Memory is the deepest moat. Auto Dream and Memories are steps forward but they're implicit — the system decides what to remember. Atelier's explicit, human-curated, promotable memory (card → project level) is a genuinely different approach. Nobody else has built visible, editable, phase-persistent project memory.


### Shift 4: IDE-Bound → IDE-Agnostic

What's happening: The best orchestration tools are decoupling from specific IDEs.

Evidence:

  • Cline Kanban: CLI-agnostic, works with any agent
  • T3 Code: BYOK, works with Codex and Claude (more coming)
  • Cloud agents (Codex Cloud, Jules, Copilot Coding Agent): no IDE required
  • Claude Code: terminal-native, IDE-independent
  • MCP: universal protocol connecting any tool to any agent

What this means for Atelier: Atelier should NOT be an IDE. It should be a standalone orchestration and oversight layer that works with whatever agents and tools people already use. This is the T3 Code insight, extended beyond the building phase.


### Shift 5: Code Generation → Full-Cycle Assistance

What's happening: The industry is saturating on "write code faster." The frontier is moving upstream (research, design, planning) and downstream (testing, deployment, iteration).

Evidence:

  • Kiro: specs before code (upstream)
  • Figma AI: agents designing on canvas (upstream)
  • Jules: 60 concurrent background tasks (downstream — maintenance, tech debt)
  • Devin: migrations, upgrades, cleanup (downstream)
  • Linear: "issue tracking is dead" — AI workflows for the full product development lifecycle
  • Anthropic Trend #7: non-engineers using agents — the cycle extends beyond engineering

What this means for Atelier: This is Atelier's biggest strategic opportunity. Every competitor is fighting over the building phase. Atelier can own the full cycle: research → design → build → test → iterate. The card-type system enables this — each card type is a phase-specific surface.


### Shift 6: Developer-Only → Cross-Functional

What's happening: AI tools are expanding beyond software engineers to designers, PMs, marketers, and founders.

Evidence:

  • Anthropic Trend #5 & #7: "Sales, legal, marketing, and operations teams are using agents to solve local process problems without waiting on engineering queues."
  • Non-technical founders using Lovable/v0/Bolt hit the wall at month 3 — demand for better tools exists
  • 51% of Figma users working on AI products are building agents
  • Linear positioning as "workflows shared by humans and agents, from drafting PRDs to pushing PRs"

What this means for Atelier: The PM-clarify card type, the design-flow card type — these aren't developer features. They're cross-functional features. Atelier could be the surface where a founder, a designer, and an agent collaborate on a product, not just where a developer codes.


## The Three Moats Available to Atelier

After mapping the full landscape, three defensible positions emerge:

### Moat 1: Full-Cycle Coverage (Research → Ship)

Why it's a moat: Everyone is fighting over the building phase. The research phase, design phase, and testing phase are dramatically underserved. No tool spans the full cycle.

Competitive defense: Cline Kanban and T3 Code would need to build entirely new surfaces for research, design, and testing. That's a product direction change, not a feature addition. They've optimized for "kanban for code agents" — pivoting to "studio for the full project lifecycle" is structurally different.

Risk: Scope is enormous. The moat is also a trap if you try to do everything at once.

### Moat 2: Explicit, Human-Curated Project Memory

Why it's a moat: Memory is recognized as the defining challenge of 2026. Every approach so far is implicit (Auto Dream, Memories) or static (CLAUDE.md). Explicit, promotable, phase-persistent memory that a human curates is a genuinely new paradigm.

Competitive defense: Building a real memory system requires deep product thinking about what to remember, how to present it, when to surface it, and how to let users curate it. This is design engineering work — exactly your skillset. Cline and T3 Code are engineering-led tools that would approach memory as a data problem, not a UX problem.

Risk: Memory is hard to get right. Bad memory (too much, too noisy, stale) is worse than no memory. The curation UX is load-bearing.

### Moat 3: Design Taste as Product Differentiator

Why it's a moat: "When average output becomes easier to generate, taste becomes more valuable." (Figma 2025 AI Report). Every competitor is built by engineers for engineers. Atelier, built by a design engineer, can be the tool where the product itself demonstrates the taste it helps users achieve.

Competitive defense: Taste can't be replicated by adding features. It's in the interaction design, the animation quality, the spatial reasoning of the interface. This is the Rauno/Emil/Linear moat applied to the developer tools space.

New data supporting this moat:

  • Figma 2025: 78% say AI makes them efficient, only 47% say it makes them better. 31-point gap. Only 32% of designers trust AI output.
  • Vercel pays design engineers $196K-$294K (SF base). The craft is valued at the highest level.
  • Rauno Freiberg (Staff Design Engineer, Vercel): works at "the intersection of design and engineering" — created cmdk (millions of downloads/week), Devouring Details (20-chapter interactive manual). Proof that craft-in-public builds reputation without a company title.
  • 52% of designers report design becomes MORE important for AI products (Figma 2025)
  • 85% of designers consider AI skills essential to future success

Risk: Taste alone doesn't sell software. It must be combined with functional superiority (moats 1 or 2).


## The Spec-Driven + Memory Convergence (Atelier's Unique Position)

The most interesting strategic insight from the research is that two emerging patterns — spec-driven development (Kiro's approach) and persistent memory (the Auto Dream/Memories direction) — are converging but nobody has connected them yet.

Spec-driven development says: write the requirements and design before writing code. The spec is the source of truth.

Persistent memory says: carry context across sessions and phases. Decisions and reasoning should persist.

The convergence nobody has built: A system where:

  1. Research produces structured findings (phase memory)
  2. Findings feed into requirements and design specs (spec-driven)
  3. Specs guide agent execution (structured building)
  4. Decisions at each stage are recorded as memory (persistent context)
  5. Memory feeds back into future specs and future building (learning loop)

This is what Atelier's card system + memory system could be. Each card type is a phase-specific spec surface. Memory flows between cards. The spec and the memory are the same thing at different levels of abstraction.

No competitor has this. Kiro does specs but not memory. Auto Dream does memory but not specs. Cline Kanban does orchestration but neither specs nor memory. T3 Code does multi-project management but neither specs nor memory.


## The "Project as Container" Opportunity

One problem is completely unsolved in software: project-level isolation.

A software project = git repo + dev server + browser profile + terminal + editor workspace + environment variables + localhost port + database. Currently, these are scattered across the OS with no grouping.

What exists:

  • Dev containers (Docker-based isolation) — powerful but heavyweight
  • Git worktrees (file-level isolation) — used by Cline Kanban and T3 Code
  • Browser profiles (Chrome/Firefox) — manual, not developer-tool-integrated

What nobody has built:

  • One-click project switching that changes: terminal, browser profile, dev server, port, environment, editor workspace — all at once
  • Port management: automatic port allocation per project, no collisions
  • Cookie/auth isolation: each project gets its own localhost origin
  • Ambient awareness: which project's agent just finished?

Atelier's "Capsules" concept from the overview.md addresses this, but it's listed as "Layer 3 — not in v1." The research suggests it's more important than originally assessed — multi-project fragmentation is one of the top 3 pain points across all user segments.


## The Design Phase Gap (Atelier's Potential Unique Moat)

The research reveals something striking: every tool in the landscape is engineering-first. None center design.

| Tool | Engineering | Design | |------|-----------|--------| | Cursor | Deep code editing | None | | Claude Code | Powerful reasoning | None | | Cline Kanban | Task orchestration | None | | T3 Code | Multi-repo management | None | | Kiro | Spec-driven building | Specs include data models, not UX | | v0 | Component generation | Generates UI, doesn't do design thinking | | Figma AI | Design assistance | Doesn't connect to building workflow | | Linear | Issue tracking | Viewing code diffs, not design thinking |

The gap: No tool helps with the design THINKING that should happen between "research findings" and "code." No tool structures:

  • User flow design (not just individual screens)
  • Interaction specification (timing, easing, spatial continuity)
  • Design system compliance checking
  • Design decision documentation ("we chose this layout because...")
  • The translation of design intent into agent-consumable constraints

Why this matters for Atelier specifically:

  1. The founder IS a design engineer — this is the authentic skill edge
  2. The "taste gap" is confirmed by every source (Figma report, NN/g, industry consensus)
  3. Design thinking as a product feature can't be replicated by adding an AI feature — it requires deep product understanding
  4. The design phase connects research (upstream) to building (downstream) — owning it means owning the full-cycle handoff

## What Users Actually Pay For (The Value Equation)

From the pricing research, clear patterns emerge:

What makes users pay:

  • A $50/month tool needs to save at least 4 hours of work
  • "Close the loop" — tools that advise but don't execute live in dangerous territory
  • Vertical specialization beats horizontal generics
  • Measurable business impact, not soft ROI

The free competitor problem: Cline Kanban is free. T3 Code is free (BYOK). Cursor charges $20/mo because it provides enough value over free VS Code.

For Atelier to justify paid pricing, it must deliver value that free tools don't:

  1. Full-cycle coverage (free tools are building-only)
  2. Persistent project memory (free tools are session-amnesiac)
  3. Design-aware surfaces (free tools are engineering-only)
  4. Environment isolation (free tools don't do this)

The value proposition isn't "manage your agents" (Cline Kanban does that for free). It's "manage your entire project — from idea to iteration — with memory, structure, and taste."


## Risk Assessment

### What could kill Atelier before it ships:

  1. Cursor/Windsurf adds project-level features. If Cursor adds multi-project management, memory, and task orchestration, the window closes. Cursor has $2B ARR and massive engineering resources.

    • Mitigation: Cursor is an IDE. IDEs are editing environments. Atelier is a project studio. The frame is different. But the risk is real.
  2. Cline Kanban adds card-type-specific UIs. If Cline adds research surfaces, design surfaces, and memory, it's a free alternative.

    • Mitigation: Cline is engineering-led. Adding design-quality surfaces is a product direction change they're unlikely to make. Their DNA is terminal-first.
  3. The scope is too large. Atelier's vision touches every phase of the development cycle. Trying to build everything results in building nothing.

    • Mitigation: Start with 2 card types (research + backend-endpoint) as the PRD specifies. Prove the morphing bet. Expand after validation.
  4. Memory is too hard to get right. Bad memory is worse than no memory. Stale facts, noisy surfaces, curation fatigue.

    • Mitigation: Start manual and minimal. Let users promote facts explicitly. Don't try to auto-generate memory.
  5. "Project as container" requires native OS access. Port isolation, browser profiles, ambient presence — these are desktop-app features that a web app can't do.

    • Mitigation: Start web-first for the studio UI. Capsules (native layer) comes in v2 only after the studio proves its value.

## The Decision Framework for Part 5 (problem-space.md)

After all this research, here are the questions Part 5 needs to answer:

  1. Primary segment: Solo builder? Startup team? Agency? (Evidence points to solo builder as the beachhead — smallest sales cycle, most pain, founder is the user.)

  2. Primary phases: Which 2-3 phases does Atelier own first? (Evidence points to: Research + Building + the handoff between them. Design is the medium-term moat.)

  3. Relationship to existing tools: Does Atelier replace or complement? (Evidence points to complement: Atelier orchestrates agents from Claude Code/Codex/Cursor, doesn't replace them.)

  4. Memory model: Explicit human curation (Atelier PRD), implicit automated (Auto Dream), or hybrid? (Evidence points to hybrid: auto-suggest + human curate.)

  5. Pricing position: Free (compete with Cline Kanban), freemium, or paid-only? (Evidence points to freemium: free for 1 project, paid for memory + multi-project + morphing.)

  6. Architecture: Web app? Desktop app? CLI tool? (Evidence points to web app first, desktop app later for Capsules.)

  7. MCP-native? (Yes — architecturally load-bearing. The substrate should be MCP-addressable from day one, as the original overview.md specifies.)



## The Spec-as-Memory Insight (from Addy Osmani)

Osmani's research on writing specs for AI agents reveals something directly applicable to Atelier's architecture.

The 6-area framework (from analysis of 2,500+ agent configuration files): Commands, Testing, Project Structure, Code Style, Git Workflow, Boundaries.

The three-tier boundaries system: Always Do (safe to execute) → Ask First (needs human approval) → Never Do (hard stops).

The key finding: "Most agent files fail because they're too vague." And: "As you pile on more instructions, the model's performance in adhering to each one drops significantly." (The "curse of instructions.")

What this means for Atelier: Each card type should generate its own minimal, focused spec — not a giant document. A research card produces a research spec (sources to explore, synthesis format). A backend card produces an API spec (endpoints, schemas, constraints). These card-level specs ARE the memory — they're structured, they carry forward, and they're minimal enough to be effective.

This connects to Kiro's approach but goes further: Kiro generates specs once at the start. Atelier's cards generate living specs that update as work progresses. The spec IS the card's memory. The project memory is the union of all card specs.


## Enterprise Precedent: EY.ai PDLC (March 2026)

Ernst & Young launched EY.ai PDLC in March 2026 — an AI-native approach to software delivery that replaces traditional linear development with an AI-driven model. It orchestrates a "collaborative mesh" of AI agents with human oversight across the full software lifecycle — from requirements and architecture through code, testing, infrastructure and operations.

Why this matters for Atelier: It proves that the "full lifecycle, agent-orchestrated, human-curated" concept is real — a Big 4 consulting firm built their version for enterprise. Atelier is the same concept for solo builders and small teams, with design taste and visible memory instead of enterprise compliance.

The existence of EY.ai PDLC validates the market direction. Atelier doesn't compete with it — they serve entirely different segments. But the fact that Ernst & Young invested in building this confirms that "full lifecycle AI orchestration" is a real category, not a theoretical concept.


## The Linear Threat and Opportunity (March 2026)

Linear declared "issue tracking is dead" and is pivoting to agent orchestration. 25% of new issues in enterprise workspaces are now created autonomously by agents. Coding agents installed in 75% of enterprise workspaces. Workload volume up 5x in three months.

Why this matters:

  • Linear starts with issues → adds agents. Atelier starts with agents → adds project structure. They're converging.
  • If Linear ships a compelling coding agent + MCP integration, it could absorb Atelier's orchestration value.
  • Basecamp is making the same move ("agent-driven" with CLI access for any AI agent). This is an industry trend.

The mitigation: Atelier's differentiation is in the phases Linear won't touch — research, design, memory, and the craft layer. Linear will never be a design tool. It will be a strong orchestrator for the build phase but won't solve the full cycle.

The opportunity: Linear's pivot proves the "agent orchestration hub" category is real. The market expects this tool to exist. Atelier's bet: the best version of this for solo builders and small teams is design-first and full-cycle, not text-first and build-only.

NEW: Cursor Glass threat (April 2, 2026): Cursor 3 "Glass" rebuilt the IDE around an Agents Window — orchestration as the primary surface, not the editor. Cloud agents in isolated VMs. Multi-repo layout. Cloud Handoff maintains context between local and cloud. This is the most direct competitive threat because Glass explicitly moves from "IDE" to "orchestration surface." Mitigation: Glass is still repo-centric and build-phase-only. No research, no design, no memory curation, no card-type morphing. And $200/mo for Ultra.

NEW: Claude Managed Agents opportunity (April 8, 2026): Anthropic launched infrastructure-as-a-service for agents — sandboxed execution, multi-agent coordination, long-running sessions, $0.08/agent runtime hour + model usage. This is potentially Atelier's most important integration: dispatch tasks from Atelier's board → Managed Agents executes → Atelier curates results and feeds memory. Intent, oversight, and memory on Atelier's side; execution on Anthropic's infrastructure.


## The Agentic UX Playbook (Emerging, April 2026)

Six patterns are crystallizing for how to design agent-first products. These directly inform Atelier's card-type system:

| Pattern | What it does | Atelier application | |---------|-------------|-------------------| | Intent Preview | Agent shows plan before acting | Each card type IS an intent preview — morphed to the task | | Autonomy Dial | Per-task-type autonomy levels | Per-card-type autonomy: research auto, backend manual | | Explainable Rationale | "Because you said X, I did Y" | Memory rail provides rationale source | | Confidence Signal | Agent surfaces certainty | Card glow maps to confidence (amber = uncertain) | | Action Audit & Undo | Chronological logs with undo | Board IS the audit log; git-backed undo | | Escalation Pathway | Agent recognizes uncertainty, asks | Blocked-on-human cards in board view |

Key insight: "Autonomy emerges from technical capability, but trustworthiness stems from intentional design." The design taste that Atelier brings IS the trust-building mechanism. This reframes Atelier's design advantage from aesthetic to functional — good design creates the psychological safety that enables delegation.

See deep-dive-agentic-ux.md for the full pattern analysis.


## The Runtime Isolation Gap (Confirmed, April 2026)

The "project-as-container" opportunity from the original strategic analysis is confirmed as deeper than assumed. Research reveals six specific collision types that git worktrees don't solve: port conflicts, shared databases, bind-mount collisions, secret leakage, browser state contamination, and log ambiguity.

The competitive landscape for isolation:

  • Coasts: most complete model (Coastfile + per-service strategies + dynamic ports)
  • Container Use (Dagger): containerized sandboxes per agent
  • Google Scion: manager-worker with isolated containers
  • Worktree CLI (AirOps): isolated environments on demand
  • Nobody has shipped all five dimensions: ports + services + data + secrets + observability

The multi-agent ceiling: "3-4 parallel agents is a practical ceiling when a single reviewer integrates results." Gains are bounded by review throughput, not generation speed. This means Atelier's board — which makes review and oversight the primary UX — is solving the actual bottleneck, not the perceived one (generation speed).

The Quality Gates Model (Osmani): Three mandatory safeguards:

  1. Plan Approval — agents write implementation plans before coding; leads reject architectural mistakes before they compound
  2. Hooks on Task Completion — automated checks (lint, tests) fire before task state changes; failed hooks force continued work
  3. Loop Guardrails with Reflection — hard MAX_ITERATIONS=8; before retry, agents must answer "What failed? What specific change fixes it?"

Plus: a dedicated @reviewer teammate (Opus-class model, read-only) triggered on every TaskCompleted event. 1 reviewer per 3-4 builders. This is the verification architecture Atelier's card-type system should implement — each card has built-in quality gates that match the task type.

Frontier model accuracy: 70%+ on single-issue tasks, below 25% on multi-file patches. Task decomposition quality is the difference between these numbers. Atelier's card-type system enforces bounded, testable units — which is the prerequisite for agent success.

See deep-dive-runtime-isolation.md for the full analysis.


## The Agentic Engineering Paradigm (Validated, April 2026)

Andrej Karpathy officially declared vibe coding "passé" in early 2026, coining "agentic engineering" as the successor paradigm. This isn't just a label — it's the structured methodology that validates Atelier's core thesis.

The four practices of agentic engineering:

  1. Context Engineering — curate what agents see (CLAUDE.md, card-level specs)
  2. Task Decomposition — break features into single-session chunks (card-level tasks)
  3. Verification Loops — write failing tests first; agents iterate until they pass
  4. Checkpoint Discipline — commit working states before risky changes; rollback beats fixing forward

Critical new data:

  • Anthropic's multi-agent system (Opus lead + Sonnet specialists) outperformed single-agent Opus by 90.2% through context isolation — each agent maintains only task-relevant information
  • System-level success rate: 80-87% even when individual agent success is 95-98% — coordination is where failures compound
  • C Compiler case study: 16 agents hit same non-parallelizable bug, all produce identical fixes, $20K wasted on 100K Rust lines
  • CodeScene: Teams with Code Health 9.5+ see 2-3x agentic productivity gains — clean architecture enables delegation
  • Only 3% of developers "highly trust" AI-generated code despite 84% using it

What this means for Atelier:

  • Card types are context engineering surfaces (validated by the paradigm)
  • Card-level task decomposition is not just UX — it's the prerequisite for agent success (70% vs 25% on bounded vs unbounded tasks)
  • Quality gates per card type = verification loops built into the workflow
  • Git-backed card history = checkpoint discipline at the card level
  • The product IS the agentic engineering workflow, visualized

The economics to know:

  • Three-agent pilot: $5-50/session
  • Production three-agent workflow: $18,000-90,000/monthly
  • Solo founders won't hit these numbers, but they need cost visibility per card/feature

## The Solo Founder as Primary User (Confirmed, April 2026)

The "solo builder" segment from the problem-space map is confirmed as the right beachhead. Key data:

  • 36.3% of 2026 ventures are solo-founded
  • Solo founders replace headcount with AI subscriptions ($200-500/month), achieving 10-50x higher capital efficiency
  • 60,000+ GitHub repos include agent instruction files (CLAUDE.md, AGENTS.md)
  • Gartner: 1,445% surge in enterprise inquiries about multi-agent orchestration
  • The "Vibe CEO" model: 80% strategy/judgment, 20% execution — enabled by agents

The workflow Atelier replaces:

  1. Claude/ChatGPT for strategy (no project continuity)
  2. Perplexity for research (no connection to build phase)
  3. Cursor for development (no research context)
  4. Linear/Notion for tracking (no agent integration)
  5. Manual CLAUDE.md maintenance (context engineering overhead)

Atelier collapses this into: research → design → build → test in one workspace with persistent memory. The context engineering that solo founders do manually (CLAUDE.md, .cursorrules) becomes automatic through card-level specs.

Price ceiling: $0-$150/month. Free tools handle 80% of needs. Atelier must justify paid pricing through time savings on context management, project switching, and review overhead.


## Sources

All sources are cited in problem-space.md and competitive-landscape.md. Key sources for this synthesis:

$
? keys · ⌘K palette · esc back