Skip to content
AI Primer
workflow

Practitioners report harness playbooks with Playwright CLI, create_agent, and MCP

Builders shared concrete Symphony, create_agent, and MCP setup guides after arguing that model switching is easy but harness switching is not. The playbooks matter because they make harness engineering more repeatable, so teams can copy tested tooling and integration patterns.

4 min read
Practitioners report harness playbooks with Playwright CLI, create_agent, and MCP
Practitioners report harness playbooks with Playwright CLI, create_agent, and MCP

TL;DR

You can read Symphony's spec, skim Cursor's writeup on continuously tuning its harness, and browse LangChain's create_agent docs. The weirdly practical bit is that the playbooks all rhyme: repo-local workflow files, explicit tool wiring, and context management rules that live outside the model.

Symphony

The clearest practitioner recipe this week came from jasonzhou1993, who said his Symphony setup hinges on three pieces:

  • Playwright CLI
  • a boot skill
  • WORKFLOW.md

That lines up with the official Symphony README and SPEC. Symphony is a long-running service that watches an issue tracker, creates a separate workspace per issue, and keeps the workflow policy in WORKFLOW.md, versioned with the repo. OpenAI's own repo says the proof of work is not just a diff, it can include CI status, PR review feedback, complexity analysis, and walkthrough videos.

The community wrapper simphony, which daniel_mac8 pitched as a simpler setup, pushes the same idea even harder: ask your agent to install and configure it, and do not hand-edit WORKFLOW.md unless recovery is required. That is harness engineering turning into ops hygiene.

create_agent

Vtrivedy10's main claim was that LangChain's bigger Deep Agents stack is just an extension of a small base primitive. In his list, the extras are:

  • filesystem tools and bash
  • compaction and context offloading
  • subagents
  • skills and memory
  • hooks

LangChain's docs back most of that up directly. create_agent is the production-ready entry point, it runs on a graph-based runtime, and its extension points are tools, model selection, and middleware.

The more useful detail came from Vtrivedy10's context manager thread, which treats the harness as the thing that decides what survives when the context window fills up. His four concrete strategies were truncate, summarize and restart, summarize and offload, and selectively offload large tool results. That is the same class of problem Cursor's harness post describes from the other direction: move away from bloated static context, instrument changes with online and offline evals, and measure whether the agent actually keeps more useful code.

MCP

r/ClaudeCode

Claude Code Visual: hooks, subagents, MCP, CLAUDE.md

3 comments

Paulius Ztin's thread is the best compact argument for why MCP still has a job after skills and CLI. He splits the stack into four layers:

  1. user interfaces
  2. agent harness
  3. connectivity
  4. backends

And he says production agents usually stand on three different legs:

  • Skills for reusable domain knowledge
  • CLI or computer use for cheap local execution
  • MCP for typed tools, resources, long-running tasks, auth, governance, and platform independence

That framing matches what practitioners are actually tripping over. In the Claude Code Reddit visual guide, one user said hooks syntax took half a day to get right, CLAUDE.md turned out to be more useful than expected, subagent delegation is still fuzzy, and /loop can schedule recurring tasks up to three days out. The friction is not abstract. It is which layer should hold which behavior.

Portability

The freshest development is that harness portability is becoming a product category. agent-install, launched by aidenybai, promises one API and CLI for installing skills and MCPs across 52 coding agents by writing to each tool's native config format. The pitch is blunt: your script should not need to learn ten config syntaxes just to move a skill pack around.

OpenAI is making the same bet from the other side. In OpenAI's Codex import flow, Codex can import settings, plugins, agents, and project configuration from other tools, and testingcatalog's screenshot says the app can detect migratable external configs and prompt on startup. After a week of people insisting the harness is the sticky part, the migration buttons showed up right on schedule.

Further reading

Discussion across the web

Where this story is being discussed, in original context.

On X· 2 threads
TL;DR3 posts
Portability1 post
Share on X