AI systems that write, edit, review, or debug code.
Claude Code lead Boris Cherny published a feature guide covering mobile handoff, /teleport, /remote-control, scheduled loops, hooks, Dispatch, browser testing, worktrees, /batch, and CLI flags like --bare. The guide shows Claude Code being used as a persistent automation surface, so teams can evaluate whether to lean on remote sessions and repo-scale fan-out.
OpenCode said all Go models now run under zero-retention agreements, then clarified that hosted routes use the same providers customers get direct and explained why higher subscription tiers are risky to price. The clarification matters for users debating telemetry, proxying, and how local the web UI really is, so teams should verify their data path.
OpenClaw's 2026.3.28 update switched xAI integration to the Responses API and added plugin prompts that can request user permission during behavior. One user reported a /v1/models failure after upgrading with a missing operator.write scope, so teams should check auth changes before deploying.
Users showed a flow where a rough Figma sketch and file link are handed to Claude Code, which fleshes out styled mockups and extra components inside Figma. The handoff keeps UI iteration in the design tool before code generation, so teams can keep design review upstream of frontend implementation.
Users report new request-per-minute caps that trigger after three to four concurrent agents, and Boris Cherny says efficiency work is underway. The issue hits the multi-agent workflows Anthropic has been promoting, separate from five-hour usage buckets.
Hermes Agent v0.5.0 adds 400+ models via Nous Portal, Hugging Face access, Exa support, GPT-5.4 behavior tweaks, and a published changelog. The release broadens provider coverage and hardens the runtime without changing the terminal-first workflow.
The ATLAS harness says a frozen Qwen3-14B Q4 model on one RTX 5060 Ti reached 74.6% pass@1-v(k=3) on LiveCodeBench v5 through multi-pass repair and selection. The result shifts comparison toward harness design, though HN commenters note it is not a one-shot head-to-head with hosted frontier models.
Z.ai said GLM-5.1 is now available to all GLM Coding Plan users and highlighted a 5am to 11am PT switch window. The update broadens access beyond the initial rollout, though early practitioner tests reported weaker Repo bench and tool-calling behavior than 5.0.
Composio shipped Universal CLI as a shell-first interface to its integrations, moving install, search, and agent workflows out of MCP setup. The release targets users who want simpler agent tool access after complaints that MCP stacks are harder to install, slower, and less stable.
Z.ai made GLM-5.1 available to all Coding Plan users and documented how to route coding agents to it by changing the model name in config. Early harness benchmarks place it near Opus 4.6 on coding evals, but BridgeBench users report much slower tokens per second.
OpenAI published a Codex use-case gallery with one-click workflows, and shadcn/ui and Box shipped official plugins. Teams can now install reusable app and web workflows directly instead of wiring each integration by hand.
Claude Code can now run recurring prompts and background pull-request work on Anthropic-managed cloud environments from the web, desktop, or `/schedule`. That makes long-running repo tasks less dependent on a local machine, but users report task caps and restricted egress.
Hermes Agent now treats Hugging Face as a first-class inference provider and surfaces 28 curated models in its picker, plus a custom path to the broader catalog. That broadens model choice for a persistent local agent workflow without requiring users to wire a provider manually.
OpenCode shipped terminal, desktop, and `opencode serve` workflows for an open-source coding agent with LSP support, plugins, and more than 75 providers. Users should look at the multi-backend web sessions, IPC plugins, and sandboxed local setup as the main differentiators.
Cline launched Kanban, a local multi-agent board that runs Claude, Codex, and Cline CLI tasks in isolated worktrees with dependency chains and diffs. Teams can use it as a visual control layer for parallel coding agents on repo chores that split cleanly.
OpenAI rolled out Codex plugins across the app, CLI, and IDE extensions, with app auth, reusable skills, and optional MCP servers. Teams should test plugin-backed workflows and permission models before broad rollout.
Every opened Plus One, a hosted OpenClaw that lives in Slack, comes preloaded with internal skills, and works with a ChatGPT subscription or other API keys. It lowers the ops burden for deployed coworkers, so teams can test packaged agents before building their own stack.
Rork added Max Publishing to generate icons, screenshots, listing text, review metadata, and submission steps for App Store releases, and also shipped an App Store MCP. Use it first on non-critical apps and keep a manual review gate.
Imbue released Latchkey, a library that prepends ordinary curl calls so local agents can use SaaS and internal APIs while credentials stay on the developer machine. Try it where agents need many HTTP integrations but should not see raw secrets.
OpenCode is adding remote sandboxes, synced state across laptop, server, and cloud, and more product surface inside its plugin system. That makes long-running off-laptop workflows more practical, but operators should still review telemetry, sandbox, and exposure defaults.
Claude Code 2.1.84 adds an opt-in PowerShell tool, new task and worktree hooks, safer MCP limits, and better startup and prompt-cache behavior. Anthropic also documented auto mode’s action classifier and added iMessage as a channel, so teams should review permissions and remote-control workflows.
GPT-5.4 mini and nano bring 400K context, multimodal input, and the full GPT-5.4 reasoning-mode ladder at lower prices. Early benchmarking suggests nano is the strongest cost-performance tier for agentic tasks, but both models spend far more output tokens than peers.
Expect wraps browser QA for Claude Code, Codex, or Cursor into a CLI that records bug videos and feeds failures back into a fix loop. It gives coding agents a tighter UI validation cycle without requiring a custom browser harness.
OpenClaw 2026.3.24 adds native Microsoft Teams, OpenWebUI sub-agent access, Slack reply buttons, and a control surface for skills and tools. The release expands where the runtime can plug into enterprise workflows, while also increasing the surface area teams need to secure.
Cursor shipped Instant Grep, a local regex index built from n-grams, inverted indexes, and Bloom filters that drops large-repo searches from seconds to milliseconds. Faster candidate retrieval shortens the coding-agent loop, especially when ripgrep-style scans become the bottleneck.
OpenClaw shipped version 2026.3.22 with ClawHub, OpenShell plus SSH sandboxes, side-question flows, and more search and model options, then followed with a 2026.3.23 patch. Teams get a broader plugin surface, but should patch quickly and review plugin trust boundaries as the ecosystem grows.
Claude can now drive macOS apps, browser tabs, the keyboard, and the mouse from Claude Cowork and Claude Code, with permission prompts when it needs direct screen access. That makes legacy desktop workflows automatable, and Anthropic is pairing the push with more background-task support for longer agent loops.
Vals AI switched SWE-Bench Verified from SWE-Agent to the bash-only mini-swe-agent harness, aligning results more closely with the official benchmark setup. Top score dipped slightly to 78.8%, but the change reduces harness-specific confounds when comparing models.
Agent Computer launched cloud desktops that boot in under half a second and expose persistent disks, shared credentials, SSH access, and ACP control for agents. It gives coding agents a faster place to run tools and reuse auth, but teams still need to design safe session and credential boundaries.
OpenHands introduced EvoClaw, a benchmark that reconstructs milestone DAGs from repo history to test continuous software evolution instead of isolated tasks. The first results show agents can clear single tasks yet still collapse under regressions and technical debt over longer runs.
PlayerZero launched an AI production engineer and claims its world model can simulate failures before release, trace incidents to exact PRs, and beat existing tools on real production test cases. If those numbers hold, the interesting shift is from code generation to debugging, testing, and observability after code ships.
Vercel Emulate added a programmatic API for creating, resetting, and closing local GitHub, Vercel, and Google emulators inside automated tests. That makes deterministic integration tests easier to wire into CI and agent loops without manual setup.
Anthropic is testing a new /init flow that interviews users and configures Claude.md, hooks, and skills in new or existing repos. Try it in a sandbox repo, then watch for skills behavior differences between chat and web surfaces.
OpenClaw's maintainer asked users to switch to the dev channel and stress normal workflows before a large release that may break plugins. Watch harness speed, context plugins, and permission boundaries closely while the SDK refactor lands.
Nous Research said Hermes Agent crossed 10,000 stars, while users reported easy migrations from OpenClaw and stable long-running use. If you test it, focus on persistent memory, MCP browser control, and delegation behavior under real workloads.
Skyler Miao said MiniMax M2.7 open weights are due in roughly two weeks, with updates tuned for agent tasks. Separate replies also confirm multimodal M3, so local-stack builders should watch both the drop and the benchmark setup.
LangChain published a free course on taking agents from first run to production-ready systems with LangSmith loops for observability and evals. The timing lines up with new NVIDIA integration messaging, so teams can study process and stack choices together.
CopilotKit shipped hooks that let agents inspect app state and call frontend actions, then paired them with Shadify for ShadCN-based UI composition. It gives embedded agents a cleaner path from chat to in-app behavior.
Vercel Labs published a stateful service emulator for GitHub, Vercel, and Google integrations instead of relying on brittle mocks. It is useful when agents or CI need deterministic auth and third-party API flows in local or sandboxed runs.
Claude Code can now run scheduled cloud tasks against remote repos and MCP-connected tools, while Anthropic is also pushing reusable agent SDK and skill controls. Test remote automation paths carefully, because messaging and multi-repo edge cases still surface in practice.
Vercel's Next.js evals place Composer 2 second, ahead of Opus and Gemini despite the recent Kimi-base controversy. The result matters because it separates base-model branding from measured task performance on a real framework workflow.
Agent Flywheel lays out a planning-first workflow built on beads, agent mail, swarms, and TUI inspection for very large coding runs. It is useful because the guide exposes coordination primitives and review loops, not just benchmark screenshots.
Conductor now bundles plan mode, fast mode, skills, repo quick start, and an experimental merge-conflict UI around Codex sessions. Try it if you want a higher-level harness for long-running code agents, but watch the foreground chat UX on larger tasks.
OpenCode can now run from AWS CloudShell via npx and inherit AWS auth plus Bedrock models; the same update also brought Firecrawl, India billing, and heap-snapshot debugging. It is becoming a real ops workflow, not just a local terminal toy.
A developer says an autoresearch loop hill-climbed a vibecoded Rust engine to 2718 Elo after running more than 70 experiments under a 500 ms move budget. The real takeaway is the workflow: automated experiment loops can optimize code against a measurable target.
ACE open-sources a platform that turns AGENTS.md instructions into evolving playbooks backed by execution history, with hosted and self-hosted options. It is a notable response to prompt drift and prompt extraction, because procedures become revisable operating docs instead of static prompts.
WSJ reported that OpenAI is consolidating ChatGPT, Codex, and its browser into a single desktop app to simplify heavy-use workflows. If it ships, developers would get one workspace for chat, coding, and browsing instead of today's fragmented clients.
Anthropic rolled Projects into Cowork on the Claude desktop app, giving each project its own local folder, persistent instructions, and import paths from existing work. It makes Cowork more practical for ongoing tasks, though teams should test current folder-location limits.
Cursor and Kimi said Composer 2 starts from Kimi K2.5, with continued pretraining and RL added on top after developers spotted Kimi model IDs in traffic. Teams should benchmark it as a productized open-base stack, not a from-scratch model.
Next.js 16.2 adds version-matched AGENTS.md docs, a terminal browser for inspecting running apps, browser-error forwarding, and a dev-server lock file. It gives coding agents better frontend context and cuts duplicate-server and client-side debugging waste.