LLM Watch

LLM Watch

The Week in AI Agents

AI Agents of the Week: Papers You Should Know About

Get ahead of the curve with LLM Watch

Apr 19, 2026
∙ Paid

Executive Summary

Natural language instructions are failing to control autonomous AI agents - and this week’s research makes that case with striking empirical clarity. Across eight papers, we see an industry grappling with the limits of prompt engineering and pivoting hard toward structural solutions: deterministic infrastructure, explainable governance, transferable memory, and reasoning-aware reward systems. The message is consistent: talking to agents is not enough. We need to engineer around them.

The Rise of Harness Engineering: The single most compelling thread this week is the emergence of “harness engineering” as a distinct discipline - designing the complete infrastructure necessary to transform unconstrained agents into controllable, auditable, production-reliable systems. Sema Code decouples AI coding agents from their delivery interfaces, packaging the reasoning kernel as a standalone, embeddable library that any runtime can drive programmatically. Its companion framework, SemaClaw, extends this philosophy to personal AI agents with DAG-based orchestration, behavioral safety systems, and a three-tier context management architecture. Together, they argue that as model capabilities converge, the harness layer - not the model itself - is becoming the primary site of architectural differentiation.

Agent Observability and Enterprise Trust: Deploying agents at scale without adequate governance is producing a phenomenon researchers call “Agent Sprawl,” and this week two papers dissect the consequences. An empirical study of 4,550 agentic pull requests in Do AI Coding Agents Log Like Humans? reveals that agents fail to comply with constructive natural language logging requests 67% of the time, forcing human developers to perform 72.5% of post-generation log repairs as “silent janitors.” Meanwhile, Agentic Explainability at Scale addresses the corporate fears that accompany this governance vacuum, proposing design-time and runtime explainability techniques - including a prototype “Agentic AI Card” - to make agent-to-agent communication and decision-making transparent to enterprise stakeholders.

Advancing Agent Cognition - Reasoning, Memory, and Decision-Making: Three papers push forward the internal cognitive machinery of agents. Exploration and Exploitation Errors Are Measurable introduces policy-agnostic metrics that independently quantify how well agents balance exploring a problem space versus exploiting acquired knowledge, finding that even frontier models struggle - and that minimal harness engineering significantly improves both dimensions. Memory Transfer Learning demonstrates that cross-domain memory improves average coding agent performance by 3.7% across six benchmarks, but only when memories are stored as high-level abstract insights rather than low-level code traces. And RationalRewards shows that teaching reward models to produce explicit, multi-dimensional critiques before scoring transforms them from passive evaluators into active optimization tools, achieving state-of-the-art preference prediction with 10 - 20x less training data.

Multi-Modal World Simulation: Standing apart from the text-centric agent papers, HY-World 2.0 advances the frontier of 3D world generation and simulation. Its multi-modal pipeline accepts text, images, or video and produces navigable 3D Gaussian Splatting scenes through a four-stage method encompassing panorama generation, trajectory planning, world expansion, and world composition. For agents that must perceive and act in physical or simulated environments, this kind of infrastructure could prove foundational.


User's avatar

Continue reading this post for free, courtesy of Pascal Biese.

Or purchase a paid subscription.
© 2026 Pascal Biese · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture