arXiv:2512.20660v1 Announce Type: cross
Abstract: Current approaches to AI coding agents appear to blur the lines between the Large Language Model (LLM) and the agent itself, asking the LLM to make decisions best left to deterministic processes. This leads to systems prone to stochastic failures such as gaming unit tests or hallucinating syntax. Drawing on established software engineering practices that provide deterministic frameworks for managing unpredictable processes, this paper proposes setting the control boundary such that the LLM is treated as a component of the environment environment — preserving its creative stochasticity — rather than the decision-making agent.
A textbfDual-State Architecture is formalized, separating workflow state (deterministic control flow) from environment state (stochastic generation). textbfAtomic Action Pairs couple generation with verification as indivisible transactions, where textbfGuard Functions act as sensing actions that project probabilistic outputs onto observable workflow state. The framework is validated on three code generation tasks across 13 LLMs (1.3B–15B parameters). For qualified instruction-following models, task success rates improved by up to 66 percentage points at 1.2–2.1$times$ baseline computational cost. The results suggest that architectural constraints can substitute for parameter scale in achieving reliable code generation.
AI Wrapped: The 14 AI terms you couldn’t avoid in 2025
If the past 12 months have taught us anything, it’s that the AI hype train is showing no signs of slowing. It’s hard to believe


