Computation Beyond Tool Use: Executing Programs Inside a Transformer
When LLMs need to compute something reliably, we've settled into a familiar pattern: the model writes code, an external interpreter runs it, the result gets injected back into the context, and the model carries on. It works well enough that we rarely stop to question the architecture. But