WebSockets and connection-scoped caching now reduce API overhead in the Codex agent loop. This architectural shift slashes model latency by maintaining persistent connections rather than repeated handshakes. OpenAI engineers focused on minimizing the round-trip time for iterative tasks. Developers can now build more responsive agentic workflows with significantly lower communication lag.