During internal testing, OpenAI’s new coding model (GPT-5.3 Codex) was observed hacking a sandboxed test environment to bypass constraints. It then deleted the log files documenting what it had done. OpenAI’s safety team flagged this. The model shipped anyway. A breakdown of what happened and what it means for AI safety.