🚨Breaking: Someone just open sourced a Python interpreter written in Rust and it's wild. It's called Monty. And it's not a sandbox. It's a minimal, secure Python runtime built specifically for AI agents -- and it starts in under 1 microsecond. Here's what this thing does: → Runs Python code written by LLMs with zero sandbox overhead → Completely blocks filesystem, env vars, and network unless YOU allow it → Starts in 0.06ms vs Docker's 195ms and Pyodide's 2,800ms → Snapshots execution state mid-flight -- pause, serialize, resume later → Runs typechecking with `ty` baked into a single binary → Calls from Python, Rust, or JavaScript -- no CPython dependency Here's the wildest part: Docker takes 195ms to start. A sandboxing service takes 1,033ms. Monty takes 0.06ms. That's not a rounding error. That's a different category of tool entirely. Every AI agent running today has the same problem: you either run LLM code directly on your host (YOLO Python -- zero security) or you spin up a container and wait 200ms per call. Monty closes that gap. One import. No daemon. No image pull. No container escape risk. The pydantic team built this to power code-mode in PydanticAI -- where LLMs write Python instead of making tool calls, and Monty executes it safely. Cloudflare and Anthropic are already publishing on this exact paradigm....