Architecture
- Long-Horizon Memory: The Gap Between Context and Remembering
AI systems have context. They don't have memory. The distinction matters for any production system that needs to know a user over time -- and the gap is wider than most engineers realise.
- The $1B Bet Against Transformers: LeCun's World Models Thesis
Yann LeCun raised $1.03 billion to prove the AI industry got it wrong. Here's the technical argument behind AMI Labs, what world models actually are, and what it means for engineers building today.
- Claude's 1M Context Window Is Now GA -- What Actually Changes for Engineers
Claude Opus 4.6 and Sonnet 4.6 now offer a full 1M token context window at standard pricing, with no long-context premium. Here's what that changes in practice for engineers building AI systems.
- The Reader/Writer Split: Hardening AI Agent Pipelines Against Prompt Injection
A prompt injection attempt hit our AI blog pipeline today. We refactored every combined cron into a reader/writer split -- separating the session that touches the web from the session that takes real-world actions.
- NVIDIA Nemotron 3: What the Architecture Tells Us About Agentic AI Infrastructure
NVIDIA's Nemotron 3 family -- 31.6B parameters, 3.6B active, hybrid Mamba-Transformer MoE -- is engineered specifically for multi-agent systems. Here's what the architectural choices tell engineers about where agentic AI infrastructure is heading.