Skip to main content

Everything is RL: Automatically Creating Optimal Agents with Evolutionary Methods

· 11 min read

The way most people build software with LLMs today is: write something, look at it, ask an LLM to make it better, look at it again. Vibecoding. It's fast, it's surprisingly effective, and it's still fundamentally local search guided by human intuition.

Human intuition is a bad guide for knowing if something actually got better. You can feel like an iteration was good without having moved the metric that matters. You can iterate eight times and end up with a more sophisticated-looking solution that scores lower than what you started with. I know this because I ran the experiment.

The New Normal? Early Results from InceptionLabs' Diffusion-Based LLM Look Promising

· 7 min read

The longer your context, the slower transformer-based LLMs get. It's not a tuning problem, it's architectural: autoregressive models produce one token at a time and attend over the full context on each step. At 20,000 tokens, you're paying for 20,000 tokens of attention on every single generation step.

InceptionLabs' Mercury-2 uses a diffusion architecture that generates output in parallel across the full sequence, so its latency doesn't scale the same way. I benchmarked it against GPT-4.1-nano and GPT-5-nano in a RAG pipeline at two context lengths. At short context, Mercury-2 finishes last on every metric. At 21k tokens, it's 5x faster than the alternatives and the only model that stays under 1.5 seconds. The crossover is around 4,500 tokens.

The Three AI Frontiers for 2026 - A Prediction and What to Look Out For

· 5 min read

I've lived and breathed AI and agents for 365 days of 2025 (and 2026 won't be any different). The AI boom of the last few years has been incredible to watch, but I keep coming back to the same few issues:

  1. Energy is not scaling with demand.
  2. When we say want autonomous AI agents, we don't really want autonomous AI agents.
  3. Data is not scaling with demand and synthetic data will never be the whole story.