Posts

We finally conquered alchemy

But without economies of scale

Five Hinge‑Questions That Decide Whether AGI Is Five Years Away or Twenty

For people who care about falsifiable stakes rather than vibes

Born on Third Base

The Case for Inheriting Nothing and Building Everything

One-dimensional vs multi-dimensional features in interpretability

How to stop conflating with 768 dimensions

LLMs are really good at k-order thinking (where k is even)

You still need to tell a language model you want to cure cancer before it can help you cure cancer.

Information bounds in quantum gravity

How information theory links quantum mechanics and general relativity

Books I read (am reading) in 2025

A dynamic list

The nihilism of NeurIPS

Some thoughts on our future in AI research

Can quantised autoencoders find and interpret circuits in language models?

Using VQ-VAEs and categorical decision trees to do automatic circuit identification in LLMs.

Learning compressed representations and GPT-5 speculation

Why language models probably get too much from the abstraction we give them for free