Normalising a Short AGI Horizon

AI 2027 was recently released and is another great forecast of how the next few years could play out, similar to another favourite of mine, Situational Awareness. While listening to the authors on the Dwarkesh Podcast I tried to look a bit deeper on my own thoughts. Why do I also believe AGI is coming soon? Why does it still seem so fantastical, despite my belief? To help myself, I decided to jot down the key points that have influenced that thinking. It’s far from an exhaustive list, but are some of the latest facts that help me to stay grounded in my predictions.

Read More

Think it Faster in AI

Ever since reading “How could I have thought that faster?” this year, I have been trying to put it into practice. Working with AI models, I found there are plenty of opportunities. One can spend hours on some buggy code just to find the bug they fixed wasn’t the real problem after all, or that someone had already solved it five years ago on Stack Overflow. One can invests hours into modelling to realise that there was a far simpler approach if you just thought about it from another angle…

Read More

Test-Time Compute Matters: From LLMs to Search

Test-time compute has recently been popularised with LLMs (e.g. o1/o3, DeepSeek-R1 and other reasoning models) as its application has allowed LLMs to perform significantly better on complex reasoning problems (e.g. mathematics, reasoning). This article explores why it works, why it’s not new, and how it’s been employed across different AI paradigms.

Read More

How to Run Cheap LLM Experiments

As a researcher, it’s common to have way more ideas than you have time to experiment with. This is especially true in the world of language modeling when you consider the cost of running such experiments. In this post I touch on some of the methods that I’ve seen to run experiments with language models without breaking the bank.

Read More