Comment by simonw

14 days ago

I'm hoping that Llama 4 goes the same way as Llama 3.

The first Llama 3 models released were 8B and 70B in April 2024.

Llama 3.1 came later in July at 8B, 70B, and 405B.

Llama 3.2 in September got really interesting: 1B, 3B, 11B and 90B.

Then Llama 3.3 in December was 70B but claimed performance similar to the earlier Llama 3.1 405B!

Llama 4 is 109B and 400B, both of which were trained with the help of the 2T(?) "Behemoth".

I'm hoping we'll see further releases in the Llama 4 series that are smaller. I'm particularly excited to see if they produce a ~24B model, since that appears to be the sweet spot for running models on my 64GB laptop while still being able to have other applications running at the same time. Mistral Small 3.1 is a 24B model and is absolutely superb.

(Fleshed this comment out a bit on my blog: https://simonwillison.net/2025/Apr/5/llama-4-notes/#my-hopes...)