Comment by esafak

8 months ago

Proving diversity of thought is a good thing. A controversial observation in 2025's USA ;)

A counterpoint to this is Sourcegraph's Amp, which is all in on Anthropic because they "believe that building deeply into the model’s capabilities yields the best product, vs. building for the lowest common denominator across many models." https://ampcode.com/fif#model-selector

When I embark on a project, I usually ask Gemini to architect and implement the first pass, then iterate with Claude.

Importantly it’s not just any model they’re “alloying”. It’s only the two most capable models where there’s objective evidence that the combination is better than the individual parts.

In this way, it’s not really a “lowest common denominator” as you get to pick the highest performing combination (with solo models just being a special case).

  • In humans, diversity of thought [patterns, not just diversity of knowledge] increases the quality beyond its parts.

    I suspect this model alloy tactic always works, just only seems impressive when it does so with the top models and achieves otherwise unattainable quality.

    One paper of many such new (and nuanced) wisdom of crowds resources:

    Cultural diversity and wisdom of crowds are mutually beneficial and evolutionarily stable https://www.nature.com/articles/s41598-021-95914-7