Comment by skate
2 months ago
As others pointed out this problem isn't special.
Grok 4 heavy Thought for 4m 17s
{"decoded_prefix": "nqxznhzhvqvvjddqiterrqdboctzzmoxmhyzlcfe", "last_10": "kfohgkrkoj", "vowel_counts": {"a": 7, "e": 18, "i": 7, "o": 12, "u": 6}}
it did count another e, but that's a known point of failure for LLMs which i assume you put in intentionally.
>Counting e's shows at least 10 more, so total e's are <at least> 17.
I guess GPT-5 with thinking is still a bit ahead of grok. I wonder what the secret sauce is.