← Back to context

Comment by SamDc73

7 days ago

I mean they are only running a small version of codex can they run the full one? Or the technology isn't there yet?

1000 tokens/sec for a highly specialised model is where we are going to see agents requiring.

Dedicated knowledge, fast output, rapid iteration.

I have been trying out SMOL models as coding models don't need to the full corpus of human history.

My most recent build was good but too small.

I am thinking of a model that is highly tuned to coding and agentic loops.