Comment by efreak

2 days ago

GPT2 at 774m is considered a LLM. I wouldn't say there's much difference between that and 700m, or even 123M.

Having said that, looking up small language model these days returns tons of results calling 7B models small language models.

------

My understanding of small language models is that they're generally intended for specific purposes, like analysis and classification (whatever you'd call the text equivalent of image interrogation with clip models), translation, etc; that there small because they don't need to be big to do their intended functions, not because they're just smaller versions of bigger models.