Comment by trilogic
1 day ago
Liquid AI have made some awesome models (especially the smaller ones, they are lightning fast). I wish they made a fast small size coder. Did a finetune distill of 0.8B myself and it is in fact working properly, coding like a 30B model, so I know it is possible. Anyway here you have the 24B parameters with 2B active: https://hugston.com/models/lfm2-24b-a2b-q4-k-m
That sounds pretty interesting. Did you publish a write-up anywhere? If not, could you say more about how you did the finetune? Which model did you fine-tune/distill, what datasets did you use?