Comment by progbits

3 months ago

That's not how the censorship works, the model doesn't know. Either some information has been excluded from training data set, some answers penalized in training or, most commonly, just another LLM/simple regex on the output that cuts out the response

I’m talking about pruning a local LLM not using their service. There are plenty of ways to prune and distill. Heck DeepSeek was distilled from other models. You could simply run a distillation using Hex, then convert those outputs back to the target language.