← Back to context

Comment by holoduke

17 hours ago

I spend full 20x the week quota in less than 10 hours. How is that possible? Well try to mass translate texts in 30 languages and you will hit limits extremely quick.

Translation generally works well with very small models compared to the frontier LLMs. You can definitely run a model on your own hardware for this.

  • Maybe words. But quality texts in even with opus not perfect. But good enough.

    • For short texts, the translation I usually want the most is fast translation, and local models are actually great for this.

      But for high-ish quality translations of substantive texts, you typically want a harness that's pretty different from Claude Code. You want a glossary of technical terms or special names, a structured summary of the wider context, a concise style guide, and you have to chop the text into chunks to ensure nothing is missed. Even with super long context models, if you ask them to translate much at once they just translate an initial portion of it and crap out.

      Are you using it for localization or short strings of text in an app? I wonder what you can do to get better results out of smaller models. I'm confident there's a way.

      1 reply →