← Back to context

Comment by deepsquirrelnet

1 day ago

This is pretty cool. I have a similar model that’s 8 days into training on msmarco.

So far I only have the “cold start” data posted, but I’m planning on posting a full distillation dataset.

https://huggingface.co/datasets/dleemiller/lm25

What kind of hardware setup would be needed to replicate the paper’s results?

  • I am training phi-4 (14B) using a single A6000. There’s some tricks you have to use to keep VRAM consumption down - mainly LoRA and quantization.

    There’s a package called “unsloth” that integrates with huggingface’s TRL library that can help.