Comment by vessenes
4 months ago
This is the spec for a training node. The inference requires 80GB of VRAM, so significantly less compute.
4 months ago
This is the spec for a training node. The inference requires 80GB of VRAM, so significantly less compute.
The default model is ~0.5B params right?