Comment by ANewFormation
1 year ago
LLMs are 100% deterministic. The facade of randomness is injected solely by a superfluous rng factor.
1 year ago
LLMs are 100% deterministic. The facade of randomness is injected solely by a superfluous rng factor.
Even if in the forward pass there would be no "temperature" tilting, the NN training would still be performed through different processes on different implementations, making the outputs "personal".