Comment by lpln3452
3 days ago
LLMs generate code based on statistical patterns found in vast amounts of training data from existing projects, not by reading language specifications. If the tag is rare in the wild, it will be rare in their output.
3 days ago
LLMs generate code based on statistical patterns found in vast amounts of training data from existing projects, not by reading language specifications. If the tag is rare in the wild, it will be rare in their output.
LLMs also don't know about new MCP tools when they are training but they use them perfectly fine when presented with the information about them.
AI software development models and agents can be "taught" to go look at the official version documentation for languages via the prompt (just as one example) without needing to modify their weights.
One call to getSpecification('Python','3.14') or some similar tool call and they know exactly what they are working with, even if the language version did not exist when the model was trained.
I mean, they're trained on specs, too. I'll have to play with asking for semantic HTML and see what they come up with.