Language models learn grammaticality as a distinct concept from string probability, hidden in their internal representations rather than reflected in output probabilities—you can extract this knowledge with a simple linear probe.
Language models generate grammatical text but their probability scores don't clearly distinguish grammatical from ungrammatical sentences.