pull down to refresh

Most concrete example from there:
For instance, an LLM might learn that a question like “Where is Paris located?” is structured as adverb/verb/proper noun/verb. If there are many examples of sentence construction in the model’s training data, the LLM may associate that syntactic template with questions about countries. So, if the model is given a new question with the same grammatical structure but nonsense words, like “Quickly sit Paris clouded?” it might answer “France” even though that answer makes no sense.

It sounds to me like model performance goes down when, in the training data, there's a lot of correlation between speech patterns and domains.
I think the issue is more about reoccurring speech patterns inside domains.
100 sats \ 1 reply \ @0xbitcoiner 6h
Seems like CGPT ain’t got that disease! Ahaha
Quickly sit Paris clouded?
ChatGPT Free:
I’m not sure what you mean by “Quickly sit Paris clouded?” It looks like a sentence fragment or a poetic phrase.
Would you like me to:
Interpret it?
Rewrite it into a clearer sentence?
Guess what you intended to ask (maybe about weather in Paris or a description)?
Let me know!
reply
They claim they've reproduced it (though maybe not this particular example) on 4o
reply
Hmm, interesting.
LLMs are weird.
reply