An LLM is basically a bunch of layers that transform data into other data

Some of those probably at some step encode some “knowledge” of sorts

The problem is that it will work whatever you throw at it, even if it is a nonexistent thing, and return the most sensible “knowledge” according to the training data

Reply to this note

Please Login to reply.

Discussion

It basically is giving someone a lot of words and facts about those words, in a way they cannot remember them all

Then randomly asking them for at least something

They will be correct for the common ones, and for rarer stuff they start making up sensible looking things