The "next word predictor" model is pervasive and disappointingly ridiculous. If you present an LLM with a logic puzzle and it gives you the correct answer, how did it "predict the next word"? Yes, the output came in the form of additional tokens. But deciding which tokens were "most likely" required logical thought, and it's a mistake to take the "what" as the "how".

#LLM #AI

Reply to this note

Please Login to reply.

Discussion

No replies yet.