The "next word predictor" model is pervasive and disappointingly ridiculous. If you present an LLM with a logic puzzle and it gives you the correct answer, how did it "predict the next word"? Yes, the output came in the form of additional tokens. But deciding which tokens were "most likely" required logical thought, and it's a mistake to take the "what" as the "how".
#LLM #AI