The LLM:Probabilistic Predictor An LLM (Large Language Model) does not have a math engine. It is a Next-Token Predictor. When you ask it a question, it is performing a high-speed search through a high-dimensional space of text patterns. The process: It views your query as a sequence of tokens, converts them into vectors, and uses Self-Attention to weigh the importance of those tokens. The outcome: It is always calculating probability. When it produces 2 as the answer to 1 + 1=, it isn't "adding"
Comment
Sign in to join the discussion.
Loading comments…