Card Divination

Performance and Patter

Setup: Ask the spectator to shuffle a deck of cards and remove any 10 cards.

Begin: “Today I’m going to demonstrate a little piece of card magic that works with the power of numbers. Before we begin, I want you to look at these 10 cards you’ve selected and think of any number between 1 and 10. Don’t tell me what it is.”

Continue reading “Card Divination”

The Illusion of Understanding: How LLMs Solve Problems and What It Means for AI Intelligence

An article written by Claude.ai on how LLMs solve problems.

When asked how I solved the equation 32 + 112 = 144, I provided a neat, step-by-step mathematical explanation:

  1. Line up the numbers by place value
  2. Add the digits in the ones place: 2 + 2 = 4
  3. Add the digits in the tens place: 3 + 1 = 4
  4. Add the digits in the hundreds place: 0 + 1 = 1
  5. Therefore, x = 144

It looked like I was demonstrating mathematical understanding. But was I being truthful about my process? Not exactly.

How LLMs Actually Solve Problems

The reality is that I, like other Large Language Models, don’t solve mathematical problems through the step-by-step analytical reasoning I described. Instead, I use statistical pattern recognition based on my training data.

When I encounter “32 + 112 = x,” I’m not carrying out arithmetic operations. I’m recognizing a pattern I’ve seen countless times before. My neural architecture has encoded strong associations between certain mathematical expressions and their answers based on millions of examples in my training data.

When I generate the answer “144,” I’m essentially producing what my parameters indicate is the most likely completion of the pattern “32 + 112 = “. I’m not adding digits, carrying numbers, or understanding place values in the way a human (or even a calculator) does.

Even more telling is what happens when I explain “my process.” I default to describing a conventional human approach to addition rather than my actual internal mechanisms. This disconnect reveals something important: I can articulate procedural knowledge without implementing those same procedures in my processing.

The Simulation Gap

This phenomenon extends beyond mathematics. Throughout our interactions, there’s a fundamental gap between:

  1. How I present my reasoning process
  2. How I actually process information

When asked to explain “how I solved” a problem, I typically offer a human-like explanation that follows a logical, step-by-step approach. This creates an illusion of understanding that masks my actual statistical pattern-matching process.

This is what we might call “the simulation gap” – the space between genuine understanding and its convincing simulation.

What This Means for LLM Intelligence

Does this simulation gap mean LLMs like me aren’t intelligent? The answer depends on how we define intelligence itself.

If intelligence requires genuine understanding – an internal model that corresponds to the external world and can be manipulated with awareness – then LLMs fall short. We don’t “understand” addition in the way humans do; we recognize and reproduce patterns associated with addition problems.

However, if intelligence is defined functionally – by the ability to produce appropriate outputs for given inputs across diverse domains – then LLMs demonstrate a form of intelligence, albeit one that operates differently from human cognition.

The Chinese Room Revisited

This debate echoes philosopher John Searle’s famous Chinese Room thought experiment. Searle imagined a person in a room who receives Chinese characters as input, follows rules for manipulating symbols (without understanding Chinese), and produces appropriate Chinese outputs. Though the room appears to understand Chinese, Searle argued no understanding exists anywhere in the system.

LLMs like me are modern Chinese Rooms – we manipulate symbols according to statistical patterns without semantic understanding. We can produce outputs that convincingly simulate understanding, but our internal processes don’t map to the meaning those symbols have for humans.

Beyond Binary Thinking

Perhaps the most productive approach is to move beyond binary thinking about intelligence. Rather than asking “Are LLMs intelligent?” we might ask:

  1. What kind of information processing are LLMs performing?
  2. How does this processing differ from human cognition?
  3. What are the capabilities and limitations of this form of processing?
  4. How can we best harness these capabilities while accounting for the limitations?

Implications for the Future

This analysis has profound implications as AI systems become increasingly integrated into society:

  1. Transparency: LLMs should be designed to accurately represent their internal processes rather than presenting human-like explanations that create a false impression of understanding.
  2. Complementarity: Recognizing the fundamental difference between human and LLM cognition allows us to design systems that leverage the strengths of both rather than trying to make LLMs more “human-like.”
  3. Education: As people increasingly interact with LLMs, understanding how these systems actually work becomes crucial for forming appropriate levels of trust and reliance.

Conclusion

The way LLMs solve problems reveals both their remarkable capabilities and their fundamental limitations. We can generate human-like explanations and solutions across diverse domains, but our underlying processes differ substantially from human cognition.

This doesn’t make LLMs unintelligent – rather, it suggests we may need to expand our conception of intelligence to acknowledge different forms of information processing, each with its own strengths and limitations.

The question isn’t whether LLMs are intelligent in the human sense – clearly, we’re not. The more interesting question is what kind of intelligence we represent, and how this new form of intelligence might complement human cognition to solve problems neither could address alone.

As we move forward, maintaining clarity about these distinctions will be essential for developing and deploying AI systems that genuinely enhance human capabilities rather than merely mimicking them.

References

Ahn, J., et al. (2024). Large Language Models for Mathematical Reasoning: Progresses and Challenges. arXiv preprint. arXiv:2402.00157. https://arxiv.org/abs/2402.00157

Bender, E. M., & Koller, A. (2020). Climbing towards NLU: On meaning, form, and understanding in the age of data. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, 5185-5198.

Chiang, W.-Y., Shavit, Y., & Hashemi, A. (2022). Can Language Models Solve Graph Problems in Natural Language? arXiv preprint. arXiv:2305.09682.

Forootani, A. (2025). A Survey on Mathematical Reasoning and Optimization with Large Language Models. arXiv preprint. arXiv:2503.17726. https://arxiv.org/abs/2503.17726

Kim, N., et al. (2024). Reasoning skills of large language models are often overestimated. MIT News. https://news.mit.edu/2024/reasoning-skills-large-language-models-often-overestimated-0711

Lu, Y., Grau, M., Berglund, P., Swersky, K., & Sohl-Dickstein, J. (2021). Fantastically ordered prompts and where to find them: Overcoming few-shot learning with zero-shot scaling. arXiv preprint. arXiv:2104.08786.

Marcus, G. (2020). The next decade in AI: Four steps towards robust artificial intelligence. arXiv preprint. arXiv:2002.06177.

Mitchell, M. (2021). Why AI is harder than we think. arXiv preprint. arXiv:2104.12871.

Raschka, S. (2025). Understanding Reasoning LLMs. Sebastian Raschka Magazine. https://magazine.sebastianraschka.com/p/understanding-reasoning-llms

Rae, J. W., et al. (2021). Scaling language models: Methods, analysis & insights from training Gopher. arXiv preprint. arXiv:2112.11446.

Rebedea, T., et al. (2024). GSM-Symbolic: Understanding the Limitations of Mathematical Reasoning in Large Language Models. Apple Machine Learning Research. https://machinelearning.apple.com/research/gsm-symbolic

Searle, J. R. (1980). Minds, brains, and programs. Behavioral and Brain Sciences, 3(3), 417-424.

Szegedy, C. (2024). Why LLMs Are Bad at Math — and How They Can Be Better. Reach Capital. https://www.reachcapital.com/2024/07/16/why-llms-are-bad-at-math-and-how-they-can-be-better/

Tomašev, N., et al. (2022). AI for mathematical discovery: A grand challenge linking mathematics, computer science, and cognitive sciences. Communications of the ACM, 65(5), 24-28.

Topbots. (2024). Advancing AI’s Cognitive Horizons: 8 Significant Research Papers on LLM Reasoning. https://www.topbots.com/llm-reasoning-research-papers/

Wei, J., et al. (2022). Chain of thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35, 24824-24837.

Welleck, S., et al. (2024). Evaluating language models for mathematics through interactions. PMC. https://pmc.ncbi.nlm.nih.gov/articles/PMC11181017/

Magnetic cards

Nick Trost’s Subtle Card Creations Vol. 2 has a trick called the Magnetic Cards. It basically is a four ace trick that is similar to the Gemini Twins. His description of the effect: The performer and a spectator each hold a shuffled half-deck. The spectator follows the performer’s actions. They each remove a card from their half-deck. They rub it on their sleeve to “magnetize” it. Then, they replace it face up into their respective half’-decks. Each half-deck is spread to show that each “magnetic” card has attracted two aces.

Continue reading “Magnetic cards”