4chan Users Discovered Chain-of-Thought Reasoning Before Google Did
In 2022, 4chan users were playing AI Dungeon. They gave the AI a math problem. The result was wrong. So they tried something simple: they asked it to show its work, solve the problem step-by-step, explain the reasoning as it went. The answer was right.
This was not a formal experiment. There was no lab coat. No peer review queue. Just people playing with an AI, discovering by accident that asking it to think out loud made it smarter. More than a year later, in January 2024, Google researchers published a paper called "Chain-of-Thought Prompting Elicits Reasoning in Large Language Models." They presented the same idea as a breakthrough.
Google did not mention 4chan. They did not mention the community discovery that preceded their publication by more than a year. They published it as if it was their insight. Today, chain-of-thought reasoning is foundational to models like Claude and ChatGPT-4o. It is the reason models like OpenAI's o1 and o3 are so capable. And the first person to discover it was someone on an anonymous image board, not a researcher at a major lab.
How It Worked
The mechanics are straightforward. When you ask an AI to solve a problem, it generates an answer all at once. That answer is often wrong because the model has not had time to "reason" through the steps. It is pattern-matching at high speed, and patterns do not always lead to truth.
But if you ask the AI to show its work, solve one step at a time, and explain the reasoning before giving the final answer, something changes. The model is forced to generate intermediate steps. Those steps act like anchors. They constrain the final answer. The result: a better answer.
This is not magic. It is not even that surprising if you think about how humans solve math problems. We do not jump to the answer. We work through steps. We write things down. We check our work. We use intermediate steps to catch errors. Why would we expect AI to be different?
But the AI research community did not make that connection. Not formally. Not until Google published a paper saying so. That paper changed how people talked about the problem. It gave the idea legitimacy. It became a technique. It got built into systems. It became standard.
The Timeline Matters
Let me be clear about the dates because the order is important.
In 2022, AI Dungeon was a popular text adventure game powered by GPT-3 and later GPT-3.5. Players asked the AI to continue narratives. Some of those players were curious about the model's limits. They played with prompting. They discovered that asking the AI to explain its reasoning improved the coherence of its output. They shared findings on forums and Reddit threads. The community knew about this technique.
In January 2024, Google published their paper on chain-of-thought prompting. The paper was rigorous. It included experiments and benchmarks. It showed that this simple technique could significantly improve reasoning performance on mathematical and commonsense tasks.
The gap: more than a year. The community had this insight. The labs caught up later.
Why This Matters
This is not a gotcha. Google did important work formalizing and testing the idea. The paper generated citations and attention that expanded adoption. That matters. But it is a reminder about where innovation actually lives.
AI breakthroughs are not all happening in Google labs. They are not all happening in OpenAI offices or Anthropic research teams. Some are happening because someone on 4chan asked a simple question and paid attention to the answer. Some are happening in Discord servers and hackathons. Some are happening in small teams that nobody has heard of.
The labs have resources and rigor. But they do not have monopoly on insight. The labs are slow sometimes. They have bureaucracy. They have incentives that point away from the kind of experimentation that might lead to breakthroughs.
A person playing a video game has different incentives. They want the game to work better. They experiment. They discover. They share.
Chain-of-Thought Today
Fast forward to 2026. Chain-of-thought is not a trick anymore. It is infrastructure. OpenAI built it into o1 and o3. These models spend time reasoning before answering. That is chain-of-thought at scale. Claude uses it. Gemini uses it. Every frontier model uses some version of this idea.
The reason o1 and o3 are so capable is because they do what 4chan players discovered in 2022: they reason step-by-step. They show their work. That makes them better at math, code, and reasoning tasks. It is the same insight. It is just built into the model itself instead of being a prompting technique.
Models that reason before answering are dramatically better than models that do not. The difference is visible in benchmarks and in practice. We know this because the community discovered it. The labs confirmed it. Everyone adopted it.
The Broader Lesson
Here is what this history tells us about AI innovation in 2026.
Breakthroughs do not belong to any one organization. The person who discovers something first might be in a research lab. They might be a startup founder. They might be a hobbyist. They might be someone playing a game online.
The labs will always have advantages: compute, talent, publication channels, credibility. But ideas come from everywhere. The best outcome is when the labs take ideas from the community, formalize them, test them, and build them into products that everyone can use.
That is what happened with chain-of-thought. The community discovered it. The labs validated it. The world got better models. That cycle needs to keep happening.
Pay attention to what people on the edges are discovering. They might be ahead of the labs more often than you think.
See chain-of-thought in action on our benchmark pages.
Compare reasoning performance across models on our Benchmarks page, explore current model capabilities, and dive deeper into AI research at our research hub. Learn what AI is and how it works.
About Kira Nolan: Kira writes about AI research, innovation, and the often surprising origins of breakthroughs at TensorFeed. Focus is on the communities and experiments that drive progress forward.