Glossary/Chain-of-Thought (CoT)

What is Chain-of-Thought Prompting?

Chain-of-thought (CoT) prompting is the technique of asking a language model to produce step-by-step reasoning before its final answer, rather than jumping straight to a conclusion. Models given room to "think out loud" reach correct answers on math, logic, and multi-step reasoning problems at much higher rates than models prompted to answer immediately.

How chain-of-thought works

The technique is structurally simple: append "Let's think step by step" or "Think through this carefully before answering" to a prompt, and the model produces intermediate reasoning before its conclusion. The 2022 Wei et al. paper from Google formalized it; modern models are RLHF-trained to use CoT by default on hard problems.

Variants:

Why CoT matters

Reasoning-heavy benchmarks (GSM8K math, BIG-Bench Hard, etc.) saw 10-50 percentage-point improvements from CoT prompting at the time it was discovered. Modern reasoning models that generate long internal chains-of-thought before answering have produced jumps in capability that pure scaling alone hadn't.

Security implications

Chain-of-thought introduces specific attack surfaces that don't exist for direct-answer models:

Practical takeaways