Skip to main content
Reasoning models are AI models with extended thinking capabilities. Rather than producing an answer immediately, they work through a problem step by step — exploring approaches, checking their reasoning, and refining their thinking before producing a final response. The result is dramatically better accuracy on complex problems.

What makes reasoning models different

Standard models generate responses token by token in a single forward pass. Reasoning models have an additional “thinking” phase: an extended internal reasoning process that runs before the final answer is produced. This thinking process is often visible as a collapsed section in ZeroTwo — you can expand it to see how the model worked through the problem. The tradeoff: reasoning models are slower and typically use more of your premium quota, but they’re significantly more accurate on hard problems where standard models stumble.

Available reasoning models in ZeroTwo

ModelProviderNotes
o3OpenAIAdvanced reasoning, among the most capable for math and logic
o4-miniOpenAIFaster and more efficient reasoning; strong balance of speed vs. depth
DeepSeek ReasonerDeepSeekExcellent for math, logical proofs, and structured analysis
Claude Sonnet / Opus with extended thinkingAnthropicConfigurable thinking depth via thinking level slider
Gemini 2.5 ProGoogleBuilt-in reasoning; also multimodal

Thinking levels

Some models — especially Claude — expose a thinking level slider in the ZeroTwo prompt bar when selected. This controls how much reasoning the model does before responding:
Thinking levelBehaviorBest for
LowLight reasoning, quick responseProblems that need some structured thinking but not exhaustive analysis
MediumBalanced depth and speedMost complex tasks — good default
HighDeep, extended reasoning; significantly slowerThe hardest problems where accuracy is paramount
Higher thinking levels use more of your premium message quota and take more time, but produce more thorough and accurate output for genuinely hard problems.
Not all reasoning models expose a thinking level slider. For models like o3 and o4-mini, the reasoning depth is handled internally by the model. For Claude, the slider gives you explicit control.

When to use reasoning models

Reasoning models provide the most benefit for tasks that require careful, multi-step thinking: Use a reasoning model for:
  • Complex math problems: multi-step calculations, proofs, statistics, algebra
  • Hard coding challenges: algorithm design, complex debugging, system architecture decisions, performance optimization
  • Logical reasoning: puzzles, deductive inference, formal logic
  • Research synthesis: pulling together insights from multiple sources or a long document into a coherent structured analysis
  • Detailed analysis: when you want the model to carefully consider multiple angles before reaching a conclusion
  • Planning tasks: anything where the right answer requires working through dependencies and ordering carefully
Examples:
  • "Prove that the sum of the first n odd numbers equals n²."
  • "Debug this recursive algorithm that's producing incorrect output for edge cases: [code]"
  • "Analyze this 50-page business strategy document and identify the three most significant risks, with supporting evidence from the text."
  • "Design a database schema for a multi-tenant SaaS application with these requirements: [requirements]"

When NOT to use reasoning models

Reasoning models are overkill — and slower — for tasks that don’t require deep analysis: Use a standard model for:
  • Simple questions with straightforward answers
  • Quick drafts and brainstorming
  • Conversational exchanges
  • Reformatting or transforming text
  • Simple code snippets
  • Summarizing short content
Standard models are faster, just as good for easy tasks, and don’t consume as much of your premium quota.

Tradeoffs at a glance

AspectStandard modelsReasoning models
Response timeSecondsSeconds to minutes
Quota cost1 premium message1+ premium messages
Simple tasksExcellentOverkill
Complex reasoningGoodExcellent
Math and logicAdequateSignificantly better
Multi-step codingGoodSignificantly better
Start with a standard model. Switch if needed. For a new task, try a standard model like Claude Sonnet 4.6 or GPT-4o first. If the answer seems off, shallow, or the problem is genuinely complex, switch to a reasoning model and try again. This strategy saves quota without sacrificing quality when it matters.

Viewing the thinking process

When a reasoning model is working on a problem, ZeroTwo displays a “Thinking…” indicator. Once complete, you can expand a Thinking section in the response to see the model’s internal reasoning process — the steps it took, the approaches it considered, and how it arrived at its answer. This is useful for:
  • Verifying the model understood the problem correctly
  • Learning from the model’s problem-solving approach
  • Catching reasoning errors before acting on the conclusion