How Context Works
Every time you send a message in ZeroTwo, the platform constructs a context payload and sends the entire thing to the selected model. This payload contains:- The full conversation history — every message from both you and the AI in the current chat, from the very first message to the one just before your current input
- Your current message — the new prompt you are sending
- System context — any active project instructions, custom instructions from Settings → Personalization, and relevant memory entries
- Tool results — if web search, file attachments, or other tools were used, their outputs are included
Context Window Limits
Every AI model has a maximum context window — the total number of tokens (roughly, words and punctuation) it can process in a single request. When a conversation grows large enough to exceed this limit, older messages are dropped to make room for new content. ZeroTwo displays context window information in the model capabilities tooltip in the model picker.Common Context Windows
| Model | Context Window | Approximate Length |
|---|---|---|
| Gemini 2.5 Pro | 1,000,000 tokens | ~750,000 words |
| Claude Sonnet 4.6 | 200,000 tokens | ~150,000 words |
| Claude Opus 4.6 | 200,000 tokens | ~150,000 words |
| GPT-4o | 128,000 tokens | ~96,000 words |
| GPT-5 | 128,000 tokens | ~96,000 words |
| Mistral Large | 128,000 tokens | ~96,000 words |
| DeepSeek Chat | 64,000 tokens | ~48,000 words |
| Smaller models | 8,000–32,000 tokens | ~6,000–24,000 words |
When Context Gets Too Long
If a conversation grows beyond the model’s context window, ZeroTwo handles it by truncating the oldest messages to make room for new content. The most recent messages are always preserved. Signs that context truncation has occurred:- The AI seems to have “forgotten” something mentioned early in the conversation
- The AI asks for information you already provided
- Responses lack awareness of earlier decisions, constraints, or established context
- Start a new chat — open a new conversation and summarize the key context in your opening message (see the tip below)
- Use a larger-context model — switch to Gemini 2.5 Pro (1M tokens) or Claude Sonnet 4.6 (200k tokens) for very long sessions
- Break long tasks into focused chats within a project — instead of one giant conversation, use a series of shorter focused chats that share context via the project instruction
Starting Fresh
Starting a new chat gives the AI a completely clean slate — no prior conversation history. This is intentional and useful:- Switch topics cleanly. When you are done with one task and starting another, a new chat prevents any spillover context from the previous conversation affecting responses.
- Reset incorrect context. If a conversation went in the wrong direction and is now full of noise, a fresh chat with a refined prompt often produces significantly better results.
- Conserve premium quota. Very long conversations generate large context payloads, which consume more tokens per message. Starting a new chat resets this.
Persistent Context Across Sessions
ZeroTwo provides several mechanisms for context that persists beyond individual chats, so the AI always has background information about you and your work.Memory
ZeroTwo’s memory system automatically learns personal facts about you from your conversations — your role, technical preferences, writing style, and more. These facts are categorized as preferences, general facts, or profile information and are automatically injected into the context of every new conversation.- Free plan: Up to 5 memory entries
- Pro+: Unlimited memory entries
Project Instructions
Each project has an optional instruction set — a system prompt that is automatically injected into every chat within that project. This is ideal for maintaining consistent AI behavior across many sessions about the same topic. Example project instruction:“You are helping me build a SaaS product called Acme. The tech stack is Next.js, Supabase, and Stripe. Always write TypeScript. Prefer functional React components. Assume production-grade code quality.”Every new chat in that project starts with this context without requiring you to re-enter it. This is the most powerful form of persistent context in ZeroTwo — it transforms a collection of individual chats into a coherent ongoing workspace.
Custom Instructions
Settings → Personalization → Custom Instructions lets you set global instructions that apply to all new chats across your entire account, regardless of which project they are in. There are two fields:- “What should ZeroTwo know about you?” — your role, background, preferences, technical context
- “How should ZeroTwo respond?” — tone, format, length, specific behaviors
How Switching Models Affects Context
When you switch models mid-conversation, the full conversation history is sent to the new model. The new model has complete context of everything discussed before the switch. Things to be aware of:- Different models may interpret or reference earlier content with slightly different nuances
- Switching to a model with a smaller context window than the current model may result in earlier messages being truncated if the conversation is already long
- Not all models support all tools — switching to a model that lacks tool-use support disables active tools like Web Search and Agent Mode
Switching models mid-conversation does not lose any context. The new model receives the complete history. Check the context window size in the model capabilities tooltip before switching on very long conversations.
Context Flow Summary
Every message you send triggers this payload construction:Related Pages
- Memory: How It Works — the persistent memory system in detail
- Model Picker — context window sizes and model capabilities
- Projects Overview — project-level instructions and persistent context
- Answer Quality and Limitations — why responses vary and how to improve them

