LAUNCHES

Chroma Releases Context-1: 20B Parameter Model for Multi-Hop Search

M megaone_admin Mar 29, 2026 2 min read
Engine Score 7/10 — Important

Chroma's release of Context-1 is a significant launch for developers, offering advanced agentic search capabilities for multi-hop retrieval and context management. Its actionability and potential industry impact on AI system development are high, despite being reported by a single Tier 2 source.

Editorial illustration for: Chroma Releases Context-1: 20B Parameter Model for Multi-Hop Search

Chroma has released Context-1, a 20 billion parameter agentic search model designed to act as a specialized retrieval subagent for complex, multi-hop queries. Rather than expanding context windows in frontier models, the company behind the popular open-source vector database is taking what it calls “a different, more surgical approach” to retrieval-augmented generation systems.

Context-1 is derived from gpt-oss-20B, a Mixture of Experts (MoE) architecture that Chroma fine-tuned using Supervised Fine-Tuning (SFT) and Reinforcement Learning via CISPO (a staged curriculum optimization). The model operates within an agent harness that allows it to interact with tools including search_corpus (hybrid BM25 + dense search), grep_corpus (regex), and read_document.

The model’s core technical innovation is “Self-Editing Context,” which addresses context window degradation during multi-step searches. As Context-1 gathers information over multiple turns, it reviews its accumulated context and executes a prune_chunks command to discard irrelevant passages, achieving a pruning accuracy of 0.94. The model averages 2.56 tool calls per turn and maintains retrieval quality within a bounded 32k context window.

According to the source material, “Context-1 doesn’t just hit a vector index once. It decomposes the high-level query into targeted subqueries, executes parallel tool calls, and iteratively searches the corpus.” This approach shifts retrieval logic responsibility from developers to the model itself, representing what Chroma describes as “decoupling search from generation.”

Chroma developed a benchmark called “context-1-data-gen” to train and evaluate the model on multi-hop reasoning tasks that require multiple steps to reach ground truth answers. The company positions this as addressing the “lost in the middle” reasoning failures that occur when large amounts of tokens are inserted into prompts, leading to higher latency and increased costs.

Share

Enjoyed this story?

Get articles like this delivered daily. The Engine Room — free AI intelligence newsletter.

Join 500+ AI professionals · No spam · Unsubscribe anytime

M
MegaOne AI Editorial Team

MegaOne AI monitors 200+ sources daily to identify and score the most important AI developments. Our editorial team reviews 200+ sources with rigorous oversight to deliver accurate, scored coverage of the AI industry. Every story is fact-checked, linked to primary sources, and rated using our six-factor Engine Score methodology.

About Us Editorial Policy