TOOL UPDATES

NVIDIA Releases 88B Parameter Model Using Neural Architecture Search

M megaone_admin Mar 26, 2026 1 min read
Engine Score 8/10 — Important

The release of an 88B parameter model by NVIDIA on Hugging Face is a significant development for the AI research and developer community. This highly actionable model provides a new tool, despite the news originating from a secondary source like Reddit.

Editorial illustration for: NVIDIA Releases 88B Parameter Model Using Neural Architecture Search

NVIDIA has released gpt-oss-puzzle-88B, an 88-billion parameter language model derived from OpenAI’s gpt-oss-120b using a post-training neural architecture search framework called Puzzle. The model is designed to improve inference efficiency for reasoning-heavy workloads while maintaining accuracy across different reasoning budgets.

The model represents a deployment-optimized version of a larger base model, using NVIDIA’s Puzzle framework to reduce computational requirements. According to the model documentation, gpt-oss-puzzle-88B is “specifically optimized for long-context and short-context” applications, though the original source text appears truncated in the available materials.

The technical implementation includes specific tokenization configurations, with the model using “<|return|>” as an end-of-sequence token and “<|endoftext|>” as a padding token. The model’s chat template supports additional parameters including “builtin_tools” (which can contain “browser” and/or “python”), “model_identity” for describing the model, and “reasoning_effort” with a default setting of “medium.”

The model architecture incorporates TypeScript-style parameter specifications for tool integration, suggesting it’s designed for applications requiring structured interaction with external tools and services. The template system includes support for rendering complex parameter types including arrays, objects, and union types.

This release follows NVIDIA’s broader strategy of optimizing large language models for specific deployment scenarios. The Puzzle framework appears to use neural architecture search techniques to create smaller, more efficient versions of existing models while preserving their core capabilities for reasoning tasks.

Share

Enjoyed this story?

Get articles like this delivered daily. The Engine Room — free AI intelligence newsletter.

Join 500+ AI professionals · No spam · Unsubscribe anytime

M
MegaOne AI Editorial Team

MegaOne AI monitors 200+ sources daily to identify and score the most important AI developments. Our editorial team reviews 200+ sources with rigorous oversight to deliver accurate, scored coverage of the AI industry. Every story is fact-checked, linked to primary sources, and rated using our six-factor Engine Score methodology.

About Us Editorial Policy