Search

Subscribe to our newsletter

Do not worry we don't spam!

GDPR Compliance

We use cookies to ensure you get the best experience on our website. By continuing to use our site, you accept our use of cookies, Privacy Policy, and Terms of Service.

Nvidia launches Nemotron 3 open models to enable multi-agent systems

Nvidia launched its Nemotron 3 family of open models as it aims to provide an efficient set of large language models that can be used by enterprises to customize and deploy in multi-agent systems.

The company said it is releasing open models, training data and libraries. Nvidia, which doesn't have to worry about monetizing models since it cashes in on GPU sales, is focused on providing tools to build agentic AI systems, which will use multiple LLMs focused on various tasks.

Nvidia is also filling in a major US open model gap. Meta's Llama hasn't been updated as the company has retooled its AI unit and may be focusing on proprietary models.

Nemotron 3 models will come in three sizes--Nano, Super and Ultra. Nemotron 3 Nano provides 4x higher throughput than Nemotron 2 and delivers the most tokens per second for multi-agent systems at scale.

The Nemotron 3 Super and Ultra models use a hybrid latent mixture-of-experts (MoE) architecture.

Nvidia outlined multiple early adopters ranging from Accenture to CrowdStrike to Oracle, Palantir and ServiceNow to name a few.

The game plan for Nvidia is to use Nemotron 3 to give developers options to mix and match open models with proprietary offerings to optimize costs.

Nemotron 3 Nano is available now on Hugging Face and inference service providers including Baseten, DeepInfra, Fireworks, FriendliAI, OpenRouter and Together AI. Nemotron is also available on platforms from Couchbase, DataRobot, H2O.ai, JFrog, Lambda and UiPath. And Nemotron 3 Nano is available on AWS via Amazon Bedrock with availability on Google Cloud, CoreWeave, Crusoe, Microsoft Foundry, Nebius, Nscale and Yotta on deck.

According to Nvidia, Nemotron 3 Super and Ultra will be available in the first half of 2026.

Prev Article
Browse More News
Next Article
Browse More News

Related to this topic: