AI Model

NVIDIA: Nemotron-4 340B Instruct

NVIDIA: Nemotron-4 340B Instruct logoNVIDIA
Text Generation
About Nemotron-4 340B Instruct

Nemotron-4-340B-Instruct is an English-language chat model optimized for synthetic data generation. This large language model (LLM) is a fine-tuned version of Nemotron-4-340B-Base, designed for single and multi-turn chat use-cases with a 4,096 token context length.

The base model was pre-trained on 9 trillion tokens from diverse English texts, 50+ natural languages, and 40+ coding languages. The instruct model underwent additional alignment steps:

1. Supervised Fine-tuning (SFT) 2. Direct Preference Optimization (DPO) 3. Reward-aware Preference Optimization (RPO)

The alignment process used approximately 20K human-annotated samples, while 98% of the data for fine-tuning was synthetically generated. Detailed information about the synthetic data generation pipeline is available in the [technical report](https://arxiv.org/html/2406.11704v1).

Specifications
Provider
NVIDIA
Context Length
4,096 tokens
Input Types
text
Output Types
text
Category
Other
Added
6/23/2024

Frequently Asked Questions

Common questions about Nemotron-4 340B Instruct

Use Nemotron-4 340B Instruct and 200+ more models

Access all the best AI models in one platform. No API keys, no switching between apps.