About Hermes 2 Mistral 7B DPO
This is the flagship 7B Hermes model, a Direct Preference Optimization (DPO) of [Teknium/OpenHermes-2.5-Mistral-7B](/models/teknium/openhermes-2.5-mistral-7b). It shows improvement across the board on all benchmarks tested - AGIEval, BigBench Reasoning, GPT4All, and TruthfulQA.
The model prior to DPO was trained on 1,000,000 instructions/chats of GPT-4 quality or better, primarily synthetic data as well as other high quality datasets.
Specifications
- Provider
- Nousresearch
- Context Length
- 8,192 tokens
- Input Types
- text
- Output Types
- text
- Category
- Mistral
- Added
- 2/21/2024