About OpenHermes 2.5 Mistral 7B
A continuation of [OpenHermes 2 model](/models/teknium/openhermes-2-mistral-7b), trained on additional code datasets. Potentially the most interesting finding from training on a good ratio (est. of around 7-14% of the total dataset) of code instruction was that it has boosted several non-code benchmarks, including TruthfulQA, AGIEval, and GPT4All suite. It did however reduce BigBench benchmark score, but the net gain overall is significant.
Specifications
- Provider
- Teknium
- Context Length
- 4,096 tokens
- Input Types
- text
- Output Types
- text
- Category
- Mistral
- Added
- 11/20/2023