teknium commited on
Commit
942b88e
1 Parent(s): f43d97f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -28,7 +28,7 @@ Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained o
28
 
29
  The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
30
 
31
- This is the SFT + DPO version of Mixtral Hermes 2, we are also be providing an SFT only version, for people to find which works best for them, which can be found here: https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT
32
 
33
  ## We are grateful to Together.ai for sponsoring our compute during the many experiments both training Mixtral and working on DPO!
34
 
 
28
 
29
  The model was trained on over 1,000,000 entries of primarily GPT-4 generated data, as well as other high quality data from open datasets across the AI landscape, achieving state of the art performance on a variety of tasks.
30
 
31
+ This is the SFT + DPO version of Mixtral Hermes 2, we have also released an SFT only version, for people to find which works best for them, which can be found here: https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-SFT
32
 
33
  ## We are grateful to Together.ai for sponsoring our compute during the many experiments both training Mixtral and working on DPO!
34