Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
ayoubkirouane
/
Mistral-SLERP-Merged7B-DPO
like
0
Text Generation
PEFT
Safetensors
HuggingFaceH4/ultrafeedback_binarized
ayoubkirouane/Orca-Direct-Preference-Optimization
trl
dpo
unsloth
conversational
License:
apache-2.0
Model card
Files
Files and versions
Community
Use this model
main
Mistral-SLERP-Merged7B-DPO
File size: 26 Bytes
4907099
1
2
3
4
{
"<|im_end|>"
:
32000
}