File size: 862 Bytes
3967cb4
 
 
 
78d4f5e
3967cb4
78d4f5e
3967cb4
78d4f5e
3967cb4
e22e1fb
3967cb4
78d4f5e
3967cb4
78d4f5e
3967cb4
78d4f5e
3967cb4
78d4f5e
 
 
 
 
3967cb4
78d4f5e
3967cb4
78d4f5e
3967cb4
78d4f5e
3967cb4
78d4f5e
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
---
library_name: transformers
tags: []
---
# MarcoroCapy-7B

This model is a DPO fine tune of [mlabonne/Marcoro14-7B-slerp](https://huggingface.co/mlabonne/Marcoro14-7B-slerp) on [argilla/distilabel-capybara-dpo-7k-binarized](https://huggingface.co/datasets/argilla/distilabel-capybara-dpo-7k-binarized)

<div align="center">  

![image/webp](https://cdn-uploads.huggingface.co/production/uploads/6455cc8d679315e4ef16fbec/oey_JDcpqQ0Lw-7KH0AIE.webp)

[<img src="https://raw.githubusercontent.com/argilla-io/distilabel/main/docs/assets/distilabel-badge-dark.png" alt="Built with Distilabel" width="200" height="32"/>](https://github.com/argilla-io/distilabel)

</div>

## Process

+ Realigned the chat template to ChatML 
+ Completed 1 Epoch
+ 5e-5 learning rate
+ Training time was about 4.5 hours on 1 H100
+ Cost was ~$20

## GGUF

TODO

## Evaluations

TODO