Update README.md
Browse files
README.md
CHANGED
@@ -20,6 +20,9 @@ Tulu is a series of language models that are trained to act as helpful assistant
|
|
20 |
Tulu V2 DPO 70B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
|
21 |
This model is a strong alternative to Llama 2 70b Chat.
|
22 |
|
|
|
|
|
|
|
23 |
|
24 |
## Model description
|
25 |
|
@@ -137,12 +140,13 @@ The following hyperparameters were used during DPO training:
|
|
137 |
If you find Tulu 2 is useful in your work, please cite it with:
|
138 |
|
139 |
```
|
140 |
-
@misc{
|
141 |
-
|
142 |
-
|
143 |
-
|
144 |
-
|
145 |
-
|
|
|
146 |
}
|
147 |
```
|
148 |
|
|
|
20 |
Tulu V2 DPO 70B is a fine-tuned version of Llama 2 that was trained on on a mix of publicly available, synthetic and human datasets using [Direct Preference Optimization (DPO)](https://arxiv.org/abs/2305.18290).
|
21 |
This model is a strong alternative to Llama 2 70b Chat.
|
22 |
|
23 |
+
For more details, read the paper: [Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2
|
24 |
+
](https://arxiv.org/abs/2311.10702).
|
25 |
+
|
26 |
|
27 |
## Model description
|
28 |
|
|
|
140 |
If you find Tulu 2 is useful in your work, please cite it with:
|
141 |
|
142 |
```
|
143 |
+
@misc{ivison2023camels,
|
144 |
+
title={Camels in a Changing Climate: Enhancing LM Adaptation with Tulu 2},
|
145 |
+
author={Hamish Ivison and Yizhong Wang and Valentina Pyatkin and Nathan Lambert and Matthew Peters and Pradeep Dasigi and Joel Jang and David Wadden and Noah A. Smith and Iz Beltagy and Hannaneh Hajishirzi},
|
146 |
+
year={2023},
|
147 |
+
eprint={2311.10702},
|
148 |
+
archivePrefix={arXiv},
|
149 |
+
primaryClass={cs.CL}
|
150 |
}
|
151 |
```
|
152 |
|