Commit
•
eda9a39
1
Parent(s):
80baa50
Update README.md
Browse files
README.md
CHANGED
@@ -21,3 +21,23 @@ configs:
|
|
21 |
- split: train
|
22 |
path: data/train-*
|
23 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
- split: train
|
22 |
path: data/train-*
|
23 |
---
|
24 |
+
|
25 |
+
# OpenRewriteEval
|
26 |
+
|
27 |
+
This dataset contains the data of the `OpenRewriteEval` benchmark used to evaluate the rewriting capabilities of an LLM. The benchmark was originally released in [RewriteLM: An Instruction-Tuned Large Language Model for Text Rewriting
|
28 |
+
](https://arxiv.org/abs/2305.15685) paper by Google. The data has been reuploaded to Hugging Face for convenience from the original GitHub repo: [RewriteLM - GitHub](https://github.com/google-research/google-research/tree/master/rewritelm).
|
29 |
+
|
30 |
+
|
31 |
+
## Citation
|
32 |
+
|
33 |
+
```
|
34 |
+
@misc{shu2023rewritelminstructiontunedlargelanguage,
|
35 |
+
title={RewriteLM: An Instruction-Tuned Large Language Model for Text Rewriting},
|
36 |
+
author={Lei Shu and Liangchen Luo and Jayakumar Hoskere and Yun Zhu and Yinxiao Liu and Simon Tong and Jindong Chen and Lei Meng},
|
37 |
+
year={2023},
|
38 |
+
eprint={2305.15685},
|
39 |
+
archivePrefix={arXiv},
|
40 |
+
primaryClass={cs.CL},
|
41 |
+
url={https://arxiv.org/abs/2305.15685},
|
42 |
+
}
|
43 |
+
```
|