Datasets:
File size: 12,889 Bytes
d93509c b08732d d93509c e2e94ea b08732d d93509c 838b419 d93509c ae3b547 f8281b1 6243051 d93509c 2effcd4 d93509c 365cc7d d93509c 365cc7d d93509c 2308981 d8526c5 2308981 1dcc886 4d6d790 1dcc886 4d6d790 1dcc886 4d6d790 1dcc886 d8526c5 d93509c d8526c5 aa79e6e 4d6d790 aa79e6e 4d6d790 aa79e6e d93509c d8526c5 d93509c 2463ddb |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 |
---
annotations_creators:
- expert-generated
language_creators:
- crowdsourced
language:
- en-UK
- en-US
- fr-FR
- fr
license:
- cc-by-sa-4.0
multilinguality:
- translation
pretty_name: DiaBLa
size_categories:
- 1K<n<10K
source_datasets:
- original
task_categories:
- translation
- machine translation
task_ids: []
---
# Dataset Card for DiaBLa: Bilingual dialogue parallel evaluation set
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
- [Languages](#languages)
- [Dataset Structure](#dataset-structure)
- [Data Instances](#data-instances)
- [Data Fields](#data-fields)
- [Data Splits](#data-splits)
- [Dataset Creation](#dataset-creation)
- [Curation Rationale](#curation-rationale)
- [Source Data](#source-data)
- [Annotations](#annotations)
- [Personal and Sensitive Information](#personal-and-sensitive-information)
- [Considerations for Using the Data](#considerations-for-using-the-data)
- [Social Impact of Dataset](#social-impact-of-dataset)
- [Discussion of Biases](#discussion-of-biases)
- [Other Known Limitations](#other-known-limitations)
- [Additional Information](#additional-information)
- [Dataset Curators](#dataset-curators)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
- [Contributions](#contributions)
## Dataset Description
- **Homepage:** [almanach.inria.fr/software_and_resources/custom/DiaBLa-en.html](http://almanach.inria.fr/software_and_resources/custom/DiaBLa-en.html)
- **Repository:** [github.com/rbawden/DiaBLa-dataset](https://github.com/rbawden/DiaBLa-dataset)
- **Paper:** [Bawden et al. (2021). DiaBLa: A Corpus of Bilingual Spontaneous Written Dialogues for Machine Translation. Language Resources and Evaluation(55). Pages 635–660. Springer Verlag. 10.1007/s10579-020-09514-4.](https://hal.inria.fr/hal-03021633)
- **Point of contact:** rachel.bawden[at]inria.fr
### Dataset Summary
The dataset is an English-French dataset for the evaluation of Machine Translation (MT) for informal, written bilingual dialogue.
The dataset contains 144 spontaneous dialogues (5,700+ sentences) between native English and French speakers, mediated by one of two neural MT systems in a range of role-play settings. See below for some basic statistics. The dialogues are accompanied by fine-grained sentence-level judgments of MT quality, produced by the dialogue participants themselves, as well as by manually normalised versions and reference translations produced a posteriori. See here for information about evaluation.
The motivation for the corpus is two-fold: to provide:
- a unique resource for evaluating MT models for dialogue (i.e. in context)
- a corpus for the analysis of MT-mediated communication
### Supported Tasks and Leaderboards
[More Information Needed]
### Languages
English (mainly UK) and French
## Dataset Structure
### Data Instances
- **Size of downloaded dataset files:** 37 MB
- **Number of parallel utterances:** 5748
Each example is highly annotated and is associated with dialogue context. An example from the test set looks as follows (only the first and last utterances of the dialogue history are shown for readability purposes):
```
{
"id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_25",
"mt": "Tu m'en veux pour \u00e7a ?",
"norm": "",
"orig": "Are you blaming me for this?",
"ref": "C'est moi que vous critiquez pour \u00e7a\u00a0?",
"utterance_meta": {
"eval_judgment": "medium",
"eval_verbatim": "",
"eval_problems": [
"coherence"
],
"lang": "english"
},
"dialogue_meta": {
"start_time": "2018-04-25T16:20:36.087170",
"end_time": "",
"translation_model": "baseline",
"final_evaluation_user1": {
"style": "average",
"coherence": "average",
"grammaticality": "good",
"meaning": "average",
"word_choice": "average"
},
"final_evaluation_user2": {
"style": "",
"coherence": "",
"grammaticality": "",
"meaning": "",
"word_choice": ""
},
"scenario": [
[
"You are both stuck in a lift at work.",
"Vous \u00eates tous les deux bloqu\u00e9(e)s dans un ascenseur au travail."
],
[
"You are an employee and you are with your boss.",
"Vous \u00eates un(e) employ\u00e9(e) et vous \u00eates avez votre patron(ne)"
],
[
"You are the boss and are with an employee.",
"Vous \u00eates le ou la patron(ne) et vous \u00eates avec un(e) employ\u00e9(e)"
]
],
"user1": {
"role_num": 1,
"role": [
"You are an employee and you are with your boss.",
"Vous \u00eates un(e) employ\u00e9(e) et vous \u00eates avez votre patron(ne)"
],
"initiated_dialogue": true,
"turn_number": 2,
"lang": "french"
},
"user2": {
"role_num": 2,
"role": [
"You are the boss and are with an employee.",
"Vous \u00eates le ou la patron(ne) et vous \u00eates avec un(e) employ\u00e9(e)"
],
"initiated_dialogue": false,
"turn_number": 1,
"lang": "english"
}
},
"dialogue_history": [
{
"id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_0",
"orig": "We appear to have stopped moving.",
"norm": "",
"mt": "On semble avoir arr\u00eat\u00e9 de bouger.",
"ref": "J'ai l'impression qu'on s'est arr\u00eat\u00e9s.",
"utterance_meta": {
"eval_judgment": "medium",
"eval_verbatim": "",
"eval_problems": [
"style"
],
"lang": "english"
}
},
[...]
{
"id": "dialogue-2018-04-25T16-20-36.087170_french_english_1_2_24",
"orig": "La sonnerie s'est arr\u00eat\u00e9, je pense que personne ne va nous r\u00e9pondre.",
"norm": "",
"mt": "The ringing stopped, and I don't think anyone's gonna answer us.",
"ref": "It stopped ringing. I don't think anybody's going to reply.",
"utterance_meta": {
"eval_judgment": "perfect",
"eval_verbatim": "",
"eval_problems": [],
"lang": "french"
}
}
]
}
```
### Data Fields
#### plain_text
- `id`: a `string` feature.
- `orig`: a `string` feature.
- `norm`: a `string` feature.
- `mt`: a `string` feature.
- `ref`: a `string` feature.
- `utterance_meta`: a dictionary feature containing:
- `eval_judgment`: a `string` feature.
- `eval_verbatim`: a `string` feature.
- `eval_problems`: a list feature containing:
- up to 5 `string` features.
- `lang`: a `string` feature.
- `dialogue_meta`: a dictionary feature containing:
- `start_time` : a `string` feature.
- `end_time`: a `string` feature.
- `translation_model`: a `string` feature.
- `final_evaluation_user1`: a dictionary feature containing:
- `style`: a `string` feature.
- `coherence`: a `string` feature.
- `grammaticality`: a `string` feature.
- `meaning`: a `string` feature.
- `word_choice`: a `string` feature.
- `final_evaluation_user2`: a dictionary feature containing:
- `style`: a `string` feature.
- `coherence`: a `string` feature.
- `grammaticality`: a `string` feature.
- `meaning`: a `string` feature.
- `word_choice`: a `string` feature.
- `scenario`: a list feature containing
- 3 lists each containing 2 `string` features.
- `user1`: a dictionary feature containing:
- `role_num`: an `int` feature.
- `role`: a list feature containing:
- 2 `string` features.
- `initiated_dialogue`: a `bool` feature.
- `turn_number`: an `int` value.
- `lang`: a `string` value.
- `user2`: a dictionary feature containing:
- `role_num`: an `int` feature.
- `role`: a list feature containing:
- 2 `string` features.
- `initiated_dialogue`: a `bool` feature.
- `turn_number`: an `int` value.
- `lang`: a `string` value.
- `dialogue_history`: a list feature containing:
- dictionary features containing:
- `id`: a `string` feature.
- `orig`: a `string` feature.
- `norm`: a `string` feature.
- `mt`: a `string` feature.
- `ref`: a `string` feature.
- `utterance_meta`: a dictionary feature containing:
- `eval_judgment`: a `string` feature.
- `eval_verbatim`: a `string` feature.
- `eval_problems`: a list feature containing:
- up to 5 `string` features.
- `lang`: a `string` feature.
### Data Splits
DiaBLa is a test set only.
| name |test |
|----------|------:|
|plain_text| 5748|
## Dataset Creation
### Curation Rationale
[More Information Needed]
### Source Data
#### Initial Data Collection and Normalization
Original data was collected through a [dedicated online chat platform](https://github.com/rbawden/diabla-chat-interface) and involved native speakers of English and of French. As well as producing the original text, participants also annotated the quality of the machine-translated outputs of their partners' utterances (which they saw instead of their partners' original text) based on their monolingual intuitions and the dialogue context.
Each dialogue is assigned one of 12 role-play scenarios and where appropriate each participant is assigned a role to play in the dialogue.
#### Who are the source language producers?
The source text producers were native French and native English volunteers (mainly British English). See the paper for very basic information concerning their backgrounds (age categories and experience in NLP).
### Annotations
#### Annotation process
On top of the original dialogue text (a mixture of utterances in English and in French), the following "annotations" are provided:
- machine translated version of the original text (produced in real time and presented during the dialogue), produced by one of two MT systems, both trained using [Marian](https://github.com/marian-nmt/marian).
- judgments of MT quality by participants (overall quality, particular problems, verbatim comments)
- manually produced normalised version of the original text (for spelling mistakes, grammatical errors, missing punctuation, etc.)
- manually produced reference translations
#### Who are the annotators?
The judgments of MT quality were produced by the dialogue participants themselves in real time. The normalised version of the text and the reference translations were manually produced by the authors of the paper. Translations were always done into the translator's native language and all translations were verified and post-edited by a bilingual English-French speaker.
### Personal and Sensitive Information
A priori the dataset does not contain personal and sensitive information. Participants were instructed not to give any personal information and to assume the roles assigned in the role play scenario. Usernames were anonymised prior to distribution and any mention of either usernames or real names in the dialogues were replaced by generic names of the same gender as the participant. Only basic user information was collected to get an idea of the distribution of participants and to potentially see how multilingual ability influences quality judgments (rough age categories, experience in NLP or research, native languages, familiarity with the other language (either English or French), other languages spoken and gender). Gender was included because it is an important factor in translation (particularly for the direction English-to-French), and this was explained in advance to the participants in the FAQs.
## Considerations for Using the Data
### Social Impact of Dataset
[More Information Needed]
### Discussion of Biases
[More Information Needed]
### Other Known Limitations
[More Information Needed]
## Additional Information
### Dataset Curators
The dataset was collected by Rachel Bawden, Eric Bilinski, Thomas Lavergne and Sophie Rosset (see citation below).
### Licensing Information
The dataset is available under a CC BY-SA 4.0 licence.
### Citation Information
If you use or are inspired by this dataset, please cite:
```
@article{bawden_DiaBLa:-A-Corpus-of_2021,
author = {Bawden, Rachel and Bilinski, Eric and Lavergne, Thomas and Rosset, Sophie},
doi = {10.1007/s10579-020-09514-4},
title = {DiaBLa: A Corpus of Bilingual Spontaneous Written Dialogues for Machine Translation},
year = {2021},
journal = {Language Resources and Evaluation},
publisher = {Springer Verlag},
volume = {55},
pages = {635--660},
url = {https://hal.inria.fr/hal-03021633},
pdf = {https://hal.inria.fr/hal-03021633/file/diabla-lre-personal-formatting.pdf},
}
```
### Contributions
This dataset was added by Rachel Bawden [@rbawden](https://github.com/rbawden). |