--- license: cc-by-sa-4.0 dataset_info: features: - name: meeting_id dtype: string - name: speaker_id dtype: string - name: audio_id dtype: string - name: audio dtype: audio - name: segments list: - name: end dtype: float64 - name: start dtype: float64 - name: transcript dtype: string - name: words list: - name: end dtype: float64 - name: start dtype: float64 - name: word dtype: string - name: transcript dtype: string splits: - name: dev num_bytes: 14155765669 num_examples: 130 - name: train num_bytes: 74754662936 num_examples: 684 - name: test num_bytes: 13775584735 num_examples: 124 download_size: 120234623488 dataset_size: 102802035597 configs: - config_name: default data_files: - split: dev path: data/dev/* - split: test path: data/test/* - split: train path: data/train/* - config_name: example data_files: - split: train path: data/example/* task_categories: - automatic-speech-recognition - voice-activity-detection language: - fr --- _Note: if the data viewer is not working, use the "example" subset._ # SUMM-RE The SUMM-RE dataset is a collection of transcripts of French conversations, aligned with the audio signal. It is a corpus of meeting-style conversations in French created for the purpose of the SUMM-RE project (ANR-20-CE23-0017). The full dataset is described in Hunter et al. (2024): "SUMM-RE: A corpus of French meeting-style conversations". - **Created by:** Recording and manual correction of the corpus was carried out by the Language and Speech Lab (LPL) at the University of Aix-Marseille, France. - **Funded by:** The National Research Agency of France (ANR) for the SUMM-RE project (ANR-20-CE23-0017). - **Shared by:** LINAGORA (coordinator of the SUMM-RE project) - **Language:** French - **License:** CC BY-SA 4.0 ## Dataset Description Data from the `dev` and `test` splits have been manually transcribed and aligned. Data from the `train` split has been automatically transcribed and aligned with the Whisper pipeline described in Yamasaki et al. (2023): "Transcribing And Aligning Conversational Speech: A Hybrid Pipeline Applied To French Conversations". The audio and transcripts used to evaluate this pipeline, a subset of the `dev` split(*), can be found on [Ortolang](https://www.ortolang.fr/market/corpora/summ-re-asru/). The `dev` and `test` splits of SUMM-RE can be used for the evaluation of automatic speech recognition models and voice activity detection for conversational, spoken French. Speaker diarization can also be evaluated if several tracks of a same meeting are merged together. SUMM-RE transcripts can be used for the training of language models. Each conversation lasts roughly 20 minutes. The number of conversations contained in each split is as follows: - `train`: 210 (x ~20 minutes = ~67 hours) - `dev`: 36 (x ~20 minutes = ~12 hours) - `test`: 37 (x ~20 minutes = ~12 hours) Each conversation contains 3-4 speakers (and in rare cases, 2) and each participant has an individual microphone and associated audio track, giving rise to the following number of tracks for each split: - `train`: 684 (x ~20 minutes = ~226 hours) - `dev`: 130 (x ~20 minutes = ~43 hours) - `test`: 124 (x ~20 minutes = ~41 hours) ## Dataset Structure To visualize an example from the corpus, select the "example" split in the Dataset Viewer. The corpus contains the following information for each audio track: - **meeting_id**, e.g. 001a_PARL, includes: - experiment number, e.g. 001 - meeting order: a|b|c (there were three meetings per experiment) - experiment type: E (experiment) | P (pilot experiment) - scenario/topic: A|B|C|D|E - meeting type: R (reporting) | D (decision) | P (planning) - recording location: L (LPL) | H (H2C2 studio) | Z (Zoom) | D (at home) - **speaker_id** - **audio_id**: meeting_id + speaker_id - **audio**: the audio track for an individual speaker - **segments**: a list of dictionaries where each entry provides the transcription of a segment with timestamps for the segment and each word that it contains. An example is: ```json [ { "start": 0.5, "end": 1.2, "transcript": "bonjour toi", "words": [ { "word": "bonjour", "start": 0.5, "end": 0.9 } { "word": "toi", "start": 0.9, "end": 1.2 } ] }, ... ] ``` - **transcript**: a string formed by concatenating the text from all of the segments (note that those transcripts implicitly include periods of silence where other speakers are speaking in other audio tracks) ## Example Use To load the full dataset ```python import datasets ds = datasets.load_dataset("linagora/SUMM-RE") ``` Use the `streaming` option to avoid downloading the full dataset, when only a split is required: ```python import datasets devset = datasets.load_dataset("linagora/SUMM-RE", split="dev", streaming=True) for sample in devset: ... ``` Load some short extracts of the data to explore the structure: ```python import datasets ds = datasets.load_dataset("linagora/SUMM-RE", "example") sample = ds["train"][0] print(sample) ``` ## Dataset Creation ### Curation Rationale The full SUMM-RE corpus, which includes meeting summaries, is designed to train and evaluate models for meeting summarization. This version is an extract of the full corpus used to evaluate various stages of the summarization pipeline, starting with automatic transcription of the audio signal. ### Source Data The SUMM-RE corpus is an original corpus designed by members of LINAGORA and the University of Aix-Marseille and recorded by the latter. #### Data Collection and Processing For details, see [Hunter et al. (2024)](https://hal.science/hal-04623038/). #### Audio Sampling Rates By default, files recorded through Zoom have a sampling rate of 32000 and other files have a sampling rate of 48000. The sampling rates for exception files are as follows: 44100 = ['071*'] 32000 = ['101*'] 22050 = ['018a_EARZ_055.wav', '018a_EARZ_056.wav', '018a_EARZ_057.wav', '018a_EARZ_058.wav', '020b_EBDZ_017.wav', '020b_EBDZ_053.wav', '020b_EBDZ_057.wav', '020b_EBDZ_063.wav', '027a_EBRH_025.wav', '027a_EBRH_075.wav', '027a_EBRH_078.wav', '032b_EADH_084.wav', '032b_EADH_085.wav', '032b_EADH_086.wav', '032b_EADH_087.wav', '033a_EBRH_091.wav', '033c_EBPH_092.wav', '033c_EBPH_093.wav', '033c_EBPH_094.wav', '034a_EBRH_095.wav', '034a_EBRH_096.wav', '034a_EBRH_097.wav', '034a_EBRH_098.wav', '035b_EADH_088.wav', '035b_EADH_096.wav', '035b_EADH_097.wav', '035b_EADH_098.wav', '036c_EAPH_091.wav', '036c_EAPH_092.wav', '036c_EAPH_093.wav', '036c_EAPH_099.wav', '069c_EEPL_156.wav', '069c_EEPL_157.wav', '069c_EEPL_158.wav', '069c_EEPL_159.wav'] #### Who are the source data producers? Corpus design and production: - University of Aix-Marseille: Océane Granier (corpus conception, recording, annotation), Laurent Prévot (corpus conception, annotatation, supervision), Hiroyoshi Yamasaki (corpus cleaning, alignment and anonymization), Roxanne Bertrand (corpus conception and annotation) with helpful input from Brigitte Bigi and Stéphane Rauzy. - LINAGORA: Julie Hunter, Kate Thompson and Guokan Shang (corpus conception) Corpus participants: - Participants for the in-person conversations were recruited on the University of Aix-Marseille campus. - Participants for the Zoom meetings were recruited through [Prolific](https://www.prolific.com/). ### Annotations Transcripts are not punctuated and all words are in lower case. Annotations follow the conventions laid out in chapter 3 of [The SPPAS Book](https://sppas.org/book_03_annotations.html) by Brigitte Bigi. Transcripts may therefore contain additional annotations in the following contexts: * truncated words, noted as a - at the end of the token string (an ex- example); * noises, noted by a * (not available for some languages); * laughter, noted by a @ (not available for some languages); * short pauses, noted by a +; * elisions, mentioned in parentheses; * specific pronunciations, noted with brackets [example,eczap]; * comments are preferably noted inside braces {this is a comment!}; * comments can be noted inside brackets without using comma [this and this]; * liaisons, noted between = (this =n= example); * morphological variants with , * proper name annotation, like $ John S. Doe $. Note that the symbols * + @ must be surrounded by whitespace. #### Annotation process [More Information Needed] #### Who are the annotators? Principal annotator for `dev`: Océane Granier Principal annotators for `test`: Eliane Bailly, Manon Méaume, Lyne Rahabi, Lucille Rico Additional assistance from: Laurent Prévot, Hiroyoshi Yamasaki and Roxane Bertrand #### Personal and Sensitive Information A portion of the `dev`split has been (semi-automatically) anonymized for the pipeline described in Yamasaki et al. (2023). ## Bias, Risks, and Limitations [More Information Needed] ### Recommendations ## Citations Please cite the papers below if using the dataset in your work. **Description of the full dataset:** Julie Hunter, Hiroyoshi Yamasaki, Océane Granier, Jérôme Louradour, Roxane Bertrand, Kate Thompson and Laurent Prévot (2024): "[SUMM-RE: A corpus of French meeting-style conversations](https://hal.science/hal-04623038/)," TALN 2024. ```bibtex @inproceedings{hunter2024summre, title={SUMM-RE: A corpus of French meeting-style conversations}, author={Hunter, Julie and Yamasaki, Hiroyoshi and Granier, Oc{\'e}ane and Louradour, J{\'e}r{\^o}me and Bertrand, Roxane and Thompson, Kate and Pr{\'e}vot, Laurent}, booktitle={Actes de JEP-TALN-RECITAL 2024. 31{\`e}me Conf{\'e}rence sur le Traitement Automatique des Langues Naturelles, volume 1: articles longs et prises de position}, pages={508--529}, year={2024}, organization={ATALA \& AFPC} } ``` **The Whisper Pipeline:** Hiroyoshi Yamasaki, Jérôme Louradour, Julie Hunter and Laurent Prévot (2023): "[Transcribing and aligning conversational speech: A hybrid pipeline applied to French conversations](https://hal.science/hal-04404777/document)," Workshop on Automatic Speech Recognition and Understanding (ASRU). ```bibtex @inproceedings{yamasaki2023transcribing, title={Transcribing And Aligning Conversational Speech: A Hybrid Pipeline Applied To French Conversations}, author={Yamasaki, Hiroyoshi and Louradour, J{\'e}r{\^o}me and Hunter, Julie and Pr{\'e}vot, Laurent}, booktitle={2023 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU)}, pages={1--6}, year={2023}, organization={IEEE} } ``` (*)The following meetings were used to evaluate the pipeline in Yamasaki et al. (2023): ```python asru = ['018a_EARZ_055', '018a_EARZ_056', '018a_EARZ_057', '018a_EARZ_058', '020b_EBDZ_017', '020b_EBDZ_053', '020b_EBDZ_057', '020b_EBDZ_063', '027a_EBRH_025', '027a_EBRH_075', '027a_EBRH_078', '032b_EADH_084', '032b_EADH_085', '032b_EADH_086', '032b_EADH_087', '033a_EBRH_091', '033a_EBRH_092', '033a_EBRH_093', '033a_EBRH_094', '033c_EBPH_091', '033c_EBPH_092', '033c_EBPH_093', '033c_EBPH_094', '034a_EBRH_095', '034a_EBRH_096', '034a_EBRH_097', '034a_EBRH_098', '035b_EADH_088', '035b_EADH_096', '035b_EADH_097', '035b_EADH_098', '036c_EAPH_091', '036c_EAPH_092', '036c_EAPH_093', '036c_EAPH_099', '069c_EEPL_156', '069c_EEPL_157', '069c_EEPL_158', '069c_EEPL_159'] ``` ## Acknowledgements We gratefully acknowledge support from the Agence Nationale de Recherche for the SUMM-RE project (ANR-20-CE23-0017).