|
--- |
|
license: cc-by-4.0 |
|
task_categories: |
|
- audio-classification |
|
language: |
|
- en |
|
tags: |
|
- speech |
|
- speech-classifiation |
|
- text-to-speech |
|
- spoofing |
|
- accents |
|
|
|
pretty_name: ARCTIC-HS |
|
size_categories: |
|
- 10K<n<100K |
|
--- |
|
|
|
# ARCTIC-HS |
|
|
|
An extension of the [CMU_ARCTIC](http://festvox.org/cmu_arctic/) and [L2-ARCTIC](https://psi.engr.tamu.edu/l2-arctic-corpus/) datasets for synthetic speech detection using text-to-speech, featured in the paper **Synthetic speech detection with Wav2Vec 2.0 in various language settings**. |
|
|
|
This dataset is 1 of 3 used in the paper, the others being: |
|
- [FLEURS-HS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs) |
|
- the default train, dev and test sets |
|
- [FLEURS-HS VITS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs-vits) |
|
- test set containing (generally) more difficult synthetic samples |
|
- separated due to different licensing |
|
|
|
## Dataset Details |
|
|
|
### Dataset Description |
|
|
|
The dataset features 3 parts obtained from the 2 original datasets: |
|
|
|
- CMU non-native US English speakers |
|
- CMU native US English speakers |
|
- L2 non-native English speakers |
|
|
|
|
|
The original ARCTIC samples are used as `human` samples, while `synthetic` samples are generated using [Google Cloud Text-To-Speech](https://cloud.google.com/text-to-speech). |
|
|
|
The resulting `symmetric` datasets features exactly twice the samples of the original ones, but we also provide: |
|
|
|
- human samples that couldn't be paired |
|
- 4 speakers in entirety we couldn't pair with a TTS voice |
|
- a small amount of utterances unrelated to the A and B ARCTIC samples |
|
- synthetic samples that couldn't be paired |
|
- mostly when a human speaker didn't read the B ARCTIC samples |
|
|
|
|
|
- **Curated by:** [KONTXT by RealNetworks](https://realnetworks.com/kontxt) |
|
- **Funded by:** [RealNetworks](https://realnetworks.com/) |
|
- **Language(s) (NLP):** English |
|
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) for the code, [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) for the dataset, however: |
|
- the human part of the dataset is under a **custom CMU license** |
|
- it should be compatible with **CC BY 4.0** |
|
- the human part of the L2 dataset is under **CC BY-NC 4.0** |
|
|
|
### Dataset Sources |
|
|
|
The original ARCTIC sets were downloaded from their original sources. |
|
|
|
- **CMU_ARCTIC Repository:** [festvox.org](http://festvox.org/cmu_arctic/) |
|
- **L2-ARCTIC Repository:** [tamu.edu](https://psi.engr.tamu.edu/l2-arctic-corpus/) |
|
- **CMU_ARCTIC Paper:** [cmu.edu](https://www.cs.cmu.edu/~awb/papers/ssw5/arctic.pdf) |
|
- **L2-ARCTIC Paper:** [tamu.edu](https://psi.engr.tamu.edu/wp-content/uploads/2018/08/zhao2018interspeech.pdf) |
|
|
|
- **Paper:** Synthetic speech detection with Wav2Vec 2.0 in various language settings |
|
|
|
## Uses |
|
|
|
This dataset is best used as a test set for accents. Each sample contains an `Audio` feature, and a label: `human` or `synthetic`. |
|
|
|
### Direct Use |
|
|
|
The following snippet of code demonstrates loading the CMU non-native US English speaker part of the dataset: |
|
|
|
```python |
|
from datasets import load_dataset |
|
|
|
arctic_hs = load_dataset( |
|
"realnetworks-kontxt/arctic-hs", |
|
"cmu_non-us", |
|
split="test", |
|
trust_remote_code=True, |
|
) |
|
``` |
|
|
|
To load a different part, change `cmu_non-us` into one of the following: |
|
- `cmu_us` for CMU native US English speakers |
|
- `l2` for L2 non-native speakers |
|
|
|
This dataset only has a `test` split. |
|
|
|
To load only the paired samples, append `_symmetric` to the name. For example, `cmu_non-us` will load the test set also containing human and synthetic samples without their counterpart, while `cmu_non-us_symmetric` will only load samples where there is both a human and synthetic variant. This is useful if you want to have perfectly balanced labels within speakers, and if you wish to exclude speakers for which there are no TTS counterparts at all. |
|
|
|
The `trust_remote_code=True` parameter is necessary because this dataset uses a custom loader. To check out which code is being ran, check out the [loading script](./arctic-hs.py). |
|
|
|
## Dataset Structure |
|
|
|
The dataset data is contained in the [data directory](https://huggingface.co/datasets/realnetworks-kontxt/arctic-hs/tree/main/data). |
|
|
|
There exists 1 directory per part. |
|
|
|
Within those directories, there are 2 further directories: |
|
- `splits` |
|
- `pairs` |
|
|
|
Within the `splits` folder, there is 1 file per split: |
|
- `test.tar.gz` |
|
|
|
Those `.tar.gz` files contain 2 directories: |
|
- `human` |
|
- `synthetic` |
|
|
|
Each of these directories contain `.wav` files. Keep in mind that these directories can't be merged as they share most of their file names. An identical file name implies a speaker-voice pair, ex. `human/arctic_a0001.wav` and `synthetic/arctic_a0001.wav`. |
|
|
|
|
|
The `pairs` folder contains a list of file names within each speaker, and whether or not there is a human-synthetic pair. Based on that metadata we determine which samples appear in `symmetric` datasets. |
|
|
|
Back to the part directories, each contain 2 metadata files, which are not used in the loaded dataset, but might be useful to researchers: |
|
- `speaker-metadata.csv` |
|
- contains the speaker IDs paired with their speech properties |
|
- `voice-metadata.csv` |
|
- contains speaker-TTS name pairs |
|
|
|
Finally, the `data` root contains a single metadata file, `prompts.csv`, which as the name would suggest, contains the prompt transcripts. The only samples for which there are no transcripts are the ARCTIC-C ones, for which we couldn't find a source in the internet. |
|
|
|
|
|
### Sample |
|
|
|
A sample contains contains an Audio feature `audio`, and a string `label`. |
|
|
|
``` |
|
{ |
|
'audio': { |
|
'path': 'ahw/human/arctic_a0001.wav', |
|
'array': array([0., 0., 0., ..., 0., 0., 0.]), |
|
'sampling_rate': 16000 |
|
}, |
|
'label': 'human' |
|
} |
|
``` |
|
|
|
## Citation |
|
|
|
The dataset is featured alongside our paper, **Synthetic speech detection with Wav2Vec 2.0 in various language settings**, which will be published on IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). We'll provide links once it's available online. |
|
|
|
**BibTeX:** |
|
|
|
Note, the following BibTeX is incomplete - we'll update it once the actual one is known. |
|
|
|
``` |
|
@inproceedings{dropuljic-ssdww2v2ivls |
|
author={Dropuljić, Branimir and Šuflaj, Miljenko and Jertec, Andrej and Obadić, Leo} |
|
booktitle={2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)} |
|
title={Synthetic speech detection with Wav2Vec 2.0 in various language settings} |
|
year={2024} |
|
volume={} |
|
number={} |
|
pages={1-5} |
|
keywords={Synthetic speech detection;text-to-speech;wav2vec 2.0;spoofing attack;multilingualism} |
|
doi={} |
|
} |
|
``` |
|
|
|
## Dataset Card Authors |
|
|
|
- [Miljenko Šuflaj](https://huggingface.co/suflaj) |
|
|
|
## Dataset Card Contact |
|
|
|
- [Miljenko Šuflaj](mailto:msuflaj@realnetworks.com) |