Datasets:
File size: 6,771 Bytes
d5ffdd6 a921e3e d5ffdd6 e7378af d5ffdd6 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 |
---
license: cc-by-4.0
task_categories:
- audio-classification
language:
- en
tags:
- speech
- speech-classifiation
- text-to-speech
- spoofing
- accents
pretty_name: ARCTIC-HS
size_categories:
- 10K<n<100K
---
# ARCTIC-HS
An extension of the [CMU_ARCTIC](http://festvox.org/cmu_arctic/) and [L2-ARCTIC](https://psi.engr.tamu.edu/l2-arctic-corpus/) datasets for synthetic speech detection using text-to-speech, featured in the paper **Synthetic speech detection with Wav2Vec 2.0 in various language settings**.
This dataset is 1 of 3 used in the paper, the others being:
- [FLEURS-HS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs)
- the default train, dev and test sets
- [FLEURS-HS VITS](https://huggingface.co/datasets/realnetworks-kontxt/fleurs-hs-vits)
- test set containing (generally) more difficult synthetic samples
- separated due to different licensing
## Dataset Details
### Dataset Description
The dataset features 3 parts obtained from the 2 original datasets:
- CMU non-native US English speakers
- CMU native US English speakers
- L2 non-native English speakers
The original ARCTIC samples are used as `human` samples, while `synthetic` samples are generated using [Google Cloud Text-To-Speech](https://cloud.google.com/text-to-speech).
The resulting `symmetric` datasets features exactly twice the samples of the original ones, but we also provide:
- human samples that couldn't be paired
- 4 speakers in entirety we couldn't pair with a TTS voice
- a small amount of utterances unrelated to the A and B ARCTIC samples
- synthetic samples that couldn't be paired
- mostly when a human speaker didn't read the B ARCTIC samples
- **Curated by:** [KONTXT by RealNetworks](https://realnetworks.com/kontxt)
- **Funded by:** [RealNetworks](https://realnetworks.com/)
- **Language(s) (NLP):** English
- **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0) for the code, [CC BY 4.0](https://creativecommons.org/licenses/by/4.0/) for the dataset, however:
- the human part of the dataset is under a **custom CMU license**
- it should be compatible with **CC BY 4.0**
- the human part of the L2 dataset is under **CC BY-NC 4.0**
### Dataset Sources
The original ARCTIC sets were downloaded from their original sources.
- **CMU_ARCTIC Repository:** [festvox.org](http://festvox.org/cmu_arctic/)
- **L2-ARCTIC Repository:** [tamu.edu](https://psi.engr.tamu.edu/l2-arctic-corpus/)
- **CMU_ARCTIC Paper:** [cmu.edu](https://www.cs.cmu.edu/~awb/papers/ssw5/arctic.pdf)
- **L2-ARCTIC Paper:** [tamu.edu](https://psi.engr.tamu.edu/wp-content/uploads/2018/08/zhao2018interspeech.pdf)
- **Paper:** Synthetic speech detection with Wav2Vec 2.0 in various language settings
## Uses
This dataset is best used as a test set for accents. Each sample contains an `Audio` feature, and a label: `human` or `synthetic`.
### Direct Use
The following snippet of code demonstrates loading the CMU non-native US English speaker part of the dataset:
```python
from datasets import load_dataset
arctic_hs = load_dataset(
"realnetworks-kontxt/arctic-hs",
"cmu_non-us",
split="test",
trust_remote_code=True,
)
```
To load a different part, change `cmu_non-us` into one of the following:
- `cmu_us` for CMU native US English speakers
- `l2` for L2 non-native speakers
This dataset only has a `test` split.
To load only the paired samples, append `_symmetric` to the name. For example, `cmu_non-us` will load the test set also containing human and synthetic samples without their counterpart, while `cmu_non-us_symmetric` will only load samples where there is both a human and synthetic variant. This is useful if you want to have perfectly balanced labels within speakers, and if you wish to exclude speakers for which there are no TTS counterparts at all.
The `trust_remote_code=True` parameter is necessary because this dataset uses a custom loader. To check out which code is being ran, check out the [loading script](./arctic-hs.py).
## Dataset Structure
The dataset data is contained in the [data directory](https://huggingface.co/datasets/realnetworks-kontxt/arctic-hs/tree/main/data).
There exists 1 directory per part.
Within those directories, there are 2 further directories:
- `splits`
- `pairs`
Within the `splits` folder, there is 1 file per split:
- `test.tar.gz`
Those `.tar.gz` files contain 2 directories:
- `human`
- `synthetic`
Each of these directories contain `.wav` files. Keep in mind that these directories can't be merged as they share most of their file names. An identical file name implies a speaker-voice pair, ex. `human/arctic_a0001.wav` and `synthetic/arctic_a0001.wav`.
The `pairs` folder contains a list of file names within each speaker, and whether or not there is a human-synthetic pair. Based on that metadata we determine which samples appear in `symmetric` datasets.
Back to the part directories, each contain 2 metadata files, which are not used in the loaded dataset, but might be useful to researchers:
- `speaker-metadata.csv`
- contains the speaker IDs paired with their speech properties
- `voice-metadata.csv`
- contains speaker-TTS name pairs
Finally, the `data` root contains a single metadata file, `prompts.csv`, which as the name would suggest, contains the prompt transcripts. The only samples for which there are no transcripts are the ARCTIC-C ones, for which we couldn't find a source in the internet.
### Sample
A sample contains contains an Audio feature `audio`, and a string `label`.
```
{
'audio': {
'path': 'ahw/human/arctic_a0001.wav',
'array': array([0., 0., 0., ..., 0., 0., 0.]),
'sampling_rate': 16000
},
'label': 'human'
}
```
## Citation
The dataset is featured alongside our paper, **Synthetic speech detection with Wav2Vec 2.0 in various language settings**, which will be published on IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW). We'll provide links once it's available online.
**BibTeX:**
Note, the following BibTeX is incomplete - we'll update it once the actual one is known.
```
@inproceedings{dropuljic-ssdww2v2ivls
author={Dropuljić, Branimir and Šuflaj, Miljenko and Jertec, Andrej and Obadić, Leo}
booktitle={2024 IEEE International Conference on Acoustics, Speech, and Signal Processing Workshops (ICASSPW)}
title={Synthetic speech detection with Wav2Vec 2.0 in various language settings}
year={2024}
volume={}
number={}
pages={1-5}
keywords={Synthetic speech detection;text-to-speech;wav2vec 2.0;spoofing attack;multilingualism}
doi={}
}
```
## Dataset Card Authors
- [Miljenko Šuflaj](https://huggingface.co/suflaj)
## Dataset Card Contact
- [Miljenko Šuflaj](mailto:msuflaj@realnetworks.com) |