File size: 22,818 Bytes
b0c87a7 d392c84 b0c87a7 d392c84 b0c87a7 d392c84 70f140e 3e1d96c d41362d d392c84 70f140e 3e1d96c d41362d d392c84 70f140e d41362d d0e6006 d392c84 3714c05 239e8b2 d392c84 3714c05 d392c84 3714c05 ac8296c 3714c05 ac8296c d392c84 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 |
A dataset for benchmarking keyphrase extraction and generation techniques from abstracts of english scientific articles. For more details about the dataset please refer the original paper - [https://arxiv.org/abs/1704.02853](https://arxiv.org/abs/1704.02853)
Original source of the data - [https://scienceie.github.io/](https://scienceie.github.io/)
## Dataset Summary
The Semeval-2017 dataset was originally proposed by *Isabelle Augenstein et al.* in the paper titled - [SemEval 2017 Task 10: ScienceIE - Extracting Keyphrases and Relations from Scientific Publications](https://arxiv.org/abs/1704.02853) in the year 2017. The dataset consists of a abstracts of 500 English scientific papers from the ScienceDirect open access publications. The selected articles were evenly distributed among the domains of Computer Science, Material Sciences and Physics. Each paper has a set of keyphrases annotated by student volunteers. Each paper was double-annotated, where the second annotation was done by an expert annotator. In case of disagreement, the annotations done by expert annotators were chosen. The original dataset was divided into train, dev and test splits, evenly distributed across the three domains. The train, dev and test splits had 350, 50 and 100 articles respectively.
The dataset shared over here categorizes the keyphrases into *extractive* and *abstractive*. **Extractive keyphrases** are those that could be found in the input text and the **abstractive keyphrases** are those that are not present in the input text. In order to get all the meta-data about the documents and keyphrases please refer to the [original source](https://scienceie.github.io/) from which the dataset was taken from. The main motivation behind making this dataset available in the form as presented over here is to make it easy for the researchers to programmatically download it and evaluate their models for the tasks of keyphrase extraction and generation. As keyphrase extraction by treating it as a sequence tagging task and using contextual language models has become popular - [Keyphrase extraction from scholarly articles as sequence labeling using contextualized embeddings](https://arxiv.org/pdf/1910.08840.pdf), we have also made the token tags available in the BIO tagging format.
## Dataset Structure
Table 1: Statistics on the length of the abstractive keyphrases for Train, Test, and Validation splits of SemEval 2017 dataset.
| | Train | Test | Validation |
|:-----------------:|:------:|:------:|:----------:|
| Single word | 11.59% | 12.47% | 12.89% |
| Two words | 30.69% | 40.92% | 33.45% |
| Three words | 19.20% | 17.50% | 19.16% |
| Four words | 10.25% | 10.94% | 9.41% |
| Five words | 7.43% | 4.60% | 8.36% |
| Six words | 5.96% | 4.37% | 6.27% |
| Seven words | 4.28% | 2.40% | 3.14% |
| Eight words | 2.59% | 1.75% | 1.34% |
| Nine words | 2.19% | 1.75% | 1.74% |
| Ten words | 1.35% | 1.31% | 0.69% |
| Eleven words | 0.96% | 0.44% | 1.04% |
| Twelve words | 1.13% | 0.44% | 1.04% |
| Thirteen words | 0% | 0.44% | 0.34% |
| Fourteen words | 0.45% | 0.22% | 0.348% |
| Fifteen words | 0.39% | 0% | 0% |
| Sixteen words | 0.17% | 0% | 0% |
| Seventeen words | 0.11% | 0.22% | 0.34% |
| Eighteen words | 0.11% | 0% | 0% |
| Nineteen words | 0.11% | 0.22% | 0.34% |
| Twenty words | 0.06% | 0% | 0% |
| Twenty-two words | 0.06% | 0% | 0% |
| Twenty-five words | 0% | 0% | 0% |
Table 2: Statistics on the length of the extractive keyphrases for Train, Test, and Validation splits of SemEval 2017 dataset.
| | Train | Test | Validation |
|:-----------------:|:------:|:------:|:----------:|
| Single word | 27.94% | 34.50% | 36.56% |
| Two words | 33.04% | 39.64% | 31.72% |
| Three words | 17.85% | 13.45% | 15.50% |
| Four words | 8.75% | 6.19% | 7.11% |
| Five words | 4.72% | 2.44% | 4.27% |
| Six words | 2.24% | 0.89% | 1.85% |
| Seven words | 1.66% | 0.73% | 1.28% |
| Eight words | 1.33% | 0.48% | 0.43% |
| Nine words | 0.54% | 0.97% | 0.14% |
| Ten words | 0.21% | 0.24% | 0.57% |
| Eleven words | 0.38% | 0.081% | 0.28% |
| Twelve words | 0% | 0.16% | 0.14% |
| Thirteen words | 0.28% | 0% | 0% |
| Fourteen words | 0.21% | 0% | 0% |
| Fifteen words | 0.071% | 0% | 0% |
| Sixteen words | 0.02% | 0.081% | 0% |
| Eighteen words | 0% | 0.081% | 0.14 |
| Nineteen words | 0.02% | 0% | 0% |
| Twenty-five words | 0.04% | 0% | 0% |
Table 3: General statistics of the Semeval 2017 dataset.
| Type of Analysis | Train | Test | Validation |
|:------------------------------------------------:|:-------------------:|:-------------------:|:-------------------:|
| Annotator Type | Authors and Readers | Authors and Readers | Authors and Readers |
| Document Type | Scientific Papers | Scientific Papers | Scientific Papers |
| No. of Documents | 350 | 100 | 50 |
| Avg. Document length (words) | 160.5 | 190.4 | 380.8 |
| Max Document length (words) | 355 | 297 | 355 |
| Max no. of abstractive keyphrases in a document | 23 | 13 | 22 |
| Min no. of abstractive keyphrases in a document | 0 | 0 | 0 |
| Avg. no. of abstractive keyphrases per document | 5.07 | 4.57 | 5.74 |
| Max no. of extractive keyphrases in a document | 29 | 27 | 30 |
| Min no. of extractive keyphrases in a document | 2 | 4 | 2 |
| Avg. no. of extractive keyphrases per document | 11.9 | 12.26 | 14.06 |
Train
- Percentage of keyphrases that are named entities: 50.09% (named entities detected using scispacy - en-core-sci-lg model)
- Percentage of keyphrases that are noun phrases: 57.65% (noun phrases detected using spacy en-core-web-lg after removing determiners)
Validation
- Percentage of keyphrases that are named entities: 60.02% (named entities detected using scispacy - en-core-sci-lg model)
- Percentage of keyphrases that are noun phrases: 62.87% (noun phrases detected using spacy en-core-web-lg after removing determiners)
Test
- Percentage of keyphrases that are named entities: 59.78% (named entities detected using scispacy - en-core-sci-lg model)
- Percentage of keyphrases that are noun phrases: 66.39% (noun phrases detected using spacy en-core-web-lg after removing determiners)
### Data Fields
- **id**: unique identifier of the document.
- **document**: Whitespace separated list of words in the document.
- **doc_bio_tags**: BIO tags for each word in the document. B stands for the beginning of a keyphrase and I stands for inside the keyphrase. O stands for outside the keyphrase and represents the word that isn't a part of the keyphrase at all.
- **extractive_keyphrases**: List of all the present keyphrases.
- **abstractive_keyphrase**: List of all the absent keyphrases.
### Data Splits
|Split| #datapoints |
|--|--|
| Train | 350 |
| Test | 100 |
| Validation | 50 |
## Usage
### Full Dataset
```python
from datasets import load_dataset
# get entire dataset
dataset = load_dataset("midas/semeval2017", "raw")
# sample from the train split
print("Sample from train dataset split")
test_sample = dataset["train"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Document BIO Tags: ", test_sample["doc_bio_tags"])
print("Extractive/present Keyphrases: ", test_sample["extractive_keyphrases"])
print("Abstractive/absent Keyphrases: ", test_sample["abstractive_keyphrases"])
print("\n-----------\n")
# sample from the validation split
print("Sample from validation dataset split")
validation_sample = dataset["validation"][0]
print("Fields in the sample: ", [key for key in validation_sample.keys()])
print("Tokenized Document: ", validation_sample["document"])
print("Document BIO Tags: ", validation_sample["doc_bio_tags"])
print("Extractive/present Keyphrases: ", validation_sample["extractive_keyphrases"])
print("Abstractive/absent Keyphrases: ", validation_sample["abstractive_keyphrases"])
print("\n-----------\n")
# sample from the test split
print("Sample from test dataset split")
test_sample = dataset["test"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Document BIO Tags: ", test_sample["doc_bio_tags"])
print("Extractive/present Keyphrases: ", test_sample["extractive_keyphrases"])
print("Abstractive/absent Keyphrases: ", test_sample["abstractive_keyphrases"])
print("\n-----------\n")
```
**Output**
```bash
Sample from train dataset split
Fields in the sample: ['id', 'document', 'doc_bio_tags', 'extractive_keyphrases', 'abstractive_keyphrases', 'other_metadata']
Tokenized Document: ['It', 'is', 'well', 'known', 'that', 'one', 'of', 'the', 'long', 'standing', 'problems', 'in', 'physics', 'is', 'understanding', 'the', 'confinement', 'physics', 'from', 'first', 'principles.', 'Hence', 'the', 'challenge', 'is', 'to', 'develop', 'analytical', 'approaches', 'which', 'provide', 'valuable', 'insight', 'and', 'theoretical', 'guidance.', 'According', 'to', 'this', 'viewpoint,', 'an', 'effective', 'theory', 'in', 'which', 'confining', 'potentials', 'are', 'obtained', 'as', 'a', 'consequence', 'of', 'spontaneous', 'symmetry', 'breaking', 'of', 'scale', 'invariance', 'has', 'been', 'developed', '[1].', 'In', 'particular,', 'it', 'was', 'shown', 'that', 'a', 'such', 'theory', 'relies', 'on', 'a', 'scale-invariant', 'Lagrangian', 'of', 'the', 'type', '[2]', '(1)L=14w2−12w−FμνaFaμν,', 'where', 'Fμνa=∂μAνa−∂νAμa+gfabcAμbAνc,', 'and', 'w', 'is', 'not', 'a', 'fundamental', 'field', 'but', 'rather', 'is', 'a', 'function', 'of', '4-index', 'field', 'strength,', 'that', 'is,', '(2)w=εμναβ∂μAναβ.', 'The', 'Aναβ', 'equation', 'of', 'motion', 'leads', 'to', '(3)εμναβ∂βw−−FγδaFaγδ=0,', 'which', 'is', 'then', 'integrated', 'to', '(4)w=−FμνaFaμν+M.', 'It', 'is', 'easy', 'to', 'verify', 'that', 'the', 'Aaμ', 'equation', 'of', 'motion', 'leads', 'us', 'to', '(5)∇μFaμν+MFaμν−FαβbFbαβ=0.', 'It', 'is', 'worth', 'stressing', 'at', 'this', 'stage', 'that', 'the', 'above', 'equation', 'can', 'be', 'obtained', 'from', 'the', 'effective', 'Lagrangian', '(6)Leff=−14FμνaFaμν+M2−FμνaFaμν.', 'Spherically', 'symmetric', 'solutions', 'of', 'Eq.', '(5)', 'display,', 'even', 'in', 'the', 'Abelian', 'case,', 'a', 'Coulomb', 'piece', 'and', 'a', 'confining', 'part.', 'Also,', 'the', 'quantum', 'theory', 'calculation', 'of', 'the', 'static', 'energy', 'between', 'two', 'charges', 'displays', 'the', 'same', 'behavior', '[1].', 'It', 'is', 'well', 'known', 'that', 'the', 'square', 'root', 'part', 'describes', 'string', 'like', 'solutions', '[3,4].']
Document BIO Tags: ['O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'B', 'I', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'O']
Extractive/present Keyphrases: ['aaμ equation of motion', 'aναβ equation of motion leads', 'confining part', 'coulomb piece', 'develop analytical approaches', 'quantum theory calculation of the static energy between two charges', 'spherically symmetric solutions', 'spontaneous symmetry breaking of scale invariance', 'string like solutions', 'the effective lagrangian', 'understanding the confinement physics from first principles']
Abstractive/absent Keyphrases: ['(2)w=εμναβ∂μaναβ', 'function of 4-index field strength', 'integrated to (4)w=−fμνafaμν+m', 'leff=−14fμνafaμν+m2−fμνafaμν', 'scale-invariant lagrangian', 'εμναβ∂βw−−fγδafaγδ=0']
-----------
Sample from validation dataset split
Fields in the sample: ['id', 'document', 'doc_bio_tags', 'extractive_keyphrases', 'abstractive_keyphrases', 'other_metadata']
Tokenized Document: ['In', 'the', 'current', 'CLSVOF', 'method,', 'the', 'normal', 'vector', 'is', 'calculated', 'directly', 'by', 'discretising', 'the', 'LS', 'gradient', 'using', 'a', 'finite', 'difference', 'scheme.', 'By', 'appropriately', 'choosing', 'one', 'of', 'three', 'finite', 'difference', 'schemes', '(central,', 'forward,', 'or', 'backward', 'differencing),', 'it', 'has', 'been', 'demonstrated', 'that', 'thin', 'liquid', 'ligaments', 'can', 'be', 'well', 'resolved', 'see', 'Xiao', '(2012).', 'Although', 'a', 'high', 'order', 'discretisation', 'scheme', '(e.g.', '5th', 'order', 'WENO)', 'has', 'been', 'found', 'necessary', 'for', 'LS', 'evolution', 'in', 'pure', 'LS', 'methods', 'to', 'reduce', 'mass', 'error,', 'low', 'order', 'LS', 'discretisation', 'schemes', '(2nd', 'order', 'is', 'used', 'here)', 'can', 'produce', 'accurate', 'results', 'when', 'the', 'LS', 'equation', 'is', 'solved', 'and', 'constrained', 'as', 'indicated', 'above', 'in', 'a', 'CLSVOF', 'method', '(see', 'Xiao,', '2012),', 'since', 'the', 'VOF', 'method', 'maintains', '2nd', 'order', 'accuracy.', 'This', 'is', 'a', 'further', 'reason', 'to', 'adopt', 'the', 'CLSVOF', 'method,', 'which', 'has', 'been', 'used', 'for', 'all', 'the', 'following', 'simulations', 'of', 'liquid', 'jet', 'primary', 'breakup.']
Document BIO Tags: ['O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'B', 'O', 'O', 'B', 'I', 'I', 'B', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O']
Extractive/present Keyphrases: ['5th order weno', 'clsvof method', 'finite difference scheme', 'finite difference schemes', 'high order discretisation scheme', 'liquid', 'low order ls discretisation schemes', 'ls', 'reduce mass error', 'vof method']
Abstractive/absent Keyphrases: ['central, forward, or backward differencing', 'ls methods', 'simulations of liquid jet primary breakup', 'thin liquid ligaments']
-----------
Sample from test dataset split
Fields in the sample: ['id', 'document', 'doc_bio_tags', 'extractive_keyphrases', 'abstractive_keyphrases', 'other_metadata']
Tokenized Document: ['Traditionally,', 'archaeologists', 'have', 'recorded', 'sites', 'and', 'artefacts', 'via', 'a', 'combination', 'of', 'ordinary', 'still', 'photographs,', '2D', 'line', 'drawings', 'and', 'occasional', 'cross-sections.', 'Given', 'these', 'constraints,', 'the', 'attractions', 'of', '3D', 'models', 'have', 'been', 'obvious', 'for', 'some', 'time,', 'with', 'digital', 'photogrammetry', 'and', 'laser', 'scanners', 'offering', 'two', 'well-known', 'methods', 'for', 'data', 'capture', 'at', 'close', 'range', '(e.g.', 'Bates', 'et', 'al.,', '2010;', 'Hess', 'and', 'Robson,', '2010).', 'The', 'highest', 'specification', 'laser', 'scanners', 'still', 'boast', 'better', 'positional', 'accuracy', 'and', 'greater', 'true', 'colour', 'fidelity', 'than', 'SfM–MVS', 'methods', '(James', 'and', 'Robson,', '2012),', 'but', 'the', 'latter', 'produce', 'very', 'good', 'quality', 'models', 'nonetheless', 'and', 'have', 'many', 'unique', 'selling', 'points.', 'Unlike', 'traditional', 'digital', 'photogrammetry,', 'little', 'or', 'no', 'prior', 'control', 'of', 'camera', 'position', 'is', 'necessary,', 'and', 'unlike', 'laser', 'scanning,', 'no', 'major', 'equipment', 'costs', 'or', 'setup', 'are', 'involved.', 'However,', 'the', 'key', 'attraction', 'of', 'SfM–MVS', 'is', 'that', 'the', 'required', 'input', 'can', 'be', 'taken', 'by', 'anyone', 'with', 'a', 'digital', 'camera', 'and', 'modest', 'prior', 'training', 'about', 'the', 'required', 'number', 'and', 'overlap', 'of', 'photographs.', 'A', 'whole', 'series', 'of', 'traditional', 'bottlenecks', 'are', 'thereby', 'removed', 'from', 'the', 'recording', 'process', 'and', 'large', 'numbers', 'of', 'archaeological', 'landscapes,', 'sites', 'or', 'artefacts', 'can', 'now', 'be', 'captured', 'rapidly,', 'in', 'the', 'field,', 'in', 'the', 'laboratory', 'or', 'in', 'the', 'museum.', 'Fig.', '2a–c', 'shows', 'examples', 'of', 'terracotta', 'warrior', 'models', 'for', 'which', 'the', 'level', 'of', 'surface', 'detail', 'is', 'considerable.']
Document BIO Tags: ['O', 'O', 'O', 'O', 'B', 'O', 'B', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'B', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'O', 'O', 'O', 'O', 'B', 'I', 'B', 'O', 'B', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'B', 'I', 'I', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O', 'O']
Extractive/present Keyphrases: ['2d line drawings', '3d models', 'archaeological landscapes', 'artefacts', 'control of camera position', 'data capture at close range', 'digital camera', 'digital photogrammetry', 'laser scanners', 'laser scanning', 'ordinary still photographs', 'prior training about the required number and overlap of photographs', 'recording process', 'sfm–mvs', 'sites', 'terracotta warrior models']
Abstractive/absent Keyphrases: ['occasional cross-sections', 'recorded sites and artefacts', 'sfm–mvs methods', 'traditional digital photogrammetry']
-----------
```
### Keyphrase Extraction
```python
from datasets import load_dataset
# get the dataset only for keyphrase extraction
dataset = load_dataset("midas/semeval2017", "extraction")
print("Samples for Keyphrase Extraction")
# sample from the train split
print("Sample from train data split")
test_sample = dataset["train"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Document BIO Tags: ", test_sample["doc_bio_tags"])
print("\n-----------\n")
# sample from the test split
print("Sample from test data split")
test_sample = dataset["test"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Document BIO Tags: ", test_sample["doc_bio_tags"])
print("\n-----------\n")
```
### Keyphrase Generation
```python
# get the dataset only for keyphrase generation
dataset = load_dataset("midas/semeval2017", "generation")
print("Samples for Keyphrase Generation")
# sample from the train split
print("Sample from train data split")
test_sample = dataset["train"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Extractive/present Keyphrases: ", test_sample["extractive_keyphrases"])
print("Abstractive/absent Keyphrases: ", test_sample["abstractive_keyphrases"])
print("\n-----------\n")
# sample from the test split
print("Sample from test data split")
test_sample = dataset["test"][0]
print("Fields in the sample: ", [key for key in test_sample.keys()])
print("Tokenized Document: ", test_sample["document"])
print("Extractive/present Keyphrases: ", test_sample["extractive_keyphrases"])
print("Abstractive/absent Keyphrases: ", test_sample["abstractive_keyphrases"])
print("\n-----------\n")
```
## Citation Information
```
@article{DBLP:journals/corr/AugensteinDRVM17,
author = {Isabelle Augenstein and
Mrinal Das and
Sebastian Riedel and
Lakshmi Vikraman and
Andrew McCallum},
title = {SemEval 2017 Task 10: ScienceIE - Extracting Keyphrases and Relations
from Scientific Publications},
journal = {CoRR},
volume = {abs/1704.02853},
year = {2017},
url = {http://arxiv.org/abs/1704.02853},
eprinttype = {arXiv},
eprint = {1704.02853},
timestamp = {Mon, 13 Aug 2018 16:46:36 +0200},
biburl = {https://dblp.org/rec/journals/corr/AugensteinDRVM17.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```
## Contributions
Thanks to [@debanjanbhucs](https://github.com/debanjanbhucs), [@dibyaaaaax](https://github.com/dibyaaaaax) and [@ad6398](https://github.com/ad6398) for adding this dataset
|