File size: 17,344 Bytes
ebed1c3 99d4479 ebed1c3 e5d8357 ebed1c3 335a647 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 e5d8357 ebed1c3 b0caa68 ebed1c3 e5d8357 ebed1c3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 |
---
language:
- es
size_categories:
- n<1K
task_categories:
- summarization
pretty_name: Resumen Noticias Clickbait
dataset_info:
features:
- name: id
dtype: int64
- name: titular
dtype: string
- name: respuesta
dtype: string
- name: pregunta
dtype: string
- name: texto
dtype: string
- name: idioma
dtype: string
- name: periodo
dtype: string
- name: tarea
dtype: string
- name: registro
dtype: string
- name: dominio
dtype: string
- name: país_origen
dtype: string
splits:
- name: train
num_bytes: 5440051
num_examples: 700
- name: validation
num_bytes: 462364
num_examples: 50
- name: test
num_bytes: 782440
num_examples: 100
download_size: 3417692
dataset_size: 6684855
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: validation
path: data/validation-*
- split: test
path: data/test-*
tags:
- summarization
- clickbait
- news
---
<p align="center">
<img src="https://huggingface.co/datasets/Iker/NoticIA/resolve/main/assets/logo.png" style="width: 50%;">
</p>
<h1 align="center">NoticIA: A Clickbait Article Summarization Dataset in Spanish.</h1>
We present NoticIA, a dataset consisting of 850 Spanish news articles featuring prominent clickbait headlines, each paired with high-quality, single-sentence generative summarizations written by humans.
- 📖 Dataset Card en Español: https://huggingface.co/datasets/somosnlp/NoticIA-it/blob/main/README_es.md
## Dataset Details
### Dataset Description
We define a clickbait article as one that seeks to attract the reader's attention through curiosity. For this purpose, the headline poses a question or an incomplete, sensationalist, exaggerated, or misleading statement. The answer to the question raised in the headline usually does not appear until the end of the article, preceded by a large amount of irrelevant content. The goal is for the user to enter the website through the headline and then scroll to the end of the article, viewing as much advertising as possible. Clickbait articles tend to be of low quality and provide no value to the reader beyond the initial curiosity. This phenomenon undermines public trust in news sources and negatively affects the advertising revenue of legitimate content creators, who could see their web traffic reduced.
We introduce NoticIA, a dataset consisting of 850 Spanish news articles with clickbait headlines, each paired with high-quality, single-sentence generative summaries written by humans. This task demands advanced skills in text comprehension and summarization, challenging the ability of models to infer and connect various pieces of information to satisfy the user's informational curiosity generated by the clickbait headline.
The project is inspired by the X/Twitter account [@ahorrandoclick1](https://x.com/ahorrandoclick1). [@ahorrandoclick1](https://x.com/ahorrandoclick1) has 300,000 followers, demonstrating the great value of summarizing clickbait news articles. However, creating these summaries manually is a labor-intensive task, and the number of clickbait news articles published greatly exceeds the number of summaries one person can perform. Therefore, there is a need for automatic summarization of clickbait news articles. Additionally, as mentioned earlier, this is an ideal task for analyzing the text comprehension capabilities of a language model in Spanish.
The following Figure illustrates examples of clickbait headlines from our dataset, together with the human-written summaries.
<p align="center">
<img src="https://raw.githubusercontent.com/ikergarcia1996/NoticIA/main/assets/examples.png" style="width: 100%;">
</p>
- **Curated by:** [Iker García-Ferrero](https://ikergarcia1996.github.io/Iker-Garcia-Ferrero/), [Begoña Altuna](https://www.linkedin.com/in/bego%C3%B1a-altuna-78014139)
- **Funded by:** SomosNLP, HuggingFace, Argilla, [HiTZ Zentroa](https://www.hitz.eus/)
- **Language(s) (NLP):** es-ES
- **License:** apache-2.0
- **Web Page:** [Github](https://github.com/ikergarcia1996/NoticIA)
### Dataset Sources
- **💻 Repository:** https://github.com/ikergarcia1996/NoticIA
- **📖 Paper:** [NoticIA: A Clickbait Article Summarization Dataset in Spanish](https://arxiv.org/abs/2404.07611)
- **🤖 Pre Trained Models** [https://huggingface.co/collections/Iker/noticia-and-clickbaitfighter-65fdb2f80c34d7c063d3e48e](https://huggingface.co/collections/Iker/noticia-and-clickbaitfighter-65fdb2f80c34d7c063d3e48e)
- **🔌 Demo:** https://huggingface.co/spaces/somosnlp/NoticIA-demo
- **▶️ Video presentation (Spanish):** https://youtu.be/xc60K_NzUgk?si=QMqk6OzQZfKP1EUS
- **🐱💻 Hackathon #Somos600M**: https://somosnlp.org/hackathon
## Uses
This dataset has been compiled for use in scientific research. Specifically, for use in the evaluation of language models in Spanish.
Commercial use of this dataset is subject to the licenses of each news and media outlet. If you want to make commercial use of the dataset you will need to have
the express permission of the media from which the news has been obtained.
### Direct Use
- 📈 Evaluation of Language Models in Spanish.
- 🤖 Instruction-Tuning of Spanish Language Models
- 📚 Develop new datasets on top of our data
- 🎓 Any other academic research purpose.
### Out-of-Scope Use
We expressly prohibit the use of these data for two use cases that we consider to be that may be harmful: The training of models that generate sensational headlines or clickbait, and the training of models that generate articles or news automatically.
## Dataset Structure
The dataset is ready to be used to evaluate language models. For this aim, we have developed a *prompt* that makes use of the news headline and text.
The prompt is as follows:
```python
def clickbait_prompt(
headline: str,
body: str,
) -> str:
"""
Generate the prompt for the model.
Args:
headline (`str`):
The headline of the article.
body (`str`):
The body of the article.
Returns:
`str`: The formatted prompt.
"""
return (
f"Ahora eres una Inteligencia Artificial experta en desmontar titulares sensacionalistas o clickbait. "
f"Tu tarea consiste en analizar noticias con titulares sensacionalistas y "
f"generar un resumen de una sola frase que revele la verdad detrás del titular.\n"
f"Este es el titular de la noticia: {headline}\n"
f"El titular plantea una pregunta o proporciona información incompleta. "
f"Debes buscar en el cuerpo de la noticia una frase que responda lo que se sugiere en el título. "
f"Responde siempre que puedas parafraseando el texto original. "
f"Usa siempre las mínimas palabras posibles. "
f"Recuerda responder siempre en Español.\n"
f"Este es el cuerpo de la noticia:\n"
f"{body}\n"
)
```
The expected output of the model is the summary. Below is an example of how to evaluate `gemma-2b` in our dataset:
```
from transformers import pipeline
from datasets import load_dataset
generator = pipeline(model="google/gemma-2b-it",device_map="auto")
dataset = load_dataset("somosnlp/NoticIA-it",split="test")
outputs = generator(dataset[0]["prompt"], return_full_text=False,max_length=4096)
print(outputs)
```
The dataset includes the following fields:
- **ID**: id of the example
- **Titular (headline)**: headline of the article
- **Respuesta (response)**: Summary written by a human being
- **Pregunta (question)**: Prompt ready to be used as input to a language model.
- **Texto (text)**: Text of the article, obtained from the HTML.
- **idioma (language)**: ISO code of the language. In the case of Spanish, it also includes the geographic variant ("Mexican Spanish" = es_mx, "Ecuadorian Spanish" = es_ec, ...).
- **Tarea (task)** Task of the example. Every example has the task `resumen` (`summary`)
- **Registro (Language Register)**: `coloquial`, `medio` o `culto` (`colloquial`, `medium` or `educated`)
- **Dominio (Domain)**: The domain (`prensa`, `press`) and the subdomain.
- **País de origen (Country of origin)**: Country of origin of the data.
*The Idioma (language), Registro (Language Register), Dominio (Domain) and País de origen (Country of origin) labels have been automatically generated using GPT3.5-Turbo.*
## Dataset Creation
### Curation Rationale
NoticIA offers an ideal scenario to test the ability of language models to understand Spanish texts. This task is complex, involving discerning the hidden question in a clickbait headline or identifying the information that the user is actually seeking. This challenge involves filtering large volumes of superfluous content to find and succinctly summarize the relevant information accurately.
In addition, by making our data and models public, we aim to exert pressure against the use of deceptive tactics by online news providers to increase advertising revenue,
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
#### Data Collection and Processing
We have compiled clickbait news using the timeline of the X/Twitter user [@ahorrandoclick1](https://x.com/ahorrandoclick1). To do this, we extracted the URLs of the news mentioned by the user. Additionally, we have added about 100 clickbait news articles chosen by us. The following image shows the source of the news in the dataset.
<p align="center">
<img src="https://raw.githubusercontent.com/ikergarcia1996/NoticIA/main/assets/noticia_dataset.png" style="width: 50%;">
</p>
We have classified each of the news articles based on the category to which they belong. As can be seen, our dataset includes a wide variety of categories.
<p align="center">
<img src="https://raw.githubusercontent.com/ikergarcia1996/NoticIA/main/assets/categories_distribution_spanish.png" style="width: 50%;">
</p>
#### Annotation process
Although [@ahorrandoclick1](https://x.com/ahorrandoclick1) provides summaries of clickbait news, these summaries do not follow any guidelines, and in many cases, their summaries do not refer to the text, but are rather of the style *"This is advertising"*, *"They still haven't realized that..."*. Therefore, we have manually generated the summaries for the 850 news articles. To do this, we have defined strict annotation guidelines, available at the following link: [https://huggingface.co/spaces/Iker/ClickbaitAnnotation/blob/main/guidelines.py](https://huggingface.co/spaces/Iker/ClickbaitAnnotation/blob/main/guidelines.py).
The dataset has been annotated by [Iker García-Ferrero](https://ikergarcia1996.github.io/Iker-Garcia-Ferrero/) and [Begoña Altuna](https://www.linkedin.com/in/bego%C3%B1a-altuna-78014139), and this process has taken approximately 40 hours.
### Dataset Statistics
We have divided the dataset into three splits, which facilitates the training of models. As can be seen in the following table, the summaries of the news are extremely concise.
They respond to the clickbait headline using the fewest words possible.
| | Train | Validation | Test | Total |
|--------------------|-------|------------|------|-------|
| Number of articles | 700 | 50 | 100 | 850 |
| Average number of words in headlines | 16 | 17 | 17 | 17 |
| Average number of words in news text | 544 | 663 | 549 | 552 |
| Average number of words in summaries | 12 | 11 | 11 | 12 |
[More Information Needed]
#### Who are the annotators?
- [Iker García-Ferrero](https://ikergarcia1996.github.io/Iker-Garcia-Ferrero/): PhD Student HiTZ, the Basque center for language technology
- [Begoña Altuna](https://www.linkedin.com/in/bego%C3%B1a-altuna-78014139): Postdoctoral research fellow at HiTZ, the Basque center for language technology
### Annotation Validation
To validate the dataset, the 100 summaries from the Test set were annotated by two annotators. This data is available here: https://huggingface.co/datasets/Iker/NoticIA_Human_Validation
The overall agreement between the annotators was high, as they provided exactly the same answer in 26% of the cases and provided responses that partially shared information in 48% of the cases (same response but with some variation in the words used).
This demonstrates that it was easy for humans to find the information referred to by the headline. We also identified a list of cases where the annotators provided different but equally valid responses, which accounts for 18% of the cases.
Lastly, we identified 8 cases of disagreement. In 3 cases, one of the annotators made an incorrect summary,
likely due to fatigue after annotating multiple examples. In the remaining 5 cases, the disagreement was due to contradictory information in the article and
different interpretations of this information. In these cases, determining the correct summary is subject to the reader's interpretation.
Regarding the evaluation of the guidelines, overall, they were not ambiguous, although the request to select the minimum number of words to generate a
valid summary is sometimes interpreted differently by the annotators: For example, the minimum length could be understood as focusing on the question in the headline or a minimum well-formed phrase.
# Massive Evaluation of Language Models
As is customary in summary tasks, we use the ROUGE scoring metric to automatically evaluate the summaries produced by models. Our main metric is ROUGE-1, which considers whole words as basic units. To calculate the ROUGE score, we lowercase both summaries and remove punctuation marks. In addition to the ROUGE score, we also consider the average length of the summaries. For our task, we aim for the summaries to be concise, an aspect that the ROUGE score does not evaluate. Therefore, when evaluating models, we consider both the ROUGE-1 score and the average length of the summaries. Our goal is to find a model that achieves the highest possible ROUGE score with the shortest possible summary length, balancing quality and brevity.
We have evaluated the best current instruction-following language models. We used the previously defined prompt. The prompt is converted into the specific chat template of each model.
The code to reproduce the results is available at the following link: [https://github.com/ikergarcia1996/NoticIA](https://github.com/ikergarcia1996/NoticIA)
<p align="center">
<img src="https://huggingface.co/datasets/somosnlp/Resumen_Noticias_Clickbait/resolve/main/Results_zero.png" style="width: 100%;">
</p>
## Bias, Risks, and Limitations
The dataset contains a small number of articles from Latin America; however, the vast majority of the articles are from Spanish news sources. Therefore, this dataset will evaluate the proficiency of language models in Spanish from Spain.
Although explicitly prohibited, a bad actor could use our data to train models that can generate clickbait articles automatically, contributing to polluting the internet with low-quality content. In any case, we consider the advantages of having a text comprehension dataset to evaluate language models in Spanish to be superior to the possible risks.
## License
We release our annotations under the Apache 2.0 license. However, commercial use of this dataset is subject to the licenses of each news and media outlet.
## Citation
If you use this dataset, please cite our paper: [NoticIA: A Clickbait Article Summarization Dataset in Spanish](https://arxiv.org/abs/2404.07611)
**BibTeX:**
```
@misc{garcíaferrero2024noticia,
title={NoticIA: A Clickbait Article Summarization Dataset in Spanish},
author={Iker García-Ferrero and Begoña Altuna},
year={2024},
eprint={2404.07611},
archivePrefix={arXiv},
primaryClass={cs.CL}
}
```
## More Information
This project was developed during the [Hackathon #Somos600M](https://somosnlp.org/hackathon) organized by SomosNLP. Demo endpoints were sponsored by HuggingFace.
**Team:**
- [Iker García-Ferrero](https://huggingface.co/Iker)
- [Begoña Altura](https://huggingface.co/baltuna)
**Contact**: {iker.garciaf,begona.altuna}@ehu.eus
This dataset was created by [Iker García-Ferrero](https://ikergarcia1996.github.io/Iker-Garcia-Ferrero/) and [Begoña Altuna](https://www.linkedin.com/in/bego%C3%B1a-altuna-78014139). We are researchers in NLP at the University of the Basque Country, within the [IXA](https://www.ixa.eus/) research group, and we are part of [HiTZ, the Basque Language Technology Center](https://www.hitz.eus/es).
<div style="display: flex; justify-content: space-around; width: 100%;">
<div style="width: 50%;" align="left">
<a href="http://ixa.si.ehu.es/">
<img src="https://raw.githubusercontent.com/ikergarcia1996/Iker-Garcia-Ferrero/master/icons/ixa.png" width="50" height="50" alt="Ixa NLP Group">
</a>
</div>
<div style="width: 50%;" align="right">
<a href="http://www.hitz.eus/">
<img src="https://raw.githubusercontent.com/ikergarcia1996/Iker-Garcia-Ferrero/master/icons/Hitz.png" width="300" height="50" alt="HiTZ Basque Center for Language Technologies">
</a>
</div>
</div>
|