Iker's picture
Update README.md
63e3973 verified
|
raw
history blame
4.45 kB
metadata
language:
  - es
license: cc-by-nc-sa-4.0
multilinguality:
  - monolingual
size_categories:
  - n<1K
source_datasets:
  - original
task_categories:
  - summarization
pretty_name: NoticIA Human Validation
dataset_info:
  features:
    - name: web_url
      dtype: string
    - name: web_headline
      dtype: string
    - name: summary
      dtype: string
    - name: summary2
      dtype: string
    - name: web_text
      dtype: string
  splits:
    - name: test
      num_examples: 100
configs:
  - config_name: default
    data_files:
      - split: test
        path: test.jsonl
tags:
  - summarization
  - clickbait
  - news

"A Clickbait Article Summarization Dataset in Spanish."

This repository contains the manual annotations from a second human to validate the test set of the NoticIA dataset.

The full NoticIA dataset is available here: https://huggingface.co/datasets/Iker/NoticIA

Data explanation

  • web_url (int): The URL of the news article
  • web_headline (str): The headline of the article, which is a Clickbait.
  • summary (str): The original summary in the NoticIA dataset.
  • summary2 (str): The second summary written by another human to validate the quality of summary
  • web_text (int): The body of the article.

Dataset Description

Dataset Usage

# pip install datasets evaluate rouge-score
from datasets import load_dataset
from evaluate import load

dataset = load_dataset("Iker/NoticIA_Human_Validation",split="test")
rouge = load("rouge")
results = rouge.compute(
                predictions=[x["summary2"] for x in dataset],
                references=[[x["summary"]] for x in dataset],
                use_aggregator=True,
            )
print(results)

Uses

This dataset is intended to build models tailored for academic research that can extract information from large texts. The objective is to research whether current LLMs, given a question formulated as a Clickbait headline, can locate the answer within the article body and summarize the information in a few words. The dataset also aims to serve as a task to evaluate the performance of current LLMs in Spanish.

Out-of-Scope Use

You cannot use this dataset to develop systems that directly harm the newspapers included in the dataset. This includes using the dataset to train profit-oriented LLMs capable of generating articles from a short text or headline, as well as developing profit-oriented bots that automatically summarize articles without the permission of the article's owner. Additionally, you are not permitted to train a system with this dataset that generates clickbait headlines.

This dataset contains text and headlines from newspapers; therefore, you cannot use it for commercial purposes unless you have the license for the data.

Dataset Creation

The dataset has been meticulously created by hand. We utilize two sources to compile Clickbait articles:

  • The Twitter user @ahorrandoclick1, who reposts Clickbait articles along with a hand-crafted summary. Although we use their summaries as a reference, most of them have been rewritten (750 examples from this source).
  • The web demo ⚔️ClickbaitFighter⚔️, which operates a pre-trained model using an early iteration of our dataset. We collect all the model inputs/outputs and manually correct them (100 examples from this source).

Who are the annotators?

The dataset was originally by Iker García-Ferrero and has been validated by Begoña Altura. The annotation took ~40 hours.

Citation

Paper coming soon, for now, you can use this citation:

@misc{garcia-ferrero-etal-2023-noticia,
    title = "TNoticIA: A Clickbait Article Summarization Dataset in Spanish.",
    author = "Garc{\'\i}a-Ferrero, Iker  and
      Altuna, Bego{\~n}a  and
    year = "2024",
    url = "https://github.com/ikergarcia1996/NoticIA",
}