--- | |
language: en | |
license: apache-2.0 | |
--- | |
HF-version model for PRIMERA: Pyramid-based Masked Sentence Pre-training for Multi-document Summarization (ACL 2022). | |
The original code can be found [here](https://github.com/allenai/PRIMER). You can find the script and notebook to train/evaluate the model in the original github repo. | |
* Note: due to the difference between the implementations of the original Longformer and the Huggingface LED model, the results of converted models are slightly different. We run a sanity check on both fine-tuned and non fine-tuned models on the **Multinews dataset**, and show the results below: | |
| Model | Rouge-1 | Rouge-2 | Rouge-L | | |
| --- | ----------- |----------- |----------- | | |
| PRIMERA | 42.0 | 13.6 | 20.8| | |
| PRIMERA-hf | 41.7 |13.6 | 20.5| | |
| PRIMERA(finetuned) | 49.9 | 21.1 | 25.9| | |
| PRIMERA-hf(finetuned) | 49.9 | 20.9 | 25.8| | |
You can use it by | |
``` | |
from transformers import ( | |
AutoTokenizer, | |
LEDConfig, | |
LEDForConditionalGeneration, | |
) | |
tokenizer = AutoTokenizer.from_pretrained('allenai/PRIMERA') | |
config=LEDConfig.from_pretrained('allenai/PRIMERA') | |
model = LEDForConditionalGeneration.from_pretrained('allenai/PRIMERA') | |
``` |