add paper link
Browse files
README.md
CHANGED
@@ -84,6 +84,8 @@ A fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/goo
|
|
84 |
- 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
|
85 |
- all training used 16384 token input / 1024 max output
|
86 |
|
|
|
|
|
87 |
## How-To in Python
|
88 |
|
89 |
Install/update transformers `pip install -U transformers`
|
|
|
84 |
- 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
|
85 |
- all training used 16384 token input / 1024 max output
|
86 |
|
87 |
+
Read the paper by Guo et al. here: [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/pdf/2112.07916.pdf)
|
88 |
+
|
89 |
## How-To in Python
|
90 |
|
91 |
Install/update transformers `pip install -U transformers`
|