pszemraj commited on
Commit
d97a777
1 Parent(s): f42b69e

add paper link

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -84,6 +84,8 @@ A fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/goo
84
  - 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
85
  - all training used 16384 token input / 1024 max output
86
 
 
 
87
  ## How-To in Python
88
 
89
  Install/update transformers `pip install -U transformers`
 
84
  - 20+ epochs of fine-tuning from the base model on V100/A100 GPUs
85
  - all training used 16384 token input / 1024 max output
86
 
87
+ Read the paper by Guo et al. here: [LongT5: Efficient Text-To-Text Transformer for Long Sequences](https://arxiv.org/pdf/2112.07916.pdf)
88
+
89
  ## How-To in Python
90
 
91
  Install/update transformers `pip install -U transformers`