|
# Linformer: Self-Attention with Linear Complexity (Wang et al., 2020) |
|
|
|
This example contains code to train Linformer models as described in our paper |
|
[Linformer: Self-Attention with Linear Complexity](https://arxiv.org/abs/2006.04768). |
|
|
|
## Training a new Linformer RoBERTa model |
|
|
|
You can mostly follow the [RoBERTa pretraining README](/examples/roberta/README.pretraining.md), |
|
updating your training command with `--user-dir examples/linformer/linformer_src --arch linformer_roberta_base`. |
|
|
|
## Citation |
|
|
|
If you use our work, please cite: |
|
|
|
```bibtex |
|
@article{wang2020linformer, |
|
title={Linformer: Self-Attention with Linear Complexity}, |
|
author={Wang, Sinong and Li, Belinda and Khabsa, Madian and Fang, Han and Ma, Hao}, |
|
journal={arXiv preprint arXiv:2006.04768}, |
|
year={2020} |
|
} |
|
``` |
|
|