Anja Reusch
added README.md
3ff6c8a
|
raw
history blame
1.07 kB
metadata
language:
  - en
tags:
  - mathematics
  - math-aware
datasets:
  - MathematicalStackExchange

Math-aware RoBERTa

This repository contains our pre-trained RoBERTa-based model for ARQMath 3. It was initialised from RoBERTa-base and further pre-trained on Math StackExchange in only one stage. We also added more LaTeX tokens to the tokenizer to enable a better tokenization of mathematical formulas. This model is not yet fine-tuned on a specific task.

For further details, please read our paper: http://ceur-ws.org/Vol-3180/paper-07.pdf.

Usage

You can use this model to further fine-tune it on any math-aware task you have in mind, e.g., classification, question-answering, etc. . Please note, that the model in this repository is only pre-trained and not fine-tuned.

Citation

If you find this model useful, consider citing our paper:

@article{reusch2022transformer,
  title={Transformer-Encoder and Decoder Models for Questions on Math},
  author={Reusch, Anja and Thiele, Maik and Lehner, Wolfgang},
  year={2022},
  organization={CLEF}
}