|
--- |
|
license: mit |
|
datasets: |
|
- bsmock/pubtables-1m |
|
tags: |
|
- table structure recognition |
|
- table extraction |
|
--- |
|
|
|
# Model Card for TATR-v1.1-Pub |
|
|
|
This repo contains the model weights for TATR (Table Transformer) v1.1 trained on the PubTables-1M dataset, using the training details in the paper: ["Aligning benchmark datasets for table structure recognition"](https://arxiv.org/abs/2303.00716). |
|
|
|
These model weights are intended to be used with [the Microsoft implementation of Table Transformer (TATR)](https://github.com/microsoft/table-transformer). |
|
|
|
This model (v1.1) was trained with additional image cropping compared to [v1.0](https://huggingface.co./bsmock/tatr-pubtables1m-v1.0) and works best on tightly cropped table images (5 pixels or less). |
|
It was also trained for more epochs, and as a result it outperforms the original model on PubTables-1M. |
|
|
|
Evaluation metrics in the paper were computed with the PubTables-1M v1.1 dataset, which tightly crops the table images in the test and validation splits. |
|
Table images in PubTables-1M v1.0, on the other hand, have ~30 pixels of padding in all three splits (train, test, and val). |
|
|
|
Model weights that can be loaded into the Hugging Face implementation of TATR are coming soon. |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
- **Developed by:** Brandon Smock and Rohith Pesala, while at Microsoft |
|
- **License:** MIT |
|
- **Finetuned from model:** DETR ResNet-18 |
|
|
|
### Model Sources |
|
|
|
Please see the following for more details: |
|
|
|
- **Repository:** ["https://github.com/microsoft/table-transformer"](https://github.com/microsoft/table-transformer) |
|
- **Paper:** ["Aligning benchmark datasets for table structure recognition"](https://arxiv.org/abs/2303.00716) |