Text Generation
PEFT
Safetensors
French
English
bloom
File size: 3,296 Bytes
922bdc3
 
 
 
 
 
 
 
 
0381c7e
922bdc3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
f39fe6b
922bdc3
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
---
base_model: cmarkea/bloomz-3b-dpo-chat  
library_name: peft  
license: apache-2.0  
datasets:  
- cmarkea/table-vqa  
language:  
- fr  
- en  
pipeline_tag: text-generation
---

## Model Description

**cmarkea/bloomz-3b-dpo-table-qa-latex** is a fine-tuned version of the **[cmarkea/bloomz-3b-dpo-chat](https://huggingface.co./cmarkea/bloomz-3b-dpo-chat)** model, specialized for table-based question answering (QA) tasks. The model has been trained on the **[table-vqa](https://huggingface.co./datasets/cmarkea/table-vqa)** dataset, which was developed by Crédit Mutuel Arkéa, and it processes tables provided in their LaTeX source format. 

This model is optimized for multilingual environments, supporting both French and English, and is especially effective in extracting and interpreting tabular data from documents. It has been fine-tuned for 2 days on an A100 40GB GPU and operates in bfloat16 precision to maximize resource efficiency.

### Key Features
- **Domain:** Table-based question answering (QA), particularly for extracting information from LaTeX-format tables.
- **Language Support:** French and English, making it suitable for multilingual environments.
- **Model Type:** Text-to-text language model.
- **Precision:** bfloat16, optimizing computational efficiency.
- **Training Duration:** 2 days on A100 40GB GPU.
- **Fine-Tuning Method:** Full fine-tuning.

This model is highly applicable in fields where tabular data needs to be queried and analyzed, such as financial reports, academic papers, and technical documentation.

## Usage

Here’s an example of how to use this model for table-based question answering:

```python
import torch
from transformers import pipeline

device = 0 if torch.cuda.is_available() else -1
table = '''\begin{tabular}{|c|c|c|}
        \hline
         Model & MAE-{$TKE$}-low &  MAE-{$TKE$}-high\\
         \hline
         U-FNET & 0.0048 & $1.09 \times 10^{-5}$\\
         \hline
    \end{tabular}'''
question = "What is the MAE-TKE-high value for the U-FNET model?"
prompt = table + '\n' + question

model = pipeline("text-generation", "cmarkea/bloomz-3b-dpo-chat", device=device)
result = model(f"</s>{prompt}<s>", max_new_tokens=512)
print(result)
```

The model processes tables written in LaTeX format, so be sure to provide your tables in that form.

## Performance

This model was evaluated on 200 question-answer pairs extracted from 100 tables in the **[table-vqa](https://huggingface.co./datasets/cmarkea/table-vqa)** test set. Each table had two question-answer pairs: one in French and one in English.

The evaluation used the **[LLM-as-Juries](https://arxiv.org/abs/2404.18796)** method, employing three judge models (GPT-4o, Gemini1.5 Pro, and Claude 3.5-Sonnet). The scoring was adapted to the table QA context, with a scale from 0 to 5 to ensure precision in assessing the model’s performance.

Here’s a visualization of the results:

![constellation](https://cdn-uploads.huggingface.co/production/uploads/6568aae588bfbc261a5b6548/J3c-phqVkZ4s_nlua7w5k.png)


## Citation

```bibtex
@online{AgDePaligemmaTableQALatex,
  AUTHOR = {Tom Agonnoude, Cyrile Delestre},
  URL = {https://huggingface.co./cmarkea/bloomz-3b-dpo-table-qa-latex},
  YEAR = {2024},
  KEYWORDS = {Table understanding, LaTeX, Multilingual, QA},
}
```