AmbatronBERTa / README.md
Peerawat2024's picture
Update README.md
aea6cc2 verified
metadata
license: unknown
base_model:
  - airesearch/wangchanberta-base-att-spm-uncased

AmbatronBERTa

AmbatronBERTa is a Thai language model fine-tuned specifically for text classification tasks, built upon the WangchanBERTa architecture.

Model Description

AmbatronBERTa is designed to handle the complexities of the Thai language. It has been fine-tuned on a dataset of over 3,000 research papers to improve classification accuracy. Leveraging the transformer-based WangchanBERTa, it efficiently captures the nuances of Thai text, making it suitable for classifying documents across multiple fields.

Developers

AmbatronBERTa was developed by students at King Mongkut's University of Technology North Bangkok:

  • Peerawat Banpahan
  • Waris Thongpho

Use Cases

AmbatronBERTa can be applied to a wide range of tasks, such as:

  • Research Classification: Categorizing academic papers into relevant topics.
  • Document Organization: Classifying articles, blogs, and other documents by themes.
  • Sentiment Analysis: Analyzing sentiment in Thai-language texts across various contexts.

How to Use

To use AmbatronBERTa with the transformers library:

from transformers import AutoTokenizer, AutoModelForSequenceClassification

# Load the tokenizer and model
tokenizer = AutoTokenizer.from_pretrained("Peerawat2024/AmbatronBERTa")
model = AutoModelForSequenceClassification.from_pretrained("Peerawat2024/AmbatronBERTa")