RoBERTa base Fine-Tuned for Proposal Sentence Classification

Overview

  • Language: English
  • Model Name: oeg/BERT-Repository-Proposal

Description

This model is a fine-tuned bert base uncased model trained to classify sentences into two classes: proposal and non-proposal sentences. The training data includes sentences proposing a software or data repository. The model is trained to recognize and classify these sentences accurately.

How to use

To use this model in Python:

from transformers import RobertaForSequenceClassification, RobertaTokenizer
import torch

tokenizer = RobertaTokenizer.from_pretrained("bert-repo-proposal-tokenizer")
model = RobertaForSequenceClassification.from_pretrained("bert-repo-proposal-model")

sentence = "Your input sentence here."
inputs = tokenizer(sentence, return_tensors="pt")
outputs = model(**inputs)
probabilities = torch.nn.functional.softmax(outputs.logits, dim=1)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.