NLP_app / pages /ToxicCommentDetector.py
Awlly's picture
adjusted visuals of the app
2f2be53
raw
history blame contribute delete
601 Bytes
import streamlit as st
from models.toxicity_MODEL import text2toxicity
st.title('Toxicity Detection')
st.write('This tool classifies text as toxic or non-toxic using RuBERT.')
user_input = st.text_area("Enter text to classify", "Type your text here...")
if st.button('Classify'):
toxicity_score = text2toxicity(user_input)
st.write('Toxicity score:', toxicity_score)
# Optional: Interpret the score for the user
if toxicity_score > 0.5:
st.write("This text is likely to be considered toxic.")
else:
st.write("This text is likely to be considered non-toxic.")