MedTrChat / app.py
harishnair04's picture
fix: tf rebuild
b56bfa5
import gradio as gr
from huggingface_hub import InferenceClient
import keras
import keras_nlp
import os
os.environ["KERAS_BACKEND"] = "jax"
"""
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co./docs/huggingface_hub/v0.22.2/en/guides/inference
"""
css = """
html, body {
margin: 0;
padding: 0;
height: 100%;
overflow: hidden;
}
body::before {
content: '';
position: fixed;
top: 0;
left: 0;
width: 100vw;
height: 100vh;
background-image: url('https://png.pngtree.com/background/20230413/original/pngtree-medical-color-cartoon-blank-background-picture-image_2422159.jpg');
background-size: cover;
background-repeat: no-repeat;
opacity: 0.60;
background-position: center;
z-index: -1;
}
.gradio-container {
display: flex;
flex-direction: column;
justify-content: center;
align-items: center;
height: 100vh;
}
"""
gemma_model = keras_nlp.models.GemmaCausalLM.from_preset("hf://harishnair04/gemma_instruct_medtr_2b")
def respond(input):
template = "Instruction:\n{instruction}\n\nResponse:\n{response}"
prompt = template.format(
instruction=input,
response="",
)
out = gemma_model.generate(prompt, max_length=1024)
ind = out.index('Response') + len('Response')+2
return out[ind:]
chat_interface = gr.Interface(
respond,
inputs="text",
outputs="text",
title="Gemma instruct 2b_en finetuned on medical transcripts",
description="Gemma instruct 2b_en finetuned on medical transcripts",
css=css
)
chat_interface.launch()