|
import streamlit as st |
|
from langchain.document_loaders import PyPDFLoader, DirectoryLoader |
|
from langchain import PromptTemplate |
|
from langchain.embeddings import HuggingFaceEmbeddings |
|
from langchain.vectorstores import FAISS |
|
from langchain.llms import CTransformers |
|
from langchain.chains import RetrievalQA |
|
|
|
DB_FAISS_PATH = 'vectorstores/db_faiss' |
|
|
|
custom_prompt_template = """Use the following pieces of information to answer the user's question. |
|
If you don't know the answer, just say that you don't know, don't try to make up an answer. |
|
Context: {context} |
|
Question: {question} |
|
Only return the helpful answer below and nothing else. |
|
Helpful answer: |
|
""" |
|
|
|
def set_custom_prompt(): |
|
prompt = PromptTemplate(template=custom_prompt_template, |
|
input_variables=['context', 'question']) |
|
return prompt |
|
|
|
def retrieval_qa_chain(llm, prompt, db): |
|
qa_chain = RetrievalQA.from_chain_type(llm=llm, |
|
chain_type='stuff', |
|
retriever=db.as_retriever(search_kwargs={'k': 2}), |
|
return_source_documents=True, |
|
chain_type_kwargs={'prompt': prompt} |
|
) |
|
return qa_chain |
|
|
|
def load_llm(): |
|
llm = CTransformers( |
|
model="TheBloke/Llama-2-7B-Chat-GGML", |
|
model_type="llama", |
|
max_new_tokens=512, |
|
temperature=0.5 |
|
) |
|
return llm |
|
|
|
def qa_bot(query): |
|
embeddings = HuggingFaceEmbeddings(model_name="sentence-transformers/all-MiniLM-L6-v2", |
|
model_kwargs={'device': 'cpu'}) |
|
db = FAISS.load_local(DB_FAISS_PATH, embeddings) |
|
llm = load_llm() |
|
qa_prompt = set_custom_prompt() |
|
qa = retrieval_qa_chain(llm, qa_prompt, db) |
|
|
|
|
|
response = qa({'query': query}) |
|
return response['result'] |
|
|
|
def add_vertical_space(spaces=1): |
|
for _ in range(spaces): |
|
st.markdown("---") |
|
|
|
def main(): |
|
st.set_page_config(page_title="Llama-2-GGML Medical Chatbot") |
|
|
|
with st.sidebar: |
|
st.title('Llama-2-GGML Medical Chatbot! 馃殌馃') |
|
st.markdown(''' |
|
## About |
|
|
|
The Llama-2-GGML Medical Chatbot uses the **Llama-2-7B-Chat-GGML** model and was trained on medical data from **"The GALE ENCYCLOPEDIA of MEDICINE"**. |
|
|
|
### 馃攧Bot evolving, stay tuned! |
|
## Useful Links 馃敆 |
|
- **Model:** [Llama-2-7B-Chat-GGML](https://huggingface.co./TheBloke/Llama-2-7B-Chat-GGML) 馃摎 |
|
- **GitHub:** [ThisIs-Developer/Llama-2-GGML-Medical-Chatbot](https://github.com/ThisIs-Developer/Llama-2-GGML-Medical-Chatbot) 馃挰 |
|
''') |
|
add_vertical_space(1) |
|
st.write('Made by [@ThisIs-Developer](https://huggingface.co./ThisIs-Developer)') |
|
|
|
st.title("Llama-2-GGML Medical Chatbot") |
|
st.markdown( |
|
""" |
|
<style> |
|
.chat-container { |
|
display: flex; |
|
flex-direction: column; |
|
height: 400px; |
|
overflow-y: auto; |
|
padding: 10px; |
|
color: white; /* Font color */ |
|
} |
|
.user-bubble { |
|
background-color: #007bff; /* Blue color for user */ |
|
align-self: flex-end; |
|
border-radius: 10px; |
|
padding: 8px; |
|
margin: 5px; |
|
max-width: 70%; |
|
word-wrap: break-word; |
|
} |
|
.bot-bubble { |
|
background-color: #363636; /* Slightly lighter background color */ |
|
align-self: flex-start; |
|
border-radius: 10px; |
|
padding: 8px; |
|
margin: 5px; |
|
max-width: 70%; |
|
word-wrap: break-word; |
|
} |
|
</style> |
|
""" |
|
, unsafe_allow_html=True) |
|
|
|
conversation = st.session_state.get("conversation", []) |
|
|
|
query = st.text_input("Ask your question here:", key="user_input") |
|
if st.button("Get Answer"): |
|
if query: |
|
with st.spinner("Processing your question..."): |
|
conversation.append({"role": "user", "message": query}) |
|
|
|
answer = qa_bot(query) |
|
conversation.append({"role": "bot", "message": answer}) |
|
st.session_state.conversation = conversation |
|
else: |
|
st.warning("Please input a question.") |
|
|
|
chat_container = st.empty() |
|
chat_bubbles = ''.join([f'<div class="{c["role"]}-bubble">{c["message"]}</div>' for c in conversation]) |
|
chat_container.markdown(f'<div class="chat-container">{chat_bubbles}</div>', unsafe_allow_html=True) |
|
|
|
if __name__ == "__main__": |
|
main() |
|
|