Update app.py
Browse files
app.py
CHANGED
@@ -337,6 +337,28 @@ if __name__ == '__main__':
|
|
337 |
|
338 |
client, collection_name, llm, dense_model, sparse_model = load_models_and_documents()
|
339 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
340 |
st.sidebar.title("Upload your file")
|
341 |
uploaded_files = st.sidebar.file_uploader("Choose a file", accept_multiple_files=True, type=['docx', 'doc', 'odt', 'pptx', 'ppt', 'xlsx', 'csv', 'tsv', 'eml', 'msg', 'rtf', 'epub', 'html', 'xml', 'pdf', 'png', 'jpg', 'heic','txt'])
|
342 |
print(f'uploaded-files : {uploaded_files}')
|
@@ -358,26 +380,4 @@ if __name__ == '__main__':
|
|
358 |
dense_embeddings,
|
359 |
sparse_embeddings
|
360 |
)
|
361 |
-
)
|
362 |
-
|
363 |
-
if "messages" not in st.session_state:
|
364 |
-
st.session_state.messages = []
|
365 |
-
|
366 |
-
for message in st.session_state.messages:
|
367 |
-
with st.chat_message(message["role"]):
|
368 |
-
st.markdown(message["content"])
|
369 |
-
|
370 |
-
if prompt := st.chat_input("Message Video Game Assistant"):
|
371 |
-
st.chat_message("user").markdown(prompt)
|
372 |
-
st.session_state.messages.append({"role": "user", "content": prompt})
|
373 |
-
|
374 |
-
ai_response = main(prompt, client, collection_name, llm, dense_model, sparse_model)
|
375 |
-
response = f"Echo: {ai_response}"
|
376 |
-
with st.chat_message("assistant"):
|
377 |
-
message_placeholder = st.empty()
|
378 |
-
full_response = ""
|
379 |
-
for chunk in re.split(r'(\s+)', response):
|
380 |
-
full_response += chunk + " "
|
381 |
-
time.sleep(0.01)
|
382 |
-
message_placeholder.markdown(full_response + "▌")
|
383 |
-
st.session_state.messages.append({"role": "assistant", "content": full_response})
|
|
|
337 |
|
338 |
client, collection_name, llm, dense_model, sparse_model = load_models_and_documents()
|
339 |
|
340 |
+
if "messages" not in st.session_state:
|
341 |
+
st.session_state.messages = []
|
342 |
+
|
343 |
+
for message in st.session_state.messages:
|
344 |
+
with st.chat_message(message["role"]):
|
345 |
+
st.markdown(message["content"])
|
346 |
+
|
347 |
+
if prompt := st.chat_input("Message Video Game Assistant"):
|
348 |
+
st.chat_message("user").markdown(prompt)
|
349 |
+
st.session_state.messages.append({"role": "user", "content": prompt})
|
350 |
+
|
351 |
+
ai_response = main(prompt, client, collection_name, llm, dense_model, sparse_model)
|
352 |
+
response = f"Echo: {ai_response}"
|
353 |
+
with st.chat_message("assistant"):
|
354 |
+
message_placeholder = st.empty()
|
355 |
+
full_response = ""
|
356 |
+
for chunk in re.split(r'(\s+)', response):
|
357 |
+
full_response += chunk + " "
|
358 |
+
time.sleep(0.01)
|
359 |
+
message_placeholder.markdown(full_response + "▌")
|
360 |
+
st.session_state.messages.append({"role": "assistant", "content": full_response})
|
361 |
+
|
362 |
st.sidebar.title("Upload your file")
|
363 |
uploaded_files = st.sidebar.file_uploader("Choose a file", accept_multiple_files=True, type=['docx', 'doc', 'odt', 'pptx', 'ppt', 'xlsx', 'csv', 'tsv', 'eml', 'msg', 'rtf', 'epub', 'html', 'xml', 'pdf', 'png', 'jpg', 'heic','txt'])
|
364 |
print(f'uploaded-files : {uploaded_files}')
|
|
|
380 |
dense_embeddings,
|
381 |
sparse_embeddings
|
382 |
)
|
383 |
+
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|