devve1 commited on
Commit
31b1908
1 Parent(s): 3c8327a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +23 -23
app.py CHANGED
@@ -337,6 +337,28 @@ if __name__ == '__main__':
337
 
338
  client, collection_name, llm, dense_model, sparse_model = load_models_and_documents()
339
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
340
  st.sidebar.title("Upload your file")
341
  uploaded_files = st.sidebar.file_uploader("Choose a file", accept_multiple_files=True, type=['docx', 'doc', 'odt', 'pptx', 'ppt', 'xlsx', 'csv', 'tsv', 'eml', 'msg', 'rtf', 'epub', 'html', 'xml', 'pdf', 'png', 'jpg', 'heic','txt'])
342
  print(f'uploaded-files : {uploaded_files}')
@@ -358,26 +380,4 @@ if __name__ == '__main__':
358
  dense_embeddings,
359
  sparse_embeddings
360
  )
361
- )
362
-
363
- if "messages" not in st.session_state:
364
- st.session_state.messages = []
365
-
366
- for message in st.session_state.messages:
367
- with st.chat_message(message["role"]):
368
- st.markdown(message["content"])
369
-
370
- if prompt := st.chat_input("Message Video Game Assistant"):
371
- st.chat_message("user").markdown(prompt)
372
- st.session_state.messages.append({"role": "user", "content": prompt})
373
-
374
- ai_response = main(prompt, client, collection_name, llm, dense_model, sparse_model)
375
- response = f"Echo: {ai_response}"
376
- with st.chat_message("assistant"):
377
- message_placeholder = st.empty()
378
- full_response = ""
379
- for chunk in re.split(r'(\s+)', response):
380
- full_response += chunk + " "
381
- time.sleep(0.01)
382
- message_placeholder.markdown(full_response + "▌")
383
- st.session_state.messages.append({"role": "assistant", "content": full_response})
 
337
 
338
  client, collection_name, llm, dense_model, sparse_model = load_models_and_documents()
339
 
340
+ if "messages" not in st.session_state:
341
+ st.session_state.messages = []
342
+
343
+ for message in st.session_state.messages:
344
+ with st.chat_message(message["role"]):
345
+ st.markdown(message["content"])
346
+
347
+ if prompt := st.chat_input("Message Video Game Assistant"):
348
+ st.chat_message("user").markdown(prompt)
349
+ st.session_state.messages.append({"role": "user", "content": prompt})
350
+
351
+ ai_response = main(prompt, client, collection_name, llm, dense_model, sparse_model)
352
+ response = f"Echo: {ai_response}"
353
+ with st.chat_message("assistant"):
354
+ message_placeholder = st.empty()
355
+ full_response = ""
356
+ for chunk in re.split(r'(\s+)', response):
357
+ full_response += chunk + " "
358
+ time.sleep(0.01)
359
+ message_placeholder.markdown(full_response + "▌")
360
+ st.session_state.messages.append({"role": "assistant", "content": full_response})
361
+
362
  st.sidebar.title("Upload your file")
363
  uploaded_files = st.sidebar.file_uploader("Choose a file", accept_multiple_files=True, type=['docx', 'doc', 'odt', 'pptx', 'ppt', 'xlsx', 'csv', 'tsv', 'eml', 'msg', 'rtf', 'epub', 'html', 'xml', 'pdf', 'png', 'jpg', 'heic','txt'])
364
  print(f'uploaded-files : {uploaded_files}')
 
380
  dense_embeddings,
381
  sparse_embeddings
382
  )
383
+ )