Update app.py
Browse files
app.py
CHANGED
@@ -114,13 +114,15 @@ def clear_all():
|
|
114 |
def add_text(chatbot, history, prompt, file):
|
115 |
if (file == None):
|
116 |
chatbot = chatbot +[(prompt, None)]
|
|
|
117 |
else:
|
118 |
if (prompt == ""):
|
119 |
chatbot=chatbot + [((file.name,), "Prompt fehlt!")]
|
|
|
120 |
else:
|
121 |
chatbot = chatbot +[((file.name,), None), (prompt, None)]
|
122 |
-
|
123 |
-
|
124 |
return chatbot, history, prompt, file, gr.Image(visible = False), "" #gr.Image( label=None, size=(30,30), visible=False, scale=1) #gr.Textbox(value="", interactive=False)
|
125 |
|
126 |
def add_text2(chatbot, prompt):
|
@@ -204,6 +206,7 @@ def process_image(image_path, prompt):
|
|
204 |
##################################################
|
205 |
#openassistant um uploaded Files zu analysieren
|
206 |
def create_assistant(prompt, file):
|
|
|
207 |
client = OpenAI()
|
208 |
assistant = client.beta.assistants.create(name="File Analysator",instructions=template, model="gpt-4-1106-preview",)
|
209 |
thread = client.beta.threads.create()
|
@@ -213,8 +216,6 @@ def create_assistant(prompt, file):
|
|
213 |
thread, run = create_thread_and_run(prompt, client, assistant.id)
|
214 |
run = wait_on_run(run, thread, client)
|
215 |
response = get_response(thread, client, assistant.id)
|
216 |
-
print("response...................")
|
217 |
-
print(response)
|
218 |
result = response.data[1].content[0].text.value
|
219 |
return result
|
220 |
|
@@ -228,7 +229,7 @@ def generate_auswahl(prompt, file, chatbot, history, rag_option, model_option, o
|
|
228 |
#kein Bild hochgeladen -> auf Text antworten...
|
229 |
if (file == None):
|
230 |
result = generate_text(prompt, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
231 |
-
history = history + [(prompt, result)]
|
232 |
else:
|
233 |
#Es wurde ein Bild angehängt -> wenn prompt dazu, das Bild analysieren
|
234 |
#geht nur über spezielle OpenAI-Schnittstelle...
|
@@ -238,11 +239,13 @@ def generate_auswahl(prompt, file, chatbot, history, rag_option, model_option, o
|
|
238 |
else:
|
239 |
result = generate_text_zu_doc(file, prompt, k, rag_option, chatbot)
|
240 |
|
241 |
-
history = history + [((file,), None),(prompt, result)]
|
242 |
|
243 |
chatbot[-1][1] = ""
|
|
|
244 |
for character in result:
|
245 |
chatbot[-1][1] += character
|
|
|
246 |
time.sleep(0.03)
|
247 |
yield chatbot, history, None, "Generating"
|
248 |
if shared_state.interrupted:
|
|
|
114 |
def add_text(chatbot, history, prompt, file):
|
115 |
if (file == None):
|
116 |
chatbot = chatbot +[(prompt, None)]
|
117 |
+
history = history +[(prompt, None)]
|
118 |
else:
|
119 |
if (prompt == ""):
|
120 |
chatbot=chatbot + [((file.name,), "Prompt fehlt!")]
|
121 |
+
history=history + [((file.name,), "Prompt fehlt!")]
|
122 |
else:
|
123 |
chatbot = chatbot +[((file.name,), None), (prompt, None)]
|
124 |
+
history = history +[((file.name,), None), (prompt, None)]
|
125 |
+
|
126 |
return chatbot, history, prompt, file, gr.Image(visible = False), "" #gr.Image( label=None, size=(30,30), visible=False, scale=1) #gr.Textbox(value="", interactive=False)
|
127 |
|
128 |
def add_text2(chatbot, prompt):
|
|
|
206 |
##################################################
|
207 |
#openassistant um uploaded Files zu analysieren
|
208 |
def create_assistant(prompt, file):
|
209 |
+
#zentral einmal erzeugen!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
|
210 |
client = OpenAI()
|
211 |
assistant = client.beta.assistants.create(name="File Analysator",instructions=template, model="gpt-4-1106-preview",)
|
212 |
thread = client.beta.threads.create()
|
|
|
216 |
thread, run = create_thread_and_run(prompt, client, assistant.id)
|
217 |
run = wait_on_run(run, thread, client)
|
218 |
response = get_response(thread, client, assistant.id)
|
|
|
|
|
219 |
result = response.data[1].content[0].text.value
|
220 |
return result
|
221 |
|
|
|
229 |
#kein Bild hochgeladen -> auf Text antworten...
|
230 |
if (file == None):
|
231 |
result = generate_text(prompt, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
232 |
+
#history = history + [(prompt, result)]
|
233 |
else:
|
234 |
#Es wurde ein Bild angehängt -> wenn prompt dazu, das Bild analysieren
|
235 |
#geht nur über spezielle OpenAI-Schnittstelle...
|
|
|
239 |
else:
|
240 |
result = generate_text_zu_doc(file, prompt, k, rag_option, chatbot)
|
241 |
|
242 |
+
#history = history + [((file,), None),(prompt, result)]
|
243 |
|
244 |
chatbot[-1][1] = ""
|
245 |
+
history[-1][1] = ""
|
246 |
for character in result:
|
247 |
chatbot[-1][1] += character
|
248 |
+
history[-1][1] += character
|
249 |
time.sleep(0.03)
|
250 |
yield chatbot, history, None, "Generating"
|
251 |
if shared_state.interrupted:
|