gufett0 commited on
Commit
9b05e50
·
1 Parent(s): 8b64b0a

added introductory prompt

Browse files
Files changed (2) hide show
  1. backend.py +11 -3
  2. data/paymentprova.txt +0 -0
backend.py CHANGED
@@ -38,7 +38,7 @@ Settings.llm = GemmaLLMInterface()
38
  documents_paths = {
39
  'blockchain': 'data/blockchainprova.txt',
40
  'metaverse': 'data/metaverseprova.txt',
41
- 'payment': 'data/payment'
42
  }
43
 
44
  session_state = {"index": False,
@@ -72,7 +72,7 @@ def build_index(path: str):
72
 
73
 
74
 
75
- @spaces.GPU(duration=20)
76
  def handle_query(query_str: str,
77
  chat_history: list[tuple[str, str]]) -> Iterator[str]:
78
 
@@ -137,11 +137,19 @@ def handle_query(query_str: str,
137
 
138
  #response = chat_engine.chat(query_str)
139
  for token in response.response_gen:
140
- #if not token.startswith("system:") and not token.startswith("user:"):
 
 
 
141
 
142
  outputs.append(token)
143
  print(f"Generated token: {token}")
144
  yield "".join(outputs)
 
 
 
 
 
145
 
146
 
147
  except Exception as e:
 
38
  documents_paths = {
39
  'blockchain': 'data/blockchainprova.txt',
40
  'metaverse': 'data/metaverseprova.txt',
41
+ 'payment': 'data/paymentprova.txt'
42
  }
43
 
44
  session_state = {"index": False,
 
72
 
73
 
74
 
75
+ @spaces.GPU()
76
  def handle_query(query_str: str,
77
  chat_history: list[tuple[str, str]]) -> Iterator[str]:
78
 
 
137
 
138
  #response = chat_engine.chat(query_str)
139
  for token in response.response_gen:
140
+ if not token.startswith("system:") and not token.startswith("user:"):
141
+
142
+ if token.startswith("file_path:"):
143
+ source = token
144
 
145
  outputs.append(token)
146
  print(f"Generated token: {token}")
147
  yield "".join(outputs)
148
+
149
+ outputs.append(f"Fonti utilizzate: {source}")
150
+ yield "".join(outputs)
151
+
152
+
153
 
154
 
155
  except Exception as e:
data/paymentprova.txt ADDED
The diff for this file is too large to render. See raw diff