devve1 commited on
Commit
2ab448b
1 Parent(s): 9cd87f4

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +15 -9
app.py CHANGED
@@ -33,7 +33,13 @@ from qdrant_client.models import (
33
  FusionQuery,
34
  Fusion,
35
  SearchRequest,
36
- Modifier
 
 
 
 
 
 
37
  )
38
 
39
  def make_points(texts: List[str], metadatas: List[dict], dense: List[List[float]], sparse: List[SparseEmbedding])-> List[PointStruct]:
@@ -161,26 +167,26 @@ def load_models_and_documents():
161
  client.create_collection(
162
  collection_name,
163
  {
164
- "text-dense": models.VectorParams(
165
  size=1024,
166
- distance=models.Distance.COSINE,
167
  on_disk=False
168
  )
169
  },
170
  {
171
- "text-sparse": models.SparseVectorParams(
172
- index=models.SparseIndexParams(
173
  on_disk=False
174
  ),
175
  modifier=Modifier.IDF
176
  )
177
  },
178
  2,
179
- optimizers_config=models.OptimizersConfigDiff(
180
  indexing_threshold=0,
181
  default_segment_number=4
182
  ),
183
- hnsw_config=models.HnswConfigDiff(
184
  on_disk=False,
185
  m=64,
186
  ef_construct=512
@@ -273,7 +279,7 @@ def load_models_and_documents():
273
  )
274
  client.update_collection(
275
  collection_name=collection_name,
276
- optimizer_config=models.OptimizersConfigDiff(indexing_threshold=20000)
277
  )
278
 
279
  return client, collection_name, tokenizer, model, llm, dense_model, sparse_model
@@ -379,7 +385,7 @@ if __name__ == '__main__':
379
  else:
380
  stop_token_ids = [151329, 151336, 151338]
381
  sampling_params = SamplingParams(temperature=0.75, max_tokens=35, stop_token_ids=stop_token_ids)
382
- prompt = [{"role": "user", "content": "hello"}]
383
  inputs = tokenizer.apply_chat_template(prompt, tokenize=False, add_generation_prompt=True)
384
  outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
385
 
 
33
  FusionQuery,
34
  Fusion,
35
  SearchRequest,
36
+ Modifier,
37
+ OptimizersConfigDiff,
38
+ HnswConfigDiff,
39
+ Distance,
40
+ VectorParams,
41
+ SparseVectorParams,
42
+ SparseIndexParams
43
  )
44
 
45
  def make_points(texts: List[str], metadatas: List[dict], dense: List[List[float]], sparse: List[SparseEmbedding])-> List[PointStruct]:
 
167
  client.create_collection(
168
  collection_name,
169
  {
170
+ "text-dense": VectorParams(
171
  size=1024,
172
+ distance=Distance.COSINE,
173
  on_disk=False
174
  )
175
  },
176
  {
177
+ "text-sparse": SparseVectorParams(
178
+ index=SparseIndexParams(
179
  on_disk=False
180
  ),
181
  modifier=Modifier.IDF
182
  )
183
  },
184
  2,
185
+ optimizers_config=OptimizersConfigDiff(
186
  indexing_threshold=0,
187
  default_segment_number=4
188
  ),
189
+ hnsw_config=HnswConfigDiff(
190
  on_disk=False,
191
  m=64,
192
  ef_construct=512
 
279
  )
280
  client.update_collection(
281
  collection_name=collection_name,
282
+ optimizer_config=OptimizersConfigDiff(indexing_threshold=20000)
283
  )
284
 
285
  return client, collection_name, tokenizer, model, llm, dense_model, sparse_model
 
385
  else:
386
  stop_token_ids = [151329, 151336, 151338]
387
  sampling_params = SamplingParams(temperature=0.75, max_tokens=35, stop_token_ids=stop_token_ids)
388
+ prompt = [{"role": "user", "content": f"{}\nExplain the above in one sentence:"}]
389
  inputs = tokenizer.apply_chat_template(prompt, tokenize=False, add_generation_prompt=True)
390
  outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
391