Update app.py
Browse files
app.py
CHANGED
@@ -33,7 +33,13 @@ from qdrant_client.models import (
|
|
33 |
FusionQuery,
|
34 |
Fusion,
|
35 |
SearchRequest,
|
36 |
-
Modifier
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
)
|
38 |
|
39 |
def make_points(texts: List[str], metadatas: List[dict], dense: List[List[float]], sparse: List[SparseEmbedding])-> List[PointStruct]:
|
@@ -161,26 +167,26 @@ def load_models_and_documents():
|
|
161 |
client.create_collection(
|
162 |
collection_name,
|
163 |
{
|
164 |
-
"text-dense":
|
165 |
size=1024,
|
166 |
-
distance=
|
167 |
on_disk=False
|
168 |
)
|
169 |
},
|
170 |
{
|
171 |
-
"text-sparse":
|
172 |
-
index=
|
173 |
on_disk=False
|
174 |
),
|
175 |
modifier=Modifier.IDF
|
176 |
)
|
177 |
},
|
178 |
2,
|
179 |
-
optimizers_config=
|
180 |
indexing_threshold=0,
|
181 |
default_segment_number=4
|
182 |
),
|
183 |
-
hnsw_config=
|
184 |
on_disk=False,
|
185 |
m=64,
|
186 |
ef_construct=512
|
@@ -273,7 +279,7 @@ def load_models_and_documents():
|
|
273 |
)
|
274 |
client.update_collection(
|
275 |
collection_name=collection_name,
|
276 |
-
optimizer_config=
|
277 |
)
|
278 |
|
279 |
return client, collection_name, tokenizer, model, llm, dense_model, sparse_model
|
@@ -379,7 +385,7 @@ if __name__ == '__main__':
|
|
379 |
else:
|
380 |
stop_token_ids = [151329, 151336, 151338]
|
381 |
sampling_params = SamplingParams(temperature=0.75, max_tokens=35, stop_token_ids=stop_token_ids)
|
382 |
-
prompt = [{"role": "user", "content": "
|
383 |
inputs = tokenizer.apply_chat_template(prompt, tokenize=False, add_generation_prompt=True)
|
384 |
outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
|
385 |
|
|
|
33 |
FusionQuery,
|
34 |
Fusion,
|
35 |
SearchRequest,
|
36 |
+
Modifier,
|
37 |
+
OptimizersConfigDiff,
|
38 |
+
HnswConfigDiff,
|
39 |
+
Distance,
|
40 |
+
VectorParams,
|
41 |
+
SparseVectorParams,
|
42 |
+
SparseIndexParams
|
43 |
)
|
44 |
|
45 |
def make_points(texts: List[str], metadatas: List[dict], dense: List[List[float]], sparse: List[SparseEmbedding])-> List[PointStruct]:
|
|
|
167 |
client.create_collection(
|
168 |
collection_name,
|
169 |
{
|
170 |
+
"text-dense": VectorParams(
|
171 |
size=1024,
|
172 |
+
distance=Distance.COSINE,
|
173 |
on_disk=False
|
174 |
)
|
175 |
},
|
176 |
{
|
177 |
+
"text-sparse": SparseVectorParams(
|
178 |
+
index=SparseIndexParams(
|
179 |
on_disk=False
|
180 |
),
|
181 |
modifier=Modifier.IDF
|
182 |
)
|
183 |
},
|
184 |
2,
|
185 |
+
optimizers_config=OptimizersConfigDiff(
|
186 |
indexing_threshold=0,
|
187 |
default_segment_number=4
|
188 |
),
|
189 |
+
hnsw_config=HnswConfigDiff(
|
190 |
on_disk=False,
|
191 |
m=64,
|
192 |
ef_construct=512
|
|
|
279 |
)
|
280 |
client.update_collection(
|
281 |
collection_name=collection_name,
|
282 |
+
optimizer_config=OptimizersConfigDiff(indexing_threshold=20000)
|
283 |
)
|
284 |
|
285 |
return client, collection_name, tokenizer, model, llm, dense_model, sparse_model
|
|
|
385 |
else:
|
386 |
stop_token_ids = [151329, 151336, 151338]
|
387 |
sampling_params = SamplingParams(temperature=0.75, max_tokens=35, stop_token_ids=stop_token_ids)
|
388 |
+
prompt = [{"role": "user", "content": f"{}\nExplain the above in one sentence:"}]
|
389 |
inputs = tokenizer.apply_chat_template(prompt, tokenize=False, add_generation_prompt=True)
|
390 |
outputs = llm.generate(prompts=inputs, sampling_params=sampling_params)
|
391 |
|