Update app.py
Browse files
app.py
CHANGED
@@ -10,13 +10,12 @@ import numpy as np
|
|
10 |
import streamlit as st
|
11 |
from typing import List
|
12 |
from numpy import ndarray
|
13 |
-
from
|
14 |
-
from llama_cpp import Llama, GGML_TYPE_I8
|
15 |
from optimum_encoder import OptimumEncoder
|
|
|
16 |
from unstructured.partition.auto import partition
|
17 |
from fastembed import SparseEmbedding, SparseTextEmbedding
|
18 |
from unstructured.nlp.tokenize import download_nltk_packages
|
19 |
-
from fastembed.sparse.splade_pp import supported_splade_models
|
20 |
from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
|
21 |
from langchain_experimental.text_splitter import SemanticChunker
|
22 |
from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
|
@@ -190,7 +189,8 @@ def load_models_and_documents():
|
|
190 |
)
|
191 |
|
192 |
dense_model = OptimumEncoder(
|
193 |
-
device="cuda"
|
|
|
194 |
)
|
195 |
|
196 |
sparse_model = SparseTextEmbedding(
|
|
|
10 |
import streamlit as st
|
11 |
from typing import List
|
12 |
from numpy import ndarray
|
13 |
+
from llama_cpp import Llama
|
|
|
14 |
from optimum_encoder import OptimumEncoder
|
15 |
+
from qdrant_client import QdrantClient, models
|
16 |
from unstructured.partition.auto import partition
|
17 |
from fastembed import SparseEmbedding, SparseTextEmbedding
|
18 |
from unstructured.nlp.tokenize import download_nltk_packages
|
|
|
19 |
from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
|
20 |
from langchain_experimental.text_splitter import SemanticChunker
|
21 |
from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
|
|
|
189 |
)
|
190 |
|
191 |
dense_model = OptimumEncoder(
|
192 |
+
device="cuda",
|
193 |
+
cache_dir=os.getenv('HF_HOME')
|
194 |
)
|
195 |
|
196 |
sparse_model = SparseTextEmbedding(
|