Spaces:
Starting
on
T4
Starting
on
T4
Update app.py
Browse files
app.py
CHANGED
@@ -16,7 +16,6 @@ from numpy import ndarray
|
|
16 |
from outlines import models
|
17 |
from llama_cpp import Llama
|
18 |
import hydralit_components as hc
|
19 |
-
from transformers import AutoTokenizer
|
20 |
from qdrant_client import QdrantClient
|
21 |
from optimum_encoder import OptimumEncoder
|
22 |
from huggingface_hub import snapshot_download
|
@@ -25,6 +24,7 @@ from fastembed import SparseEmbedding, SparseTextEmbedding
|
|
25 |
from unstructured.nlp.tokenize import download_nltk_packages
|
26 |
from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
|
27 |
from langchain_experimental.text_splitter import SemanticChunker
|
|
|
28 |
from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
|
29 |
from qdrant_client.models import (
|
30 |
NamedSparseVector,
|
@@ -147,6 +147,8 @@ def main(query: str, client: QdrantClient, collection_name: str, llm: Llama, den
|
|
147 |
def load_models_and_documents():
|
148 |
with st.spinner('Load models...'):
|
149 |
model_path = snapshot_download(repo_id="Ichigo2899/mistralai-Mistral-Nemo-Instruct-2407-AWQ")
|
|
|
|
|
150 |
|
151 |
llm = vllm.LLM(
|
152 |
model_path,
|
|
|
16 |
from outlines import models
|
17 |
from llama_cpp import Llama
|
18 |
import hydralit_components as hc
|
|
|
19 |
from qdrant_client import QdrantClient
|
20 |
from optimum_encoder import OptimumEncoder
|
21 |
from huggingface_hub import snapshot_download
|
|
|
24 |
from unstructured.nlp.tokenize import download_nltk_packages
|
25 |
from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
|
26 |
from langchain_experimental.text_splitter import SemanticChunker
|
27 |
+
from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
|
28 |
from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
|
29 |
from qdrant_client.models import (
|
30 |
NamedSparseVector,
|
|
|
147 |
def load_models_and_documents():
|
148 |
with st.spinner('Load models...'):
|
149 |
model_path = snapshot_download(repo_id="Ichigo2899/mistralai-Mistral-Nemo-Instruct-2407-AWQ")
|
150 |
+
|
151 |
+
tokenizer = MistralTokenizer.from_file(f"{model_path}/tekken.json")
|
152 |
|
153 |
llm = vllm.LLM(
|
154 |
model_path,
|