devve1 commited on
Commit
3e1894f
1 Parent(s): a0b85bb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -16,7 +16,6 @@ from numpy import ndarray
16
  from outlines import models
17
  from llama_cpp import Llama
18
  import hydralit_components as hc
19
- from transformers import AutoTokenizer
20
  from qdrant_client import QdrantClient
21
  from optimum_encoder import OptimumEncoder
22
  from huggingface_hub import snapshot_download
@@ -25,6 +24,7 @@ from fastembed import SparseEmbedding, SparseTextEmbedding
25
  from unstructured.nlp.tokenize import download_nltk_packages
26
  from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
27
  from langchain_experimental.text_splitter import SemanticChunker
 
28
  from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
29
  from qdrant_client.models import (
30
  NamedSparseVector,
@@ -147,6 +147,8 @@ def main(query: str, client: QdrantClient, collection_name: str, llm: Llama, den
147
  def load_models_and_documents():
148
  with st.spinner('Load models...'):
149
  model_path = snapshot_download(repo_id="Ichigo2899/mistralai-Mistral-Nemo-Instruct-2407-AWQ")
 
 
150
 
151
  llm = vllm.LLM(
152
  model_path,
 
16
  from outlines import models
17
  from llama_cpp import Llama
18
  import hydralit_components as hc
 
19
  from qdrant_client import QdrantClient
20
  from optimum_encoder import OptimumEncoder
21
  from huggingface_hub import snapshot_download
 
24
  from unstructured.nlp.tokenize import download_nltk_packages
25
  from scipy.sparse import csr_matrix, save_npz, load_npz, vstack
26
  from langchain_experimental.text_splitter import SemanticChunker
27
+ from mistral_common.tokens.tokenizers.mistral import MistralTokenizer
28
  from langchain_community.document_loaders import WikipediaLoader, WebBaseLoader
29
  from qdrant_client.models import (
30
  NamedSparseVector,
 
147
  def load_models_and_documents():
148
  with st.spinner('Load models...'):
149
  model_path = snapshot_download(repo_id="Ichigo2899/mistralai-Mistral-Nemo-Instruct-2407-AWQ")
150
+
151
+ tokenizer = MistralTokenizer.from_file(f"{model_path}/tekken.json")
152
 
153
  llm = vllm.LLM(
154
  model_path,