AalianK commited on
Commit
2df00a4
Β·
verified Β·
1 Parent(s): 728a310

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -14
app.py CHANGED
@@ -36,20 +36,20 @@ static_file_paths = [
36
  device = torch.device("cuda" if torch.cuda.is_available() else "CPU")
37
 
38
  # Load documents
39
- loaders = [PyPDFLoader(x) for x in static_file_paths]
40
- pages = []
41
- for loader in loaders:
42
- pages.extend(loader.load())
43
- text_splitter = RecursiveCharacterTextSplitter(
44
- chunk_size=600,
45
- chunk_overlap=40,
46
- )
47
- doc_splits = text_splitter.split_documents(pages)
48
- embedding = HuggingFaceEmbeddings()
49
- vectordb = Chroma.from_documents(
50
- documents=doc_splits,
51
- embedding=embedding,
52
- )
53
 
54
  # Load model
55
  tokenizer = AutoTokenizer.from_pretrained(llm_name, token=os.environ['HUGGINGFACEHUB_API_TOKEN'],)
 
36
  device = torch.device("cuda" if torch.cuda.is_available() else "CPU")
37
 
38
  # Load documents
39
+ #loaders = [PyPDFLoader(x) for x in static_file_paths]
40
+ #pages = []
41
+ #for loader in loaders:
42
+ # pages.extend(loader.load())
43
+ #text_splitter = RecursiveCharacterTextSplitter(
44
+ # chunk_size=600,
45
+ # chunk_overlap=40,
46
+ #)
47
+ #doc_splits = text_splitter.split_documents(pages)
48
+ #embedding = HuggingFaceEmbeddings()
49
+ #vectordb = Chroma.from_documents(
50
+ # documents=doc_splits,
51
+ # embedding=embedding,
52
+ #)
53
 
54
  # Load model
55
  tokenizer = AutoTokenizer.from_pretrained(llm_name, token=os.environ['HUGGINGFACEHUB_API_TOKEN'],)