aie4-final / load_existing_docs.py
pattonma
loaded more docs
d523035
raw
history blame
3.9 kB
import models
#import constants
#from langchain_experimental.text_splitter import SemanticChunker
from langchain_qdrant import QdrantVectorStore, Qdrant
from langchain_community.document_loaders import PyPDFLoader, UnstructuredURLLoader
from qdrant_client.http.models import VectorParams
import pymupdf
import requests
#qdrant = QdrantVectorStore.from_existing_collection(
# embedding=models.basic_embeddings,
# collection_name="kai_test_documents",
# url=constants.QDRANT_ENDPOINT,
#)
def extract_links_from_pdf(pdf_path):
links = []
doc = pymupdf.open(pdf_path)
for page in doc:
for link in page.get_links():
if link['uri']:
links.append(link['uri'])
return links
def load_documents_from_url(url):
try:
# Check if it's a PDF
if url.endswith(".pdf"):
try:
loader = PyPDFLoader(url)
return loader.load()
except Exception as e:
print(f"Error loading PDF from {url}: {e}")
return None
# Fetch the content and check for video pages
try:
response = requests.head(url, timeout=10) # Timeout for fetching headers
content_type = response.headers.get('Content-Type', '')
except Exception as e:
print(f"Error fetching headers from {url}: {e}")
return None
# Ignore video content (flagged for now)
if 'video' in content_type:
return None
if 'youtube' in url:
return None
# Otherwise, treat it as an HTML page
try:
loader = UnstructuredURLLoader([url])
return loader.load()
except Exception as e:
print(f"Error loading HTML from {url}: {e}")
return None
except Exception as e:
print(f"General error loading from {url}: {e}")
return None
#gather kai's docs
filepaths = ["./test_docs/Employee Statistics FINAL.pdf","./test_docs/Employer Statistics FINAL.pdf","./test_docs/Articles To Share.pdf"]
all_links = []
for pdf_path in filepaths:
all_links.extend(extract_links_from_pdf(pdf_path))
unique_links = list(set(all_links))
print(unique_links)
documents = []
for link in unique_links:
doc = load_documents_from_url(link)
#print(f"loaded doc from {link}")
if doc:
documents.extend(doc)
#print(len(documents))
semantic_split_docs = models.semanticChunker.split_documents(documents)
RCTS_split_docs = models.RCTS.split_documents(documents)
#for file in filepaths:
# loader = PyPDFLoader(file)
# documents = loader.load()
# for doc in documents:
# doc.metadata = {
# "source": file,
# "tag": "employee" if "employee" in file.lower() else "employer"
# }
# all_documents.extend(documents)
#chunk them
#semantic_split_docs = models.semanticChunker.split_documents(all_documents)
#add them to the existing qdrant client
collection_name = "docs_from_ripped_urls_recursive"
collections = models.qdrant_client.get_collections()
collection_names = [collection.name for collection in collections.collections]
# If the collection does not exist, create it
if collection_name not in collection_names:
models.qdrant_client.create_collection(
collection_name=collection_name,
vectors_config=VectorParams(size=1536, distance="Cosine")
)
qdrant_vector_store = QdrantVectorStore(
client=models.qdrant_client,
collection_name=collection_name,
embedding=models.te3_small
)
qdrant_vector_store.add_documents(RCTS_split_docs)
collection_info = models.qdrant_client.get_collection(collection_name)
print(f"Number of points in collection: {collection_info.points_count}")