Update optimum_encoder.py
Browse files- optimum_encoder.py +3 -2
optimum_encoder.py
CHANGED
@@ -8,9 +8,10 @@ from pydantic.v1 import PrivateAttr
|
|
8 |
from fastembed.common.utils import normalize
|
9 |
from semantic_router.encoders import BaseEncoder
|
10 |
from semantic_router.utils.logger import logger
|
|
|
11 |
|
12 |
|
13 |
-
class OptimumEncoder(BaseEncoder):
|
14 |
name: str = "mixedbread-ai/mxbai-embed-large-v1"
|
15 |
type: str = "huggingface"
|
16 |
score_threshold: float = 0.5
|
@@ -104,7 +105,7 @@ class OptimumEncoder(BaseEncoder):
|
|
104 |
|
105 |
return tokenizer, ort_model
|
106 |
|
107 |
-
def
|
108 |
self,
|
109 |
docs: List[str],
|
110 |
batch_size: int = 32,
|
|
|
8 |
from fastembed.common.utils import normalize
|
9 |
from semantic_router.encoders import BaseEncoder
|
10 |
from semantic_router.utils.logger import logger
|
11 |
+
from langchain_core.embeddings import Embeddings
|
12 |
|
13 |
|
14 |
+
class OptimumEncoder(BaseEncoder, Embeddings):
|
15 |
name: str = "mixedbread-ai/mxbai-embed-large-v1"
|
16 |
type: str = "huggingface"
|
17 |
score_threshold: float = 0.5
|
|
|
105 |
|
106 |
return tokenizer, ort_model
|
107 |
|
108 |
+
def embed_documents(
|
109 |
self,
|
110 |
docs: List[str],
|
111 |
batch_size: int = 32,
|