jmzzomg's picture
Update README.md
e0c24ed verified
metadata
license: mit
pipeline_tag: image-classification

ONNX port of sentence-transformers/clip-ViT-B-32.

This model is intended to be used for image classification and similarity searches.

Usage

Here's an example of performing inference using the model with FastEmbed.

from fastembed import ImageEmbedding

images = [
    "./path/to/image1.jpg",
    "./path/to/image2.jpg",
]

model = ImageEmbedding(model_name="Qdrant/clip-ViT-B-32-vision")
embeddings = list(model.embed(images))

# [
#   array([-0.1115,  0.0097,  0.0052,  0.0195, ...], dtype=float32),
#   array([-0.1019,  0.0635, -0.0332,  0.0522, ...], dtype=float32)
# ]