metadata
license: mit
datasets:
- tomg-group-umd/ContraStyles
library_name: transformers
An unofficial implementation of CSD
Inspired by vvmatorin/CSD, the difference in this implementation is that the CLIP backbone is not the openai clip class but an instance of CLIPVisionModel
.
Inference:
from PIL import Image
from transformers import AutoProcessor, AutoModel
model = AutoModel.from_pretrained("NagaSaiAbhinay/CSD", trust_remote_code=True)
processor = AutoProcessor.from_pretrained("NagaSaiAbhinay/CSD")
img = Image.open('test_image.png')
pixel_values = processor