Is vision_model config now correct?
#11
by
AisingioroHao0
- opened
image_encoder=transformers.CLIPVisionModelWithProjection.from_pretrained("laion/CLIP-ViT-bigG-14-laion2B-39B-b160k")
RuntimeError: Error(s) in loading state_dict for CLIPVisionModelWithProjection:
size mismatch for visual_projection.weight: copying a param with shape torch.Size([1280, 1664]) from checkpoint, the shape in current model is torch.Size([512, 1664]).
You may consider adding `ignore_mismatched_sizes=True` in the model `from_pretrained` method.
fixed
rwightman
changed discussion status to
closed