Error while loading the model using safe tensors
#22
by
tasheer10
- opened
Getting the following error while executing the following code :
pipe = pipeline("text-generation", model="HuggingFaceH4/starchat-beta", torch_dtype=torch.bfloat16, device_map="auto")
Error:
result[k] = f.get_tensor(k)
RuntimeError: Viewing a tensor as a new dtype with a different number of bytes per element is not supported.
tasheer10
changed discussion title from
Error while loading the model using saved tensors
to Error while loading the model using safe tensors