Text Classification
PEFT
Safetensors
English

Unable to load model with PEFT due to nested file structure

#1
by david-rai - opened

Hey folks, thanks for releasing this model! Very excited to test it out.

Currently, I am not able to load the adapter model using PEFT. For example, running the following:

model = AutoPeftModelForCausalLM.from_pretrained(
   'nvidia/llama-3.1-nemoguard-8b-content-safety',
    torch_dtype=torch_dtype,
    low_cpu_mem_usage=True,
)

will return the error ValueError: Can't find 'adapter_config.json' at 'nvidia/llama-3.1-nemoguard-8b-content-safety'.

The huggingface "Use this model" feature also detects that there is an issue:
image.png

I suspect that this is because the adapter files are stored within a nested llama-3.1-nemoguard-8b-content-safety-lora-adapter directory, as opposed to the base nvidia/llama-3.1-nemoguard-8b-content-safety directory.

Sign up or log in to comment