Spaces:
Running
Running
""" | |
Upload weights to huggingface. | |
Usage: | |
python3 -m fastchat.model.upload_hub --model-path ~/model_weights/vicuna-13b --hub-repo-id lmsys/vicuna-13b-v1.3 | |
""" | |
import argparse | |
import tempfile | |
import torch | |
from transformers import AutoTokenizer, AutoModelForCausalLM | |
def upload_hub(model_path, hub_repo_id, component, private): | |
if component == "all": | |
components = ["model", "tokenizer"] | |
else: | |
components = [component] | |
kwargs = {"push_to_hub": True, "repo_id": hub_repo_id, "private": args.private} | |
if "model" in components: | |
model = AutoModelForCausalLM.from_pretrained( | |
model_path, torch_dtype=torch.float16, low_cpu_mem_usage=True | |
) | |
with tempfile.TemporaryDirectory() as tmp_path: | |
model.save_pretrained(tmp_path, **kwargs) | |
if "tokenizer" in components: | |
tokenizer = AutoTokenizer.from_pretrained(model_path, use_fast=False) | |
with tempfile.TemporaryDirectory() as tmp_path: | |
tokenizer.save_pretrained(tmp_path, **kwargs) | |
if __name__ == "__main__": | |
parser = argparse.ArgumentParser() | |
parser.add_argument("--model-path", type=str, required=True) | |
parser.add_argument("--hub-repo-id", type=str, required=True) | |
parser.add_argument( | |
"--component", type=str, choices=["all", "model", "tokenizer"], default="all" | |
) | |
parser.add_argument("--private", action="store_true") | |
args = parser.parse_args() | |
upload_hub(args.model_path, args.hub_repo_id, args.component, args.private) | |