Compiled CodeLlama-7b-hf using optimum-neuron
optimum-cli export neuron --model codellama/CodeLlama-7b-hf --batch_size 1 --sequence_length 1024 --num_cores 2 --auto_cast_type fp16 ./models/codellama/CodeLlama-7b-hf
- Downloads last month
- 11
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.