Hugging Face
Models
Datasets
Spaces
Posts
Docs
Solutions
Pricing
Log In
Sign Up
aws-neuron
/
optimum-neuron-cache
like
12
Follow
AWS Inferentia and Trainium
61
License:
apache-2.0
Model card
Files
Files and versions
Community
240
fc35d6d
optimum-neuron-cache
/
inference-cache-config
8 contributors
History:
22 commits
dacorvo
HF staff
Add more batch_size for mistral on smaller instances
545cd4d
verified
6 months ago
gpt2.json
398 Bytes
Add more gpt2 configurations
7 months ago
llama-variants.json
2.63 kB
Use princeton-nlp/Sheared-LLaMA-1.3B as a test model
6 months ago
llama2.json
2.31 kB
Remove llama2 7B config for 24 cores
6 months ago
llama3.json
1.16 kB
Update inference-cache-config/llama3.json
6 months ago
mistral-variants.json
3.29 kB
Remove SalesForce embedding model
8 months ago
mistral.json
1.46 kB
Add more batch_size for mistral on smaller instances
6 months ago
stable-diffusion.json
1.91 kB
Create stable-diffusion.json (#43)
8 months ago