Optimum Nvidia
AI & ML interests
None defined yet.
Optimum-Nvidia allows you to easily leverage Nvidia's TensorRT-LLM Inference tool through a seemlessly integration following huggingface/transformers API.
This organisation holds prebuilt TensorRT-LLM compatible engines for various fondational models one can use, fork and deploy to get started as fast as possible and benefits from out of the box peak performances on Nvidia hardware.
Prebuilt engines will attempt (as much as possible) to be build with the best options available and will push updated models following additions to TensorRT-LLM repository. This can include (not limited to):
- Leveraging
float8
quantization on supported hardware (H100/L4/L40/RTX 40xx) - Enabling
float8
orint8
KV cache - Enabling in-flight batching for dynamic batching when used in combinaison with Nvidia Triton Inference Server
- Enabling xQA attention kernels
Current engines are targetting the following Nvidia TensorCore GPUs and can be found using specific branch matching the targetted GPU in the repo:
Feel free to open-up discussions and ask for models to support through the community tab
- The Optimum-Nvidia team at 🤗