neuralmagic/Phi-3-medium-128k-instruct-quantized.w4a16 Text Generation • Updated 10 days ago • 4.54k • 3
nm-testing/Meta-Llama-3-8B-Instruct-W8A8-FP8-Channelwise-compressed-tensors Text Generation • Updated 10 days ago • 1.49k
nm-testing/Meta-Llama-3-8B-Instruct-Non-Uniform-compressed-tensors Text Generation • Updated 10 days ago • 8
nm-testing/Meta-Llama-3-8B-Instruct-W4A16-ACTORDER-compressed-tensors-test Text Generation • Updated 10 days ago • 9
nm-testing/Meta-Llama-3-70B-Instruct-W8A8-Dynamic-Per-Token-test Text Generation • Updated 10 days ago • 10
nm-testing/Meta-Llama-3-70B-Instruct-W8A8-Dynamic-Per-Token Text Generation • Updated 10 days ago • 10
neuralmagic/Meta-Llama-3.1-8B-Instruct-FP8-dynamic Text Generation • Updated about 9 hours ago • 5k • 5
neuralmagic/Meta-Llama-3.1-70B-Instruct-FP8-dynamic Text Generation • Updated about 9 hours ago • 1.78k • 2
neuralmagic/Meta-Llama-3.1-405B-Instruct-FP8-dynamic Text Generation • Updated about 9 hours ago • 441 • 13
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w8a16 Text Generation • Updated 10 days ago • 1.14k • 1
neuralmagic/Meta-Llama-3.1-70B-Instruct-quantized.w8a8 Text Generation • Updated 9 days ago • 23.5k • 6