Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
QuantFactory
/
Meta-Llama-3-8B-Instruct-GGUF-v2
like
15
Follow
Quant Factory
225
Text Generation
GGUF
PyTorch
English
facebook
meta
llama
llama-3
Inference Endpoints
conversational
License:
llama3
Model card
Files
Files and versions
Community
2
Deploy
Use this model
New discussion
New pull request
Resources
PR & discussions documentation
Code of Conduct
Hub documentation
All
Discussions
Pull requests
View closed (0)
Method for loading with transformers without dequantizing
#2 opened 3 months ago by
ddphys
I'm experiencing the same endless loop with this version as well
#1 opened 6 months ago by
alexcardo