LLM Model VRAM Calculator
Calculate VRAM requirements for running large language models
GGUF, Prompt gen, Repo tools, followed by "Bench" and "Leaderboards". Leader boards get more specific going down. See also: "Run LLMs ..." collection.
Calculate VRAM requirements for running large language models
Refine your prompts
Request custom GGML quantized models via email
Edit GGUF metadata on Hugging Face
Convert your PEFT LoRA into GGUF
Copy a Hugging Face repository
Convert and upload models to MLX format on Hugging Face
View EQ-Bench Leaderboard for LLMs
Note A specialized Bench for evaluation of the creativity of a model with testing outputs shown as well as judgements / ratings including a model's "emotional intelligence".
Display a leaderboard with UGI scores
Note Uncensored General Intelligence. Another great source for creative and/or role play models.
Track, rank and evaluate open LLMs and chatbots
Compare model answers to questions
View LLM Performance Leaderboard
Compare Open LLM Leaderboard results
Run a Streamlit web app
Display chatbot leaderboard and statistics
Display and filter leaderboard results for LLM judges
Explore and analyze RewardBench leaderboard data
Select benchmarks and languages for text embeddings evaluation
A leaderboard for multimodal models
Explore LLM performance across hardware
Track, rank and evaluate open LLMs' CoT quality
Visualize model performance with interactive plots and tables
Track, rank and evaluate open LLMs and chatbots
Ranking for Open-sourced LLMs in different domains
More advanced and challenging multi-task evaluation
Blind vote on HF TTS models!
Explore and analyze code evaluation data
Submit code models for evaluation on benchmarks
Generate images from text descriptions
Browse and compare speech recognition model performance
Request evaluation of a speech recognition model
Display OCRBench leaderboard for model evaluations
Leaderboard and arena of Video Generation models
VLMEvalKit Evaluation Results Collection
Submit model evaluation and view leaderboard
Browse and submit LLM evaluations
Explore and filter language model benchmark results
Browse Q-Bench leaderboard for vision model performance
Display and filter LLM benchmark results
Browse and filter LLM benchmark results
View and submit LLM evaluations
Explore benchmark results for QA and long doc models
View and submit machine learning model evaluations
Browse and submit model evaluations in LLM benchmarks