The current number 1 on the LLM leaderboard, only 34B, 200k context and now quantized with AWQ. This could very well be the optimal model in terms of model size, model accuracy and model speed.
· Sign up or log in to comment