Hugging Face
Models
Datasets
Spaces
Posts
Docs
Enterprise
Pricing
Log In
Sign Up
75.2
TFLOPS
7
4
62
Benhao Tang
PRO
benhaotang
Follow
Khetterman's profile picture
Mi6paulino's profile picture
John6666's profile picture
3 followers
·
11 following
benhaotang
benhaotang.bsky.social
AI & ML interests
Physics Master student in theoretical particle physics at Universität Heidelberg, actively looking into the possibilities of integrating AI into future physics research.
Recent Activity
reacted
to
mitkox
's
post
with 👍
about 2 hours ago
llama.cpp is 26.8% faster than ollama. I have upgraded both, and using the same settings, I am running the same DeepSeek R1 Distill 1.5B on the same hardware. It's an Apples to Apples comparison. Total duration: llama.cpp 6.85 sec <- 26.8% faster ollama 8.69 sec Breakdown by phase: Model loading llama.cpp 241 ms <- 2x faster ollama 553 ms Prompt processing llama.cpp 416.04 tokens/s with an eval time 45.67 ms <- 10x faster ollama 42.17 tokens/s with an eval time of 498 ms Token generation llama.cpp 137.79 tokens/s with an eval time 6.62 sec <- 13% faster ollama 122.07 tokens/s with an eval time 7.64 sec llama.cpp is LLM inference in C/C++; ollama adds abstraction layers and marketing. Make sure you own your AI. AI in the cloud is not aligned with you; it's aligned with the company that owns it.
reacted
to
sometimesanotion
's
post
with 🔥
1 day ago
I've managed a #1 score of 41.22% average for 14B parameter models on the Open LLM Leaderboard. As of this writing, sometimesanotion/Lamarck-14B-v0.7 is #8 for all models up to 70B parameters. It took a custom toolchain around Arcee AI's mergekit to manage the complex merges, gradients, and LoRAs required to make this happen. I really like seeing features of many quality finetunes in one solid generalist model.
liked
a model
3 days ago
Lyte/Titans-MAC-test
View all activity
Organizations
None yet
spaces
2
Sort: Recently updated
Running
on
Zero
3
💬
Try Out phi4-qwq-sky-t1
Sleeping
1
💬
Try out Mistral-Small-Physics
models
10
Sort: Recently updated
benhaotang/phi4-qwq-sky-t1
Text Generation
•
Updated
3 days ago
•
44
•
1
benhaotang/phi4-qwq-sky-t1-Q4_K_M-GGUF
Updated
7 days ago
•
80
•
1
benhaotang/Phi-4-llama-t1-full
Updated
8 days ago
•
17
•
1
benhaotang/Phi-4-llama-t1-full-Q4_K_M-GGUF
Updated
8 days ago
•
34
benhaotang/mistral-small-physics-finetuned-adapter
Updated
Nov 16, 2024
•
22
•
1
benhaotang/mistral-small-physics-finetuned-bnb-4bit
Updated
Nov 16, 2024
•
11
benhaotang/Rombos-Coder-V2.5-Qwen-7b-GGUF_cline
Updated
Nov 15, 2024
•
585
•
6
benhaotang/llama3.2-1B-physics-finetuned
Text Generation
•
Updated
Nov 15, 2024
•
8
benhaotang/nemo-math-science-philosophy-12B-Q8_0-GGUF
Updated
Oct 16, 2024
•
56
benhaotang/nemo-math-science-philosophy-12B
Text Generation
•
Updated
Oct 16, 2024
•
65
•
3
datasets
None public yet