GHOST 2.0: generative high-fidelity one shot transfer of heads Paper • 2502.18417 • Published 3 days ago • 57
LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers Paper • 2502.15007 • Published 8 days ago • 151
How Much Knowledge Can You Pack into a LoRA Adapter without Harming LLM? Paper • 2502.14502 • Published 8 days ago • 80
MMTEB: Massive Multilingual Text Embedding Benchmark Paper • 2502.13595 • Published 9 days ago • 31
view article Article Train 400x faster Static Embedding Models with Sentence Transformers Jan 15 • 151
MMTEB: Massive Multilingual Text Embedding Benchmark Paper • 2502.13595 • Published 9 days ago • 31