RAGServe: Fast Quality-Aware RAG Systems with Configuration Adaptation Paper • 2412.10543 • Published Dec 13, 2024
SSMs Collection A collection of Mamba-2-based research models with 8B parameters trained on 3.5T tokens for comparison with Transformers. • 5 items • Updated 9 days ago • 27