-
Sequence Parallelism: Long Sequence Training from System Perspective
Paper • 2105.13120 • Published • 5 -
Ring Attention with Blockwise Transformers for Near-Infinite Context
Paper • 2310.01889 • Published • 11 -
Striped Attention: Faster Ring Attention for Causal Transformers
Paper • 2311.09431 • Published • 4 -
DeepSpeed Ulysses: System Optimizations for Enabling Training of Extreme Long Sequence Transformer Models
Paper • 2309.14509 • Published • 18
Maozhou Ge
Gmc2
AI & ML interests
None yet
Recent Activity
liked
a Space
about 7 hours ago
nanotron/ultrascale-playbook
upvoted
an
article
10 days ago
Open R1: Update #2
upvoted
an
article
13 days ago
Open-source DeepResearch – Freeing our search agents
Organizations
None yet
Collections
1
models
None public yet
datasets
None public yet