Native Hybrid Attention for Efficient Sequence Modeling Paper • 2510.07019 • Published Oct 8, 2025 • 16 • 2
Speed Always Wins: A Survey on Efficient Architectures for Large Language Models Paper • 2508.09834 • Published Aug 13, 2025 • 53 • 2
Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts Paper • 2503.05447 • Published Mar 7, 2025 • 8 • 2
Liger: Linearizing Large Language Models to Gated Recurrent Structures Paper • 2503.01496 • Published Mar 3, 2025 • 18 • 2
MoM: Linear Sequence Modeling with Mixture-of-Memories Paper • 2502.13685 • Published Feb 19, 2025 • 36 • 2
LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid Paper • 2502.07563 • Published Feb 11, 2025 • 23 • 2