Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts Paper • 2503.05447 • Published 6 days ago • 7
Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts Paper • 2503.05447 • Published 6 days ago • 7
Liger: Linearizing Large Language Models to Gated Recurrent Structures Paper • 2503.01496 • Published 10 days ago • 15
Liger: Linearizing Large Language Models to Gated Recurrent Structures Paper • 2503.01496 • Published 10 days ago • 15
MoM: Linear Sequence Modeling with Mixture-of-Memories Paper • 2502.13685 • Published 22 days ago • 33
MoM: Linear Sequence Modeling with Mixture-of-Memories Paper • 2502.13685 • Published 22 days ago • 33
LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid Paper • 2502.07563 • Published about 1 month ago • 24
LASP-2: Rethinking Sequence Parallelism for Linear Attention and Its Hybrid Paper • 2502.07563 • Published about 1 month ago • 24
Test-Time Preference Optimization: On-the-Fly Alignment via Iterative Textual Feedback Paper • 2501.12895 • Published Jan 22 • 57