HypeNet Collection The models for the paper: Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts • 1 item • Updated 1 day ago
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts Paper • 2601.22156 • Published Jan 29 • 14
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts Paper • 2601.22156 • Published Jan 29 • 14
Hybrid Linear Attention Done Right: Efficient Distillation and Effective Architectures for Extremely Long Contexts Paper • 2601.22156 • Published Jan 29 • 14
Kimi Linear: An Expressive, Efficient Attention Architecture Paper • 2510.26692 • Published Oct 30, 2025 • 132
StateX: Enhancing RNN Recall via Post-training State Expansion Paper • 2509.22630 • Published Sep 26, 2025 • 4
StateX: Enhancing RNN Recall via Post-training State Expansion Paper • 2509.22630 • Published Sep 26, 2025 • 4
StateX: Enhancing RNN Recall via Post-training State Expansion Paper • 2509.22630 • Published Sep 26, 2025 • 4 • 2