"long sequence modeling" Papers
12 papers found
2DMamba: Efficient State Space Model for Image Representation with Applications on Giga-Pixel Whole Slide Image Classification
Jingwei Zhang, Anh Tien Nguyen, Xi Han et al.
CVPR 2025posterarXiv:2412.00678
20
citations
Dendritic Resonate-and-Fire Neuron for Effective and Efficient Long Sequence Modeling
Dehao Zhang, Malu Zhang, Shuai Wang et al.
NeurIPS 2025oralarXiv:2509.17186
2
citations
Training Free Exponential Context Extension via Cascading KV Cache
Jeff Willette, Heejun Lee, Youngwan Lee et al.
ICLR 2025posterarXiv:2406.17808
3
citations
Why RoPE Struggles to Maintain Long-Term Decay in Long Sequences?
Wei Shen, Chao Yin, Yuliang Liu et al.
ICLR 2025poster
ZETA: Leveraging $Z$-order Curves for Efficient Top-$k$ Attention
Qiuhao Zeng, Jierui Huang, Peng Lu et al.
ICLR 2025posterarXiv:2501.14577
5
citations
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Yiran Ding, Li Lyna Zhang, Chengruidong Zhang et al.
ICML 2024poster
Motion Mamba: Efficient and Long Sequence Motion Generation
Zeyu Zhang, Akide Liu, Ian Reid et al.
ECCV 2024posterarXiv:2403.07487
108
citations
MuLTI: Efficient Video-and-Language Understanding with Text-Guided MultiWay-Sampler and Multiple Choice Modeling
Jiaqi Xu, Bo Liu, Yunkuo Chen et al.
AAAI 2024paperarXiv:2303.05707
2
citations
SeTformer Is What You Need for Vision and Language
Pourya Shamsolmoali, Masoumeh Zareapoor, Eric Granger et al.
AAAI 2024paperarXiv:2401.03540
7
citations
State-Free Inference of State-Space Models: The *Transfer Function* Approach
Rom N. Parnichkun, Stefano Massaroli, Alessandro Moro et al.
ICML 2024poster
Various Lengths, Constant Speed: Efficient Language Modeling with Lightning Attention
Zhen Qin, Weigao Sun, Dong Li et al.
ICML 2024poster
xT: Nested Tokenization for Larger Context in Large Images
Ritwik Gupta, Shufan Li, Tyler Zhu et al.
ICML 2024poster