"long sequence modeling" Papers

12 papers found

2DMamba: Efficient State Space Model for Image Representation with Applications on Giga-Pixel Whole Slide Image Classification

Jingwei Zhang, Anh Tien Nguyen, Xi Han et al.

CVPR 2025posterarXiv:2412.00678
20
citations

Dendritic Resonate-and-Fire Neuron for Effective and Efficient Long Sequence Modeling

Dehao Zhang, Malu Zhang, Shuai Wang et al.

NeurIPS 2025oralarXiv:2509.17186
2
citations

Training Free Exponential Context Extension via Cascading KV Cache

Jeff Willette, Heejun Lee, Youngwan Lee et al.

ICLR 2025posterarXiv:2406.17808
3
citations

Why RoPE Struggles to Maintain Long-Term Decay in Long Sequences?

Wei Shen, Chao Yin, Yuliang Liu et al.

ICLR 2025poster

ZETA: Leveraging $Z$-order Curves for Efficient Top-$k$ Attention

Qiuhao Zeng, Jierui Huang, Peng Lu et al.

ICLR 2025posterarXiv:2501.14577
5
citations

LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens

Yiran Ding, Li Lyna Zhang, Chengruidong Zhang et al.

ICML 2024poster

Motion Mamba: Efficient and Long Sequence Motion Generation

Zeyu Zhang, Akide Liu, Ian Reid et al.

ECCV 2024posterarXiv:2403.07487
108
citations

MuLTI: Efficient Video-and-Language Understanding with Text-Guided MultiWay-Sampler and Multiple Choice Modeling

Jiaqi Xu, Bo Liu, Yunkuo Chen et al.

AAAI 2024paperarXiv:2303.05707
2
citations

SeTformer Is What You Need for Vision and Language

Pourya Shamsolmoali, Masoumeh Zareapoor, Eric Granger et al.

AAAI 2024paperarXiv:2401.03540
7
citations

State-Free Inference of State-Space Models: The *Transfer Function* Approach

Rom N. Parnichkun, Stefano Massaroli, Alessandro Moro et al.

ICML 2024poster

Various Lengths, Constant Speed: Efficient Language Modeling with Lightning Attention

Zhen Qin, Weigao Sun, Dong Li et al.

ICML 2024poster

xT: Nested Tokenization for Larger Context in Large Images

Ritwik Gupta, Shufan Li, Tyler Zhu et al.

ICML 2024poster