2024 "attention mechanism" Papers

115 papers found • Page 1 of 3

A decoder-only foundation model for time-series forecasting

Abhimanyu Das, Weihao Kong, Rajat Sen et al.

ICML 2024oralarXiv:2310.10688

ADMap: Anti-disturbance Framework for Vectorized HD Map Construction

Haotian Hu, Fanyi Wang, Yaonong Wang et al.

ECCV 2024poster
5
citations

A Fixed-Point Approach for Causal Generative Modeling

Meyer Scetbon, Joel Jennings, Agrin Hilmkil et al.

ICML 2024posterarXiv:2404.06969

Agent Attention: On the Integration of Softmax and Linear Attention

Dongchen Han, Tianzhu Ye, Yizeng Han et al.

ECCV 2024posterarXiv:2312.08874
206
citations

Algorithm and Hardness for Dynamic Attention Maintenance in Large Language Models

Jan van den Brand, Zhao Song, Tianyi Zhou

ICML 2024posterarXiv:2304.02207

An Efficient and Effective Transformer Decoder-Based Framework for Multi-Task Visual Grounding

Wei Chen, Long Chen, Yu Wu

ECCV 2024posterarXiv:2408.01120
16
citations

An Image is Worth 1/2 Tokens After Layer 2: Plug-and-Play Inference Acceleration for Large Vision-Language Models

Liang Chen, Haozhe Zhao, Tianyu Liu et al.

ECCV 2024posterarXiv:2403.06764
343
citations

AnomalyDiffusion: Few-Shot Anomaly Image Generation with Diffusion Model

Teng Hu, Jiangning Zhang, Ran Yi et al.

AAAI 2024paperarXiv:2312.05767

Attention Disturbance and Dual-Path Constraint Network for Occluded Person Re-identification

Jiaer Xia, Lei Tan, Pingyang Dai et al.

AAAI 2024paperarXiv:2303.10976
24
citations

Attention Guided CAM: Visual Explanations of Vision Transformer Guided by Self-Attention

Saebom Leem, Hyunseok Seo

AAAI 2024paperarXiv:2402.04563
31
citations

Attention Meets Post-hoc Interpretability: A Mathematical Perspective

Gianluigi Lopardo, Frederic Precioso, Damien Garreau

ICML 2024posterarXiv:2402.03485

AttnLRP: Attention-Aware Layer-Wise Relevance Propagation for Transformers

Reduan Achtibat, Sayed Mohammad Vakilzadeh Hatefi, Maximilian Dreyer et al.

ICML 2024posterarXiv:2402.05602

AttNS: Attention-Inspired Numerical Solving For Limited Data Scenarios

Zhongzhan Huang, Mingfu Liang, Shanshan Zhong et al.

ICML 2024posterarXiv:2302.10184

BARET: Balanced Attention Based Real Image Editing Driven by Target-Text Inversion

Yuming Qiao, Fanyi Wang, Jingwen Su et al.

AAAI 2024paperarXiv:2312.05482

Bifurcated Attention for Single-Context Large-Batch Sampling

Ben Athiwaratkun, Sujan Kumar Gonugondla, Sanjay Krishna Gouda et al.

ICML 2024poster

Cached Transformers: Improving Transformers with Differentiable Memory Cached

Zhaoyang Zhang, Wenqi Shao, Yixiao Ge et al.

AAAI 2024paperarXiv:2312.12742
5
citations

CHAI: Clustered Head Attention for Efficient LLM Inference

Saurabh Agarwal, Bilge Acun, Basil Hosmer et al.

ICML 2024posterarXiv:2403.08058

Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance

Shenhao Zhu, Junming Chen, Zuozhuo Dai et al.

ECCV 2024posterarXiv:2403.14781
233
citations

Compositional Capabilities of Autoregressive Transformers: A Study on Synthetic, Interpretable Tasks

Rahul Ramesh, Ekdeep Singh Lubana, Mikail Khona et al.

ICML 2024posterarXiv:2311.12997

Correlation Matching Transformation Transformers for UHD Image Restoration

Cong Wang, Jinshan Pan, Wei Wang et al.

AAAI 2024paperarXiv:2406.00629
57
citations

CountFormer: Multi-View Crowd Counting Transformer

Hong Mo, Xiong Zhang, Jianchao Tan et al.

ECCV 2024posterarXiv:2407.02047
9
citations

Delving into Differentially Private Transformer

Youlong Ding, Xueyang Wu, Yining meng et al.

ICML 2024posterarXiv:2405.18194

DGR-MIL: Exploring Diverse Global Representation in Multiple Instance Learning for Whole Slide Image Classification

Wenhui Zhu, Xiwen Chen, Peijie Qiu et al.

ECCV 2024posterarXiv:2407.03575
24
citations

Do text-free diffusion models learn discriminative visual representations?

Soumik Mukhopadhyay, Matthew Gwilliam, Yosuke Yamaguchi et al.

ECCV 2024posterarXiv:2311.17921
26
citations

DreamMover: Leveraging the Prior of Diffusion Models for Image Interpolation with Large Motion

Liao Shen, Tianqi Liu, Huiqiang Sun et al.

ECCV 2024posterarXiv:2409.09605
13
citations

Dynamic Feature Pruning and Consolidation for Occluded Person Re-identification

YuTeng Ye, Hang Zhou, Jiale Cai et al.

AAAI 2024paperarXiv:2211.14742
14
citations

Enhancing Semantic Fidelity in Text-to-Image Synthesis: Attention Regulation in Diffusion Models

Yang Zhang, Tze Tzun Teoh, Wei Hern Lim et al.

ECCV 2024posterarXiv:2403.06381
19
citations

EquiAV: Leveraging Equivariance for Audio-Visual Contrastive Learning

Jongsuk Kim, Hyeongkeun Lee, Kyeongha Rho et al.

ICML 2024posterarXiv:2403.09502

Eureka-Moments in Transformers: Multi-Step Tasks Reveal Softmax Induced Optimization Problems

David T. Hoffmann, Simon Schrodi, Jelena Bratulić et al.

ICML 2024posterarXiv:2310.12956

Exploring Reliable Matching with Phase Enhancement for Night-time Semantic Segmentation

Yuwen Pan, Rui Sun, Naisong Luo et al.

ECCV 2024posterarXiv:2408.13838
5
citations

FaceCoresetNet: Differentiable Coresets for Face Set Recognition

Gil Shapira, Yosi Keller

AAAI 2024paperarXiv:2308.14075
3
citations

FALIP: Visual Prompt as Foveal Attention Boosts CLIP Zero-Shot Performance

Jiedong Zhuang, Jiaqi Hu, Lianrui Mu et al.

ECCV 2024posterarXiv:2407.05578
7
citations

Free-Editor: Zero-shot Text-driven 3D Scene Editing

Md Nazmul Karim, Hasan Iqbal, Umar Khalid et al.

ECCV 2024posterarXiv:2312.13663
14
citations

Gated Attention Coding for Training High-Performance and Efficient Spiking Neural Networks

Xuerui Qiu, Rui-Jie Zhu, Yuhong Chou et al.

AAAI 2024paperarXiv:2308.06582

GaussianFormer: Scene as Gaussians for Vision-Based 3D Semantic Occupancy Prediction

Yuanhui Huang, Wenzhao Zheng, Yunpeng Zhang et al.

ECCV 2024posterarXiv:2405.17429
95
citations

Generative Enzyme Design Guided by Functionally Important Sites and Small-Molecule Substrates

Zhenqiao Song, Yunlong Zhao, Wenxian Shi et al.

ICML 2024posterarXiv:2405.08205

Graph-based Forecasting with Missing Data through Spatiotemporal Downsampling

Ivan Marisca, Cesare Alippi, Filippo Maria Bianchi

ICML 2024oralarXiv:2402.10634

Graph Context Transformation Learning for Progressive Correspondence Pruning

Junwen Guo, Guobao Xiao, Shiping Wang et al.

AAAI 2024paperarXiv:2312.15971
8
citations

Graph External Attention Enhanced Transformer

Jianqing Liang, Min Chen, Jiye Liang

ICML 2024posterarXiv:2405.21061

Grid-Attention: Enhancing Computational Efficiency of Large Vision Models without Fine-Tuning

Pengyu Li, Biao Wang, Tianchu Guo et al.

ECCV 2024poster

GridFormer: Point-Grid Transformer for Surface Reconstruction

Shengtao Li, Ge Gao, Yudong Liu et al.

AAAI 2024paperarXiv:2401.02292

HENet: Hybrid Encoding for End-to-end Multi-task 3D Perception from Multi-view Cameras

Zhongyu Xia, ZhiWei Lin, Xinhao Wang et al.

ECCV 2024posterarXiv:2404.02517
19
citations

Heterogeneous Graph Reasoning for Fact Checking over Texts and Tables

Haisong Gong, Weizhi Xu, Shu Wu et al.

AAAI 2024paperarXiv:2402.13028
16
citations

Hierarchical Aligned Multimodal Learning for NER on Tweet Posts

Peipei Liu, Hong Li, Yimo Ren et al.

AAAI 2024paperarXiv:2305.08372
8
citations

High-Order Contrastive Learning with Fine-grained Comparative Levels for Sparse Ordinal Tensor Completion

Yu Dai, Junchen Shen, Zijie Zhai et al.

ICML 2024poster

How Smooth Is Attention?

Valérie Castin, Pierre Ablin, Gabriel Peyré

ICML 2024posterarXiv:2312.14820

How to Protect Copyright Data in Optimization of Large Language Models?

Timothy Chu, Zhao Song, Chiwun Yang

AAAI 2024paperarXiv:2308.12247

How Transformers Learn Causal Structure with Gradient Descent

Eshaan Nichani, Alex Damian, Jason Lee

ICML 2024posterarXiv:2402.14735

IIANet: An Intra- and Inter-Modality Attention Network for Audio-Visual Speech Separation

Kai Li, Runxuan Yang, Fuchun Sun et al.

ICML 2024oralarXiv:2308.08143

In-context Convergence of Transformers

Yu Huang, Yuan Cheng, Yingbin LIANG

ICML 2024posterarXiv:2310.05249
← PreviousNext →