Most Cited ICML "fmri signal decoding" Papers
5,975 papers found • Page 9 of 30
Conference
Interpreting the Repeated Token Phenomenon in Large Language Models
Itay Yona, Ilia Shumailov, Jamie Hayes et al.
Self-Composing Policies for Scalable Continual Reinforcement Learning
Mikel Malagón, Josu Ceberio, Jose A Lozano
Meta-Learners for Partially-Identified Treatment Effects Across Multiple Environments
Jonas Schweisthal, Dennis Frauen, M van der Schaar et al.
On Expressive Power of Looped Transformers: Theoretical Analysis and Enhancement via Timestep Encoding
Kevin Xu, Issei Sato
On the Calibration of Human Pose Estimation
Kerui Gu, Rongyu Chen, Xuanlong Yu et al.
A Closer Look at Multimodal Representation Collapse
Abhra Chaudhuri, Anjan Dutta, Tu Bui et al.
A Mixture-Based Framework for Guiding Diffusion Models
Yazid Janati, Badr MOUFAD, Mehdi Qassime et al.
Contrastive Representation for Data Filtering in Cross-Domain Offline Reinforcement Learning
Xiaoyu Wen, Chenjia Bai, Kang Xu et al.
Individual Contributions as Intrinsic Exploration Scaffolds for Multi-agent Reinforcement Learning
Xinran Li, Zifan LIU, Shibo Chen et al.
SOLD: Slot Object-Centric Latent Dynamics Models for Relational Manipulation Learning from Pixels
Malte Mosbach, Jan Ewertz, Angel Villar-Corrales et al.
R.I.P.: Better Models by Survival of the Fittest Prompts
Ping Yu, Weizhe Yuan, Olga Golovneva et al.
PARCv2: Physics-aware Recurrent Convolutional Neural Networks for Spatiotemporal Dynamics Modeling
Phong Nguyen, Xinlun Cheng, Shahab Azarfar et al.
RocketKV: Accelerating Long-Context LLM Inference via Two-Stage KV Cache Compression
Payman Behnam, Yaosheng Fu, Ritchie Zhao et al.
Libra: Building Decoupled Vision System on Large Language Models
Yifan Xu, Xiaoshan Yang, Yaguang Song et al.
Synthesizing Software Engineering Data in a Test-Driven Manner
Lei Zhang, Jiaxi Yang, Min Yang et al.
Reasoning Through Execution: Unifying Process and Outcome Rewards for Code Generation
Zhuohao Yu, Weizheng Gu, Yidong Wang et al.
Sample as you Infer: Predictive Coding with Langevin Dynamics
Umais Zahid, Qinghai Guo, Zafeirios Fountas
Oscillation-Reduced MXFP4 Training for Vision Transformers
Yuxiang Chen, Haocheng Xi, Jun Zhu et al.
When Do LLMs Help With Node Classification? A Comprehensive Analysis
Xixi Wu, Yifei Shen, Fangzhou Ge et al.
Fundamental limits of learning in sequence multi-index models and deep attention networks: high-dimensional asymptotics and sharp thresholds
Emanuele Troiani, Hugo Cui, Yatin Dandi et al.
Generalizing Knowledge Graph Embedding with Universal Orthogonal Parameterization
Rui Li, Chaozhuo Li, Yanming Shen et al.
Constrain Alignment with Sparse Autoencoders
Qingyu Yin, Chak Tou Leong, Hongbo Zhang et al.
Temperature-Annealed Boltzmann Generators
Henrik Schopmans, Pascal Friederich
Solving Poisson Equations using Neural Walk-on-Spheres
Hong Chul Nam, Julius Berner, Anima Anandkumar
How Much Can Transfer? BRIDGE: Bounded Multi-Domain Graph Foundation Model with Generalization Guarantees
Haonan Yuan, Qingyun Sun, Junhua Shi et al.
Discovering Mixtures of Structural Causal Models from Time Series Data
Sumanth Varambally, Yian Ma, Rose Yu
GRU: Mitigating the Trade-off between Unlearning and Retention for LLMs
Yue Wang, Qizhou Wang, Feng Liu et al.
CuTS: Customizable Tabular Synthetic Data Generation
Mark Vero, Mislav Balunovic, Martin Vechev
SEFE: Superficial and Essential Forgetting Eliminator for Multimodal Continual Instruction Tuning
Jinpeng Chen, Runmin Cong, Yuzhi Zhao et al.
Stereographic Spherical Sliced Wasserstein Distances
Huy Tran, Yikun Bai, Abihith Kothapalli et al.
Hybrid$^2$ Neural ODE Causal Modeling and an Application to Glycemic Response
Junyi Zou, Matthew Levine, Dessi Zaharieva et al.
Short-Long Convolutions Help Hardware-Efficient Linear Attention to Focus on Long Sequences
Zicheng Liu, Siyuan Li, Li Wang et al.
A Distributional Analogue to the Successor Representation
Harley Wiltzer, Jesse Farebrother, Arthur Gretton et al.
Improving Transformer World Models for Data-Efficient RL
Antoine Dedieu, Joseph Ortiz, Xinghua Lou et al.
Unifying 2D and 3D Vision-Language Understanding
Ayush Jain, Alexander Swerdlow, Yuzhou Wang et al.
Understanding Inter-Concept Relationships in Concept-Based Models
Naveen Raman, Mateo Espinosa Zarlenga, Mateja Jamnik
Fair Off-Policy Learning from Observational Data
Dennis Frauen, Valentyn Melnychuk, Stefan Feuerriegel
OTMatch: Improving Semi-Supervised Learning with Optimal Transport
Zhiquan Tan, Kaipeng Zheng, Weiran Huang
More Flexible PAC-Bayesian Meta-Learning by Learning Learning Algorithms
Hossein Zakerinia, Amin Behjati, Christoph Lampert
Simple Policy Optimization
Zhengpeng Xie, Qiang Zhang, Fan Yang et al.
Quantum Positional Encodings for Graph Neural Networks
Slimane Thabet, Mehdi Djellabi, Igor Sokolov et al.
Exploring the Low-Pass Filtering Behavior in Image Super-Resolution
Haoyu Deng, Zijing Xu, Yule Duan et al.
CRoFT: Robust Fine-Tuning with Concurrent Optimization for OOD Generalization and Open-Set OOD Detection
Lin Zhu, Yifeng Yang, Qinying Gu et al.
Translation Equivariant Transformer Neural Processes
Matthew Ashman, Cristiana Diaconu, Junhyuck Kim et al.
MONA: Myopic Optimization with Non-myopic Approval Can Mitigate Multi-step Reward Hacking
Sebastian Farquhar, Vikrant Varma, David Lindner et al.
Stochastic Control for Fine-tuning Diffusion Models: Optimality, Regularity, and Convergence
Yinbin Han, Meisam Razaviyayn, Renyuan Xu
Global Reinforcement Learning : Beyond Linear and Convex Rewards via Submodular Semi-gradient Methods
Riccardo De Santi, Manish Prajapat, Andreas Krause
Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment
Yifan Zhang, Ge Zhang, Yue Wu et al.
PlaySlot: Learning Inverse Latent Dynamics for Controllable Object-Centric Video Prediction and Planning
Angel Villar-Corrales, Sven Behnke
Multi-Session Budget Optimization for Forward Auction-based Federated Learning
Xiaoli Tang, Han Yu, Zengxiang Li et al.
How Transformers Learn Structured Data: Insights From Hierarchical Filtering
Jerome Garnier-Brun, Marc Mezard, Emanuele Moscato et al.
DAMA: Data- and Model-aware Alignment of Multi-modal LLMs
Jinda Lu, Junkang Wu, Jinghan Li et al.
Unisolver: PDE-Conditional Transformers Towards Universal Neural PDE Solvers
Hang Zhou, Yuezhou Ma, Haixu Wu et al.
Generative Enzyme Design Guided by Functionally Important Sites and Small-Molecule Substrates
Zhenqiao Song, Yunlong Zhao, Wenxian Shi et al.
A2Q+: Improving Accumulator-Aware Weight Quantization
Ian Colbert, Alessandro Pappalardo, Jakoba Petri-Koenig et al.
Helpful or Harmful Data? Fine-tuning-free Shapley Attribution for Explaining Language Model Predictions
Jingtan Wang, Xiaoqiang Lin, Rui Qiao et al.
Scalable Real-Time Recurrent Learning Using Columnar-Constructive Networks
Khurram Javed, Haseeb Shah, Richard Sutton et al.
HarmoniCa: Harmonizing Training and Inference for Better Feature Caching in Diffusion Transformer Acceleration
Yushi Huang, Zining Wang, Ruihao Gong et al.
Tool Unlearning for Tool-Augmented LLMs
Jiali Cheng, Hadi Amiri
Creative Text-to-Audio Generation via Synthesizer Programming
Manuel Cherep, Nikhil Singh, Jessica Shand
Improving Parallel Program Performance with LLM Optimizers via Agent-System Interfaces
Anjiang Wei, Allen Nie, Thiago Teixeira et al.
RICE: Breaking Through the Training Bottlenecks of Reinforcement Learning with Explanation
Zelei Cheng, Xian Wu, Jiahao Yu et al.
Towards Understanding Fine-Tuning Mechanisms of LLMs via Circuit Analysis
Xu Wang, Yan Hu, Wenyu Du et al.
On Measuring Long-Range Interactions in Graph Neural Networks
Jacob Bamberger, Benjamin Gutteridge, Scott le Roux et al.
Bounded and Uniform Energy-based Out-of-distribution Detection for Graphs
Shenzhi Yang, Bin Liang, An Liu et al.
BadPart: Unified Black-box Adversarial Patch Attacks against Pixel-wise Regression Tasks
Zhiyuan Cheng, Zhaoyi Liu, Tengda Guo et al.
FightLadder: A Benchmark for Competitive Multi-Agent Reinforcement Learning
Wenzhe Li, Zihan Ding, Seth Karten et al.
Reducing sequential change detection to sequential estimation
Shubhanshu Shekhar, Aaditya Ramdas
Revealing the Dark Secrets of Extremely Large Kernel ConvNets on Robustness
Honghao Chen, Zhang Yurong, xiaokun Feng et al.
Position: An Inner Interpretability Framework for AI Inspired by Lessons from Cognitive Neuroscience
Martina G. Vilas, Federico Adolfi, David Poeppel et al.
ReQFlow: Rectified Quaternion Flow for Efficient and High-Quality Protein Backbone Generation
Angxiao Yue, Zichong Wang, Hongteng Xu
Premier-TACO is a Few-Shot Policy Learner: Pretraining Multitask Representation via Temporal Action-Driven Contrastive Loss
Ruijie Zheng, Yongyuan Liang, xiyao wang et al.
Objective drives the consistency of representational similarity across datasets
Laure Ciernik, Lorenz Linhardt, Marco Morik et al.
Tandem Transformers for Inference Efficient LLMs
Aishwarya P S, Pranav Nair, Yashas Samaga et al.
Improved Off-policy Reinforcement Learning in Biological Sequence Design
Hyeonah Kim, Minsu Kim, Taeyoung Yun et al.
Lessons from Generalization Error Analysis of Federated Learning: You May Communicate Less Often!
Milad Sefidgaran, Romain Chor, Abdellatif Zaidi et al.
Feature Distribution on Graph Topology Mediates the Effect of Graph Convolution: Homophily Perspective
Soo Yong Lee, Sunwoo Kim, Fanchen Bu et al.
Efficient Pareto Manifold Learning with Low-Rank Structure
Weiyu CHEN, James Kwok
Leveraging Self-Consistency for Data-Efficient Amortized Bayesian Inference
Marvin Schmitt, Desi Ivanova, Daniel Habermann et al.
Compressed Image Generation with Denoising Diffusion Codebook Models
Guy Ohayon, Hila Manor, Tomer Michaeli et al.
Gaussian Mixture Flow Matching Models
Hansheng Chen, Kai Zhang, Hao Tan et al.
Fair Risk Control: A Generalized Framework for Calibrating Multi-group Fairness Risks
Lujing Zhang, Aaron Roth, Linjun Zhang
Safety Reasoning with Guidelines
Haoyu Wang, Zeyu Qin, Li Shen et al.
One-Shot Strategic Classification Under Unknown Costs
Elan Rosenfeld, Nir Rosenfeld
Mean Estimation in the Add-Remove Model of Differential Privacy
Alex Kulesza, Ananda Suresh, Yuyan Wang
Estimating Barycenters of Distributions with Neural Optimal Transport
Alexander Kolesov, Petr Mokrov, Igor Udovichenko et al.
AdsorbDiff: Adsorbate Placement via Conditional Denoising Diffusion
Adeesh Kolluru, John Kitchin
Nonparametric Teaching of Implicit Neural Representations
Chen Zhang, Steven T. S. Luo, Jason Chun Lok Li et al.
Multiplicative Weights Update, Area Convexity and Random Coordinate Descent for Densest Subgraph Problems
Ta Duy Nguyen, Alina Ene
MA-LoT: Model-Collaboration Lean-based Long Chain-of-Thought Reasoning enhances Formal Theorem Proving
Ruida Wang, Rui Pan, Yuxin Li et al.
Private Heterogeneous Federated Learning Without a Trusted Server Revisited: Error-Optimal and Communication-Efficient Algorithms for Convex Losses
Changyu Gao, Andrew Lowy, Xingyu Zhou et al.
How Uniform Random Weights Induce Non-uniform Bias: Typical Interpolating Neural Networks Generalize with Narrow Teachers
Gon Buzaglo, Itamar Harel, Mor Shpigel Nacson et al.
Weisfeiler Leman for Euclidean Equivariant Machine Learning
Snir Hordan, Tal Amir, Nadav Dym
Zebra: In-Context Generative Pretraining for Solving Parametric PDEs
Louis Serrano, Armand Kassaï Koupaï, Thomas Wang et al.
The Role of Learning Algorithms in Collective Action
Omri Ben-Dov, Jake Fawkes, Samira Samadi et al.
Learning to Predict Mutational Effects of Protein-Protein Interactions by Microenvironment-aware Hierarchical Prompt Learning
Lirong Wu, Yijun Tian, Haitao Lin et al.
To Cool or not to Cool? Temperature Network Meets Large Foundation Models via DRO
Zi-Hao Qiu, Siqi Guo, Mao Xu et al.
Non-Asymptotic Analysis for Single-Loop (Natural) Actor-Critic with Compatible Function Approximation
Yudan Wang, Yue Wang, Yi Zhou et al.
Cross-view Masked Diffusion Transformers for Person Image Synthesis
Trung Pham, Kang Zhang, Chang Yoo
Latent Logic Tree Extraction for Event Sequence Explanation from LLMs
Zitao Song, Chao Yang, Chaojie Wang et al.
Beyond the Calibration Point: Mechanism Comparison in Differential Privacy
Georgios Kaissis, Stefan Kolek, Borja de Balle Pigem et al.
Towards Understanding Inductive Bias in Transformers: A View From Infinity
Itay Lavie, Guy Gur-Ari, Zohar Ringel
Scaling Large Motion Models with Million-Level Human Motions
Ye Wang, Sipeng Zheng, Bin Cao et al.
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts
Mohammed Nowaz Rabbani Chowdhury, Meng Wang, Kaoutar El Maghraoui et al.
QuantSpec: Self-Speculative Decoding with Hierarchical Quantized KV Cache
Rishabh Tiwari, Haocheng Xi, Aditya Tomar et al.
Probabilistic Conceptual Explainers: Trustworthy Conceptual Explanations for Vision Foundation Models
Hengyi Wang, Shiwei Tan, Hao Wang
On Temperature Scaling and Conformal Prediction of Deep Classifiers
Lahav Dabah, Tom Tirer
Scalable Online Exploration via Coverability
Philip Amortila, Dylan Foster, Akshay Krishnamurthy
An Analysis for Reasoning Bias of Language Models with Small Initialization
Junjie Yao, zhongwang zhang, Zhi-Qin John Xu
Sortformer: A Novel Approach for Permutation-Resolved Speaker Supervision in Speech-to-Text Systems
Taejin Park, Ivan Medennikov, Kunal Dhawan et al.
TabFlex: Scaling Tabular Learning to Millions with Linear Attention
Yuchen Zeng, Tuan Dinh, Wonjun Kang et al.
Position: Stop Making Unscientific AGI Performance Claims
Patrick Altmeyer, Andrew Demetriou, Antony Bartlett et al.
STAR: Learning Diverse Robot Skill Abstractions through Rotation-Augmented Vector Quantization
Hao Li, Qi Lv, Rui Shao et al.
LARM: Large Auto-Regressive Model for Long-Horizon Embodied Intelligence
Zhuoling Li, Xiaogang Xu, Zhenhua Xu et al.
SEMU: Singular Value Decomposition for Efficient Machine Unlearning
Marcin Sendera, Łukasz Struski, Kamil Książek et al.
When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models
Haoran You, Yichao Fu, Zheng Wang et al.
Data-Juicer Sandbox: A Feedback-Driven Suite for Multimodal Data-Model Co-development
Daoyuan Chen, Haibin Wang, Yilun Huang et al.
The Elicitation Game: Evaluating Capability Elicitation Techniques
Felix Hofstätter, Teun van der Weij, Jayden Teoh et al.
Adaptively Perturbed Mirror Descent for Learning in Games
Kenshi Abe, Kaito Ariu, Mitsuki Sakamoto et al.
ArtWhisperer: A Dataset for Characterizing Human-AI Interactions in Artistic Creations
Kailas Vodrahalli, James Zou
AD3: Implicit Action is the Key for World Models to Distinguish the Diverse Visual Distractors
Yucen Wang, Shenghua Wan, Le Gan et al.
Unlocking the Power of Spatial and Temporal Information in Medical Multimodal Pre-training
Jinxia Yang, Bing Su, Xin Zhao et al.
An Online Optimization Perspective on First-Order and Zero-Order Decentralized Nonsmooth Nonconvex Stochastic Optimization
Emre Sahinoglu, Shahin Shahrampour
Synthesizing Privacy-Preserving Text Data via Finetuning *without* Finetuning Billion-Scale LLMs
Bowen Tan, Zheng Xu, Eric Xing et al.
Lightweight Dataset Pruning without Full Training via Example Difficulty and Prediction Uncertainty
Yeseul Cho, Baekrok Shin, Changmin Kang et al.
LoRA Training Provably Converges to a Low-Rank Global Minimum Or It Fails Loudly (But it Probably Won't Fail)
Junsu Kim, Jaeyeon Kim, Ernest Ryu
Leveraging VLM-Based Pipelines to Annotate 3D Objects
Rishabh Kabra, Loic Matthey, Alexander Lerchner et al.
Textual Unlearning Gives a False Sense of Unlearning
Jiacheng Du, Zhibo Wang, Jie Zhang et al.
A New Robust Partial p-Wasserstein-Based Metric for Comparing Distributions
Sharath Raghvendra, Pouyan Shirzadian, Kaiyi Zhang
Task Generalization with Autoregressive Compositional Structure: Can Learning from $D$ Tasks Generalize to $D^T$ Tasks?
Amirhesam Abedsoltan, Huaqing Zhang, Kaiyue Wen et al.
Effective and Efficient Masked Image Generation Models
Zebin You, Jingyang Ou, Xiaolu Zhang et al.
PEARL: Zero-shot Cross-task Preference Alignment and Robust Reward Learning for Robotic Manipulation
Runze Liu, Yali Du, Fengshuo Bai et al.
Selective Mixup Helps with Distribution Shifts, But Not (Only) because of Mixup
Damien Teney, Jindong Wang, Ehsan Abbasnejad
Compositional Risk Minimization
Divyat Mahajan, Mohammad Pezeshki, Charles Arnal et al.
Position: Tensor Networks are a Valuable Asset for Green AI
Eva Memmel, Clara Menzen, Jetze Schuurmans et al.
Improving LLMs for Recommendation with Out-Of-Vocabulary Tokens
Ting-Ji Huang, Jia-Qi Yang, Chunxu Shen et al.
Density-Softmax: Efficient Test-time Model for Uncertainty Estimation and Robustness under Distribution Shifts
Ha Manh Bui, Anqi Liu
Prototypical Transformer As Unified Motion Learners
Cheng Han, Yawen Lu, Guohao Sun et al.
Privacy Attacks in Decentralized Learning
Abdellah El Mrini, Edwige Cyffers, Aurélien Bellet
Toward a Unified Theory of Gradient Descent under Generalized Smoothness
Alexander Tyurin
Learning Safety Constraints for Large Language Models
Xin Chen, Yarden As, Andreas Krause
Bring Your Own (Non-Robust) Algorithm to Solve Robust MDPs by Estimating The Worst Kernel
Uri Gadot, Kaixin Wang, Navdeep Kumar et al.
GistScore: Learning Better Representations for In-Context Example Selection with Gist Bottlenecks
Shivanshu Gupta, Clemens Rosenbaum, Ethan R. Elenberg
Manifold Integrated Gradients: Riemannian Geometry for Feature Attribution
Eslam Zaher, Maciej Trzaskowski, Quan Nguyen et al.
Accelerating LLM Inference with Lossless Speculative Decoding Algorithms for Heterogeneous Vocabularies
Nadav Timor, Jonathan Mamou, Daniel Korat et al.
On the Diminishing Returns of Width for Continual Learning
Etash Guha, Vihan Lakshman
Autaptic Synaptic Circuit Enhances Spatio-temporal Predictive Learning of Spiking Neural Networks
Lihao Wang, Zhaofei Yu
BoA: Attention-aware Post-training Quantization without Backpropagation
Junhan Kim, Ho-young Kim, Eulrang Cho et al.
Sequential Disentanglement by Extracting Static Information From A Single Sequence Element
Nimrod Berman, Ilan Naiman, Idan Arbiv et al.
In-Context Deep Learning via Transformer Models
Weimin Wu, Maojiang Su, Jerry Yao-Chieh Hu et al.
A Sharper Global Convergence Analysis for Average Reward Reinforcement Learning via an Actor-Critic Approach
Swetha Ganesh, Washim Mondal, Vaneet Aggarwal
Operator SVD with Neural Networks via Nested Low-Rank Approximation
Jongha (Jon) Ryu, Xiangxiang Xu, Hasan Sabri Melihcan Erol et al.
Modular Learning of Deep Causal Generative Models for High-dimensional Causal Inference
Md Musfiqur Rahman, Murat Kocaoglu
Tilting the Odds at the Lottery: the Interplay of Overparameterisation and Curricula in Neural Networks
Stefano Mannelli, Yaraslau Ivashynka, Andrew Saxe et al.
Unsupervised Concept Discovery Mitigates Spurious Correlations
Md Rifat Arefin, Yan Zhang, Aristide Baratin et al.
Hierarchical Neural Operator Transformer with Learnable Frequency-aware Loss Prior for Arbitrary-scale Super-resolution
Xihaier Luo, Xiaoning Qian, Byung-Jun Yoon
Rethinking Specificity in SBDD: Leveraging Delta Score and Energy-Guided Diffusion
Bowen Gao, Minsi Ren, Yuyan Ni et al.
From Debate to Equilibrium: Belief‑Driven Multi‑Agent LLM Reasoning via Bayesian Nash Equilibrium
Yi Xie, Zhanke Zhou, Chentao Cao et al.
FedCal: Achieving Local and Global Calibration in Federated Learning via Aggregated Parameterized Scaler
Hongyi Peng, Han Yu, Xiaoli Tang et al.
Which Frequencies do CNNs Need? Emergent Bottleneck Structure in Feature Learning
Yuxiao Wen, Arthur Jacot
GeoMFormer: A General Architecture for Geometric Molecular Representation Learning
Tianlang Chen, Shengjie Luo, Di He et al.
A Graph is Worth $K$ Words: Euclideanizing Graph using Pure Transformer
Zhangyang Gao, Daize Dong, Cheng Tan et al.
BWS: Best Window Selection Based on Sample Scores for Data Pruning across Broad Ranges
Hoyong Choi, Nohyun Ki, Hye Won Chung
Compression via Pre-trained Transformers: A Study on Byte-Level Multimodal Data
David Heurtel-Depeiges, Anian Ruoss, Joel Veness et al.
Sliced-Wasserstein Estimation with Spherical Harmonics as Control Variates
Rémi Leluc, Aymeric Dieuleveut, François Portier et al.
VinT-6D: A Large-Scale Object-in-hand Dataset from Vision, Touch and Proprioception
Zhaoliang Wan, Yonggen Ling, Senlin Yi et al.
Differentially Private Decentralized Learning with Random Walks
Edwige Cyffers, Aurélien Bellet, Jalaj Upadhyay
REINFORCE Adversarial Attacks on Large Language Models: An Adaptive, Distributional, and Semantic Objective
Simon Geisler, Tom Wollschläger, M. Hesham Abdalla et al.
BRiTE: Bootstrapping Reinforced Thinking Process to Enhance Language Model Reasoning
Han Zhong, Yutong Yin, Shenao Zhang et al.
Counterfactual Reasoning for Multi-Label Image Classification via Patching-Based Training
Ming-Kun Xie, Jia-Hao Xiao, Pei Peng et al.
Disguised Copyright Infringement of Latent Diffusion Models
Yiwei Lu, Matthew Yang, Zuoqiu Liu et al.
Learning High-Frequency Functions Made Easy with Sinusoidal Positional Encoding
Chuanhao Sun, Zhihang Yuan, Kai Xu et al.
Generalizing from SIMPLE to HARD Visual Reasoning: Can We Mitigate Modality Imbalance in VLMs?
Simon Park, Abhishek Panigrahi, Yun Cheng et al.
Neural SPH: Improved Neural Modeling of Lagrangian Fluid Dynamics
Artur Toshev, Jonas Erbesdobler, Nikolaus Adams et al.
Q-VDiT: Towards Accurate Quantization and Distillation of Video-Generation Diffusion Transformers
Weilun Feng, Chuanguang Yang, Haotong Qin et al.
Energy-Guided Diffusion Sampling for Offline-to-Online Reinforcement Learning
Xu-Hui Liu, Tian-Shuo Liu, Shengyi Jiang et al.
Federated Offline Reinforcement Learning: Collaborative Single-Policy Coverage Suffices
Jiin Woo, Laixi Shi, Gauri Joshi et al.
The Pitfalls and Promise of Conformal Inference Under Adversarial Attacks
Ziquan Liu, Yufei Cui, Yan Yan et al.
Function Encoders: A Principled Approach to Transfer Learning in Hilbert Spaces
Tyler Ingebrand, Adam Thorpe, Ufuk Topcu
Chaos Meets Attention: Transformers for Large-Scale Dynamical Prediction
Yi He, Yiming Yang, Xiaoyuan Cheng et al.
Deep Linear Network Training Dynamics from Random Initialization: Data, Width, Depth, and Hyperparameter Transfer
Blake Bordelon, Cengiz Pehlevan
Visual Representation Learning with Stochastic Frame Prediction
Huiwon Jang, Dongyoung Kim, Junsu Kim et al.
Dialogue Without Limits: Constant-Sized KV Caches for Extended Response in LLMs
Ravi Ghadia, Avinash Kumar, Gaurav Jain et al.
Robustness of Deep Learning for Accelerated MRI: Benefits of Diverse Training Data
Kang Lin, Reinhard Heckel
Posterior Sampling-Based Bayesian Optimization with Tighter Bayesian Regret Bounds
Shion Takeno, Yu Inatsu, Masayuki Karasuyama et al.
Stochastic Q-learning for Large Discrete Action Spaces
Fares Fourati, Vaneet Aggarwal, Mohamed-Slim Alouini
Stochastic Forward–Backward Deconvolution: Training Diffusion Models with Finite Noisy Datasets
Haoye Lu, Qifan Wu, Yaoliang Yu
KernelWarehouse: Rethinking the Design of Dynamic Convolution
Chao Li, Anbang Yao
Near-Optimal Sample Complexity for MDPs via Anchoring
Jongmin Lee, Mario Bravo, Roberto Cominetti
Highly Compressed Tokenizer Can Generate Without Training
Lukas Lao Beyer, Tianhong Li, Xinlei Chen et al.
Loss Functions and Operators Generated by f-Divergences
Vincent Roulet, Tianlin Liu, Nino Vieillard et al.
Boosting Offline Optimizers with Surrogate Sensitivity
Cuong Dao, Phi Le Nguyen, Thao Nguyen Truong et al.
PARM: Multi-Objective Test-Time Alignment via Preference-Aware Autoregressive Reward Model
Baijiong Lin, Weisen Jiang, Yuancheng Xu et al.
Graph External Attention Enhanced Transformer
Jianqing Liang, Min Chen, Jiye Liang
Pre-Training Protein Bi-level Representation Through Span Mask Strategy On 3D Protein Chains
Jiale Zhao, Wanru Zhuang, Jia Song et al.
Rethinking Transformers in Solving POMDPs
Chenhao Lu, Ruizhe Shi, Yuyao Liu et al.
ProofAug: Efficient Neural Theorem Proving via Fine-grained Proof Structure Analysis
Haoxiong Liu, Jiacheng Sun, Zhenguo Li et al.
Causal Discovery with Fewer Conditional Independence Tests
Kirankumar Shiragur, Jiaqi Zhang, Caroline Uhler
MindAligner: Explicit Brain Functional Alignment for Cross-Subject Visual Decoding from Limited fMRI Data
Yuqin Dai, Zhouheng Yao, Chunfeng Song et al.
EffiCoder: Enhancing Code Generation in Large Language Models through Efficiency-Aware Fine-tuning
Dong HUANG, Guangtao Zeng, Jianbo Dai et al.