Most Cited ICLR "deep reinforcement learning" Papers
6,124 papers found • Page 17 of 31
Conference
Few-Class Arena: A Benchmark for Efficient Selection of Vision Models and Dataset Difficulty Measurement
Bryan Bo Cao, Lawrence OGorman, Michael Coss et al.
Logicbreaks: A Framework for Understanding Subversion of Rule-based Inference
Anton Xue, Avishree Khare, Rajeev Alur et al.
Bad-PFL: Exploiting Backdoor Attacks against Personalized Federated Learning
Mingyuan Fan, Zhanyi Hu, Fuyi Wang et al.
Preserving Deep Representations in One-Shot Pruning: A Hessian-Free Second-Order Optimization Framework
Ryan Lucas, Rahul Mazumder
Wicked Oddities: Selectively Poisoning for Effective Clean-Label Backdoor Attacks
Hung Quang Nguyen, Hieu Nguyen, Anh Ta et al.
CR-CTC: Consistency regularization on CTC for improved speech recognition
Zengwei Yao, Wei Kang, Xiaoyu Yang et al.
DEEM: Diffusion models serve as the eyes of large language models for image perception
Run Luo, Yunshui Li, Longze Chen et al.
TS-LIF: A Temporal Segment Spiking Neuron Network for Time Series Forecasting
Shibo Feng, Wanjin Feng, Xingyu Gao et al.
An Illustrated Guide to Automatic Sparse Differentiation
Adrian Hill, Guillaume Dalle, Alexis Montoison
Neural Functions for Learning Periodic Signal
Woojin Cho, Minju Jo, Kookjin Lee et al.
Measuring And Improving Persuasiveness Of Large Language Models
SOMESH SINGH, Yaman Singla, Harini S I et al.
Robust System Identification: Finite-sample Guarantees and Connection to Regularization
Hank Park, Grani A. Hanasusanto, Yingying Li
Discrete Diffusion Schrödinger Bridge Matching for Graph Transformation
Jun Hyeong Kim, Seonghwan Kim, Seokhyun Moon et al.
A Statistical Framework for Ranking LLM-based Chatbots
Siavash Ameli, Siyuan Zhuang, Ion Stoica et al.
Intricacies of Feature Geometry in Large Language Models
Satvik Golechha, Lucius Bushnaq, Euan Ong et al.
Shape as Line Segments: Accurate and Flexible Implicit Surface Representation
Siyu Ren, Junhui Hou
NExUME: Adaptive Training and Inference for DNNs under Intermittent Power Environments
Cyan Subhra Mishra, Deeksha Chaudhary, Jack Sampson et al.
PEARL: Parallel Speculative Decoding with Adaptive Draft Length
Tianyu Liu, Yun Li, Qitan Lv et al.
Provably Robust Explainable Graph Neural Networks against Graph Perturbation Attacks
Jiate Li, Meng Pang, Yun Dong et al.
Beyond Linear Approximations: A Novel Pruning Approach for Attention Matrix
Yingyu Liang, Jiangxuan Long, Zhenmei Shi et al.
Capturing the Temporal Dependence of Training Data Influence
Jiachen (Tianhao) Wang, Dawn Song, James Y Zou et al.
CL-MFAP: A Contrastive Learning-Based Multimodal Foundation Model for Molecular Property Prediction and Antibiotic Screening
Gen Zhou, Sugitha Janarthanan, Yutong Lu et al.
Accelerating Goal-Conditioned Reinforcement Learning Algorithms and Research
Michał Bortkiewicz, Władysław Pałucki, Vivek Myers et al.
BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval
Hongjin SU, Howard Yen, Mengzhou Xia et al.
Reconciling Model Multiplicity for Downstream Decision Making
Ally Du, Dung Daniel Ngo, Steven Wu
Unlearning or Obfuscating? Jogging the Memory of Unlearned LLMs via Benign Relearning
Shengyuan Hu, Yiwei Fu, Steven Wu et al.
Convergent Privacy Loss of Noisy-SGD without Convexity and Smoothness
Eli Chien, Pan Li
Explanations of GNN on Evolving Graphs via Axiomatic Layer edges
Yazheng Liu, Sihong Xie
ARB-LLM: Alternating Refined Binarizations for Large Language Models
Zhiteng Li, Xianglong Yan, Tianao Zhang et al.
Dynamic Modeling of Patients, Modalities and Tasks via Multi-modal Multi-task Mixture of Experts
Chenwei Wu, Zitao Shuai, Zhengxu Tang et al.
ASTrA: Adversarial Self-supervised Training with Adaptive-Attacks
Prakash Chandra Chhipa, Gautam Vashishtha, Jithamanyu Settur et al.
MambaQuant: Quantizing the Mamba Family with Variance Aligned Rotation Methods
Dawei Yang, Yuxuan Yue, Xing Hu et al.
OSTQuant: Refining Large Language Model Quantization with Orthogonal and Scaling Transformations for Better Distribution Fitting
Xing Hu, Yuan Cheng, Dawei Yang et al.
GOttack: Universal Adversarial Attacks on Graph Neural Networks via Graph Orbits Learning
Zulfikar Alom, Tran Gia Bao Ngo, Murat Kantarcioglu et al.
Misspecified $Q$-Learning with Sparse Linear Function Approximation: Tight Bounds on Approximation Error
Ally Du, Lin Yang, Ruosong Wang
Regret-Optimal List Replicable Bandit Learning: Matching Upper and Lower Bounds
Michael Chen, A. Pavan, N. V. Vinodchandran et al.
Efficient Imitation under Misspecification
Nicolas Espinosa Dice, Sanjiban Choudhury, Wen Sun et al.
FreeCG: Free the Design Space of Clebsch-Gordan Transform for Machine Learning Force Fields
Shihao Shao, Haoran Geng, Zun Wang et al.
From Layers to States: A State Space Model Perspective to Deep Neural Network Layer Dynamics
Qinshuo Liu, Weiqin Zhao, Wei Huang et al.
Grounding Video Models to Actions through Goal Conditioned Exploration
Yunhao Luo, Yilun Du
Trajectory-LLM: A Language-based Data Generator for Trajectory Prediction in Autonomous Driving
Kairui Yang, Zihao Guo, Gengjie Lin et al.
Federated $Q$-Learning with Reference-Advantage Decomposition: Almost Optimal Regret and Logarithmic Communication Cost
Zhong Zheng, Haochen Zhang, Lingzhou Xue
Robust-PIFu: Robust Pixel-aligned Implicit Function for 3D Human Digitalization from a Single Image
Kennard Chan, Fayao Liu, Guosheng Lin et al.
Probe Pruning: Accelerating LLMs through Dynamic Pruning via Model-Probing
Qi Le, Enmao Diao, Ziyan Wang et al.
Discrete Distribution Networks
Lei Yang
Adapt-$\infty$: Scalable Continual Multimodal Instruction Tuning via Dynamic Data Selection
Adyasha Maharana, Jaehong Yoon, Tianlong Chen et al.
MLPs Learn In-Context on Regression and Classification Tasks
William Tong, Cengiz Pehlevan
Re-Evaluating the Impact of Unseen-Class Unlabeled Data on Semi-Supervised Learning Model
Rundong He, Yicong Dong, Lan-Zhe Guo et al.
FreqPrior: Improving Video Diffusion Models with Frequency Filtering Gaussian Noise
Yunlong Yuan, Yuanfan Guo, Chunwei Wang et al.
ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
Chen Bo Calvin Zhang, Zhang-Wei Hong, Aldo Pacchiano et al.
Grid Cell-Inspired Fragmentation and Recall for Efficient Map Building
Jaedong Hwang, Zhang-Wei Hong, Eric Chen et al.
Be More Diverse than the Most Diverse: Optimal Mixtures of Generative Models via Mixture-UCB Bandit Algorithms
Parham Rezaei, Farzan Farnia, Cheuk Ting Li
Oryx MLLM: On-Demand Spatial-Temporal Understanding at Arbitrary Resolution
Zuyan Liu, Yuhao Dong, Ziwei Liu et al.
DynamicCity: Large-Scale 4D Occupancy Generation from Dynamic Scenes
Hengwei Bian, Lingdong Kong, Haozhe Xie et al.
Federated Granger Causality Learning For Interdependent Clients With State Space Representation
Ayush Mohanty, Nazal Mohamed, Paritosh Ramanan et al.
MambaPEFT: Exploring Parameter-Efficient Fine-Tuning for Mamba
Masakazu Yoshimura, Teruaki Hayashi, Yota Maeda
CBQ: Cross-Block Quantization for Large Language Models
Xin Ding, Xiaoyu Liu, Zhijun Tu et al.
Magnetic Preference Optimization: Achieving Last-iterate Convergence for Language Model Alignment
Mingzhi Wang, Chengdong Ma, Qizhi Chen et al.
GenVP: Generating Visual Puzzles with Contrastive Hierarchical VAEs
Kalliopi Basioti, Pritish Sahu, Qingze Liu et al.
Learning-Augmented Frequent Directions
Anders Aamand, Justin Chen, Siddharth Gollapudi et al.
SV4D: Dynamic 3D Content Generation with Multi-Frame and Multi-View Consistency
Yiming Xie, Chun-Han Yao, Vikram Voleti et al.
MA-RLHF: Reinforcement Learning from Human Feedback with Macro Actions
Yekun Chai, Haoran Sun, Huang Fang et al.
Structuring Benchmark into Knowledge Graphs to Assist Large Language Models in Retrieving and Designing Models
Hanmo Liu, Shimin Di, Jialiang Wang et al.
SRSA: Skill Retrieval and Adaptation for Robotic Assembly Tasks
Yijie Guo, Bingjie Tang, Iretiayo Akinola et al.
Learning local equivariant representations for quantum operators
YinZhangHao Zhou, Zixi Gan, Shishir Pandey et al.
Towards Auto-Regressive Next-Token Prediction: In-context Learning Emerges from Generalization
Zixuan Gong, Xiaolin Hu, Huayi Tang et al.
DynaPrompt: Dynamic Test-Time Prompt Tuning
Zehao Xiao, Shilin Yan, Jack Hong et al.
Forgetting Transformer: Softmax Attention with a Forget Gate
Zhixuan Lin, Evgenii Nikishin, Xu He et al.
Adaptive Retention & Correction: Test-Time Training for Continual Learning
Haoran Chen, Micah Goldblum, Zuxuan Wu et al.
ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability
Zhongxiang Sun, Xiaoxue Zang, Kai Zheng et al.
Rethinking Audio-Visual Adversarial Vulnerability from Temporal and Modality Perspectives
Zeliang Zhang, Susan Liang, Daiki Shimada et al.
IntersectionZoo: Eco-driving for Benchmarking Multi-Agent Contextual Reinforcement Learning
Vindula Jayawardana, Baptiste Freydt, Ao Qu et al.
High-dimension Prototype is a Better Incremental Object Detection Learner
Yanjie Wang, Liqun Chen, Tianming Zhao et al.
A Simple yet Effective $\Delta\Delta G$ Predictor is An Unsupervised Antibody Optimizer and Explainer
Lirong Wu, Yunfan Liu, Haitao Lin et al.
Optimizing Neural Network Representations of Boolean Networks
Joshua Russell, Ignacio Gavier, Devdhar Patel et al.
On the Convergence of Adaptive Gradient Methods for Nonconvex Optimization
Quanquan Gu, Jinghui Chen, Yuan Cao et al.
Time After Time: Deep-Q Effect Estimation for Interventions on When and What to do
Yoav Wald, Mark Goldstein, Yonathan Efroni et al.
Mixture of Experts Made Personalized: Federated Prompt Learning for Vision-Language Models
Jun Luo, Chen Chen, Shandong Wu
LR0.FM: LOW-RESOLUTION ZERO-SHOT CLASSIFICATION BENCHMARK FOR FOUNDATION MODELS
Priyank Pathak, Shyam Marjit, Shruti Vyas et al.
Robust Simulation-Based Inference under Missing Data via Neural Processes
Yogesh Verma, Ayush Bharti, Vikas Garg
NoVo: Norm Voting off Hallucinations with Attention Heads in Large Language Models
Zhengyi Ho, Siyuan Liang, Sen Zhang et al.
Temporal Difference Learning: Why It Can Be Fast and How It Will Be Faster
Patrick Schnell, Luca Guastoni, Nils Thuerey
High-quality Text-to-3D Character Generation with SparseCubes and Sparse Transformers.
Jiachen Qian, Hongye Yang, Shuang Wu et al.
VLM2Vec: Training Vision-Language Models for Massive Multimodal Embedding Tasks
Ziyan Jiang, Rui Meng, Xinyi Yang et al.
Ensembling Diffusion Models via Adaptive Feature Aggregation
Cong Wang, kuan tian, Yonghang Guan et al.
EVA: Geometric Inverse Design for Fast Protein Motif-Scaffolding with Coupled Flow
Yufei Huang, Yunshu Liu, Lirong Wu et al.
LongMamba: Enhancing Mamba's Long-Context Capabilities via Training-Free Receptive Field Enlargement
Zhifan Ye, Kejing Xia, Yonggan Fu et al.
CodePlan: Unlocking Reasoning Potential in Large Language Models by Scaling Code-form Planning
Jiaxin Wen, Jian Guan, Hongning Wang et al.
Targeted Attack Improves Protection against Unauthorized Diffusion Customization
Boyang Zheng, Chumeng Liang, Xiaoyu Wu
Radar: Fast Long-Context Decoding for Any Transformer
Yongchang Hao, Mengyao Zhai, Hossein Hajimirsadeghi et al.
Learning the Optimal Stopping for Early Classification within Finite Horizons via Sequential Probability Ratio Test
Akinori F. Ebihara, Taiki Miyagawa, Kazuyuki Sakurai et al.
Interleaved Scene Graphs for Interleaved Text-and-Image Generation Assessment
Dongping Chen, Ruoxi Chen, Shu Pu et al.
Active Learning for Continual Learning: Keeping the Past Alive in the Present
Jaehyun Park, Dongmin Park, Jae-Gil Lee
Arithmetic Transformers Can Length-Generalize in Both Operand Length and Count
Hanseul Cho, Jaeyoung Cha, Srinadh Bhojanapalli et al.
Leave-One-Out Stable Conformal Prediction
Kiljae Lee, Yuan Zhang
DeepGate4: Efficient and Effective Representation Learning for Circuit Design at Scale
Ziyang Zheng, Shan Huang, Jianyuan Zhong et al.
Visually Guided Decoding: Gradient-Free Hard Prompt Inversion with Language Models
Donghoon Kim, Minji Bae, Kyuhong Shim et al.
Tight Time Complexities in Parallel Stochastic Optimization with Arbitrary Computation Dynamics
Alexander Tyurin
VILA-U: a Unified Foundation Model Integrating Visual Understanding and Generation
Yecheng Wu, Zhuoyang Zhang, Junyu Chen et al.
HART: Efficient Visual Generation with Hybrid Autoregressive Transformer
Haotian Tang, Yecheng Wu, Shang Yang et al.
Classic but Everlasting: Traditional Gradient-Based Algorithms Converge Fast Even in Time-Varying Multi-Player Games
Yanzheng Chen, Jun Yu
Null Counterfactual Factor Interactions for Goal-Conditioned Reinforcement Learning
Caleb Chuck, Fan Feng, Carl Qi et al.
$\text{I}^2\text{AM}$: Interpreting Image-to-Image Latent Diffusion Models via Bi-Attribution Maps
Junseo Park, Hyeryung Jang
Aligned Datasets Improve Detection of Latent Diffusion-Generated Images
Anirudh Sundara Rajan, Utkarsh Ojha, Jedidiah Schloesser et al.
MME-RealWorld: Could Your Multimodal LLM Challenge High-Resolution Real-World Scenarios that are Difficult for Humans?
YiFan Zhang, Huanyu Zhang, Haochen Tian et al.
MuPT: A Generative Symbolic Music Pretrained Transformer
Xingwei Qu, yuelin bai, Yinghao MA et al.
Shapley-Guided Utility Learning for Effective Graph Inference Data Valuation
Hongliang Chi, Qiong Wu, Zhengyi Zhou et al.
6DGS: Enhanced Direction-Aware Gaussian Splatting for Volumetric Rendering
Zhongpai Gao, Benjamin Planche, Meng Zheng et al.
3D Vision-Language Gaussian Splatting
Qucheng Peng, Benjamin Planche, Zhongpai Gao et al.
Order-aware Interactive Segmentation
Bin Wang, Anwesa Choudhuri, Meng Zheng et al.
Long-Context LLMs Meet RAG: Overcoming Challenges for Long Inputs in RAG
Bowen Jin, Jinsung Yoon, Jiawei Han et al.
Discovering Temporally Compositional Neural Manifolds with Switching Infinite GPFA
Changmin Yu, Maneesh Sahani, Máté Lengyel
ZIP: An Efficient Zeroth-order Prompt Tuning for Black-box Vision-Language Models
Seonghwan Park, Jaehyeon Jeong, Yongjun Kim et al.
Unlearning-based Neural Interpretations
Ching Lam Choi, Alexandre Duplessis, Serge Belongie
Bridging Information Asymmetry in Text-video Retrieval: A Data-centric Approach
Zechen Bai, Tianjun Xiao, Tong He et al.
A Truncated Newton Method for Optimal Transport
Mete Kemertas, Amir-massoud Farahmand, Allan Jepson
MamBEV: Enabling State Space Models to Learn Birds-Eye-View Representations
Hongyu Ke, Jack Morris, Kentaro Oguchi et al.
Sharpness-Aware Minimization Efficiently Selects Flatter Minima Late In Training
Zhanpeng Zhou, Mingze Wang, Yuchen Mao et al.
Differentially private optimization for non-decomposable objective functions
Weiwei Kong, Andres Munoz medina, Mónica Ribero
Multi-Robot Motion Planning with Diffusion Models
Yorai Shaoul, Itamar Mishani, Shivam Vats et al.
ImDy: Human Inverse Dynamics from Imitated Observations
Xinpeng Liu, Junxuan Liang, Zili Lin et al.
ReMatching Dynamic Reconstruction Flow
Sara Oblak, Despoina Paschalidou, Sanja Fidler et al.
Feature Averaging: An Implicit Bias of Gradient Descent Leading to Non-Robustness in Neural Networks
Binghui Li, Zhixuan Pan, Kaifeng Lyu et al.
Efficient Exploration and Discriminative World Model Learning with an Object-Centric Abstraction
Anthony GX-Chen, Kenneth Marino, Rob Fergus
CFD: Learning Generalized Molecular Representation via Concept-Enhanced Feedback Disentanglement
Aming Wu, Cheng Deng
Analysis of Linear Mode Connectivity via Permutation-Based Weight Matching: With Insights into Other Permutation Search Methods
Akira Ito, Masanori Yamada, Atsutoshi Kumagai
Scalable Benchmarking and Robust Learning for Noise-Free Ego-Motion and 3D Reconstruction from Noisy Video
Xiaohao Xu, Tianyi Zhang, Shibo Zhao et al.
MIND: Math Informed syNthetic Dialogues for Pretraining LLMs
Syeda Nahida Akter, Shrimai Prabhumoye, John Kamalu et al.
Building Math Agents with Multi-Turn Iterative Preference Learning
Wei Xiong, Chengshuai Shi, Jiaming Shen et al.
Local Loss Optimization in the Infinite Width: Stable Parameterization of Predictive Coding Networks and Target Propagation
Satoki Ishikawa, Rio Yokota, Ryo Karakida
Linear Partial Gromov-Wasserstein Embedding
Yikun Bai, Abihith Kothapalli, Hengrong Du et al.
SeCom: On Memory Construction and Retrieval for Personalized Conversational Agents
Zhuoshi Pan, Qianhui Wu, Huiqiang Jiang et al.
Determine-Then-Ensemble: Necessity of Top-k Union for Large Language Model Ensembling
Yuxuan YAO, Han Wu, Mingyang LIU et al.
State Space Model Meets Transformer: A New Paradigm for 3D Object Detection
Chuxin Wang, Wenfei Yang, Xiang Liu et al.
A primer on analytical learning dynamics of nonlinear neural networks
Rodrigo Carrasco-Davis, Erin Grant
Bio-xLSTM: Generative modeling, representation and in-context learning of biological and chemical sequences
Niklas Schmidinger, Lisa Schneckenreiter, Philipp Seidl et al.
Enhancing Clustered Federated Learning: Integration of Strategies and Improved Methodologies
Yongxin Guo, Xiaoying Tang, Tao Lin
TRACE: Temporal Grounding Video LLM via Causal Event Modeling
Yongxin Guo, Jingyu Liu, Mingda Li et al.
Timer-XL: Long-Context Transformers for Unified Time Series Forecasting
Yong Liu, Guo Qin, Xiangdong Huang et al.
Neural Fluid Simulation on Geometric Surfaces
Haoxiang Wang, Tao Yu, Hui Qiao et al.
Recovering Manifold Structure Using Ollivier Ricci Curvature
Tristan L. Saidi, Abigail Hickok, Andrew J Blumberg
EgoSim: Egocentric Exploration in Virtual Worlds with Multi-modal Conditioning
Wei Yu, Songheng Yin, Steve Easterbrook et al.
PWM: Policy Learning with Multi-Task World Models
Ignat Georgiev, Varun Giridhar, Nick Hansen et al.
Breaking the $\log(1/\Delta_2)$ Barrier: Better Batched Best Arm Identification with Adaptive Grids
Tianyuan Jin, Qin Zhang, Dongruo Zhou
Adaptive $Q$-Network: On-the-fly Target Selection for Deep Reinforcement Learning
Théo Vincent, Fabian Wahren, Jan Peters et al.
Single-agent Poisoning Attacks Suffice to Ruin Multi-Agent Learning
Fan Yao, Yuwei Cheng, Ermin Wei et al.
Do Contemporary Causal Inference Models Capture Real-World Heterogeneity? Findings from a Large-Scale Benchmark
Haining Yu, Yizhou Sun
A Differentiable Rank-Based Objective for Better Feature Learning
Krunoslav Lehman Pavasovic, Giulio Biroli, Levent Sagun
Mixture of In-Context Prompters for Tabular PFNs
Derek Xu, Olcay Cirit, Reza Asadi et al.
Point-based Instance Completion with Scene Constraints
Wesley Khademi, Li Fuxin
Spectral Compressive Imaging via Unmixing-driven Subspace Diffusion Refinement
Haijin Zeng, Benteng Sun, Yongyong Chen et al.
Tracking the Copyright of Large Vision-Language Models through Parameter Learning Adversarial Images
Yubo Wang, Jianting Tang, Liu et al.
Unifying Causal Representation Learning with the Invariance Principle
Dingling Yao, Dario Rancati, Riccardo Cadei et al.
Scalable Mechanistic Neural Networks
Jiale Chen, Dingling Yao, Adeel Pervez et al.
Deep Signature: Characterization of Large-Scale Molecular Dynamics
Tiexin Qin, Mengxu ZHU, Chunyang Li et al.
PICASO: Permutation-Invariant Context Composition with State Space Models
Tian Yu Liu, Alessandro Achille, Matthew Trager et al.
InstaSHAP: Interpretable Additive Models Explain Shapley Values Instantly
James Enouen, Yan Liu
LLMs' Potential Influences on Our Democracy: Challenges and Opportunities
Yujin Potter, David Rand, Yejin Choi et al.
Active Learning for Neural PDE Solvers
Daniel Musekamp, Marimuthu Kalimuthu, David Holzmüller et al.
LLaRA: Supercharging Robot Learning Data for Vision-Language Policy
Xiang Li, Cristina Mata, Jongwoo Park et al.
Solving hidden monotone variational inequalities with surrogate losses
Ryan D'Orazio, Danilo Vucetic, Zichu Liu et al.
Universal Sharpness Dynamics in Neural Network Training: Fixed Point Analysis, Edge of Stability, and Route to Chaos
Dayal Singh Kalra, Tianyu He, Maissam Barkeshli
Poisson-Dirac Neural Networks for Modeling Coupled Dynamical Systems across Domains
Razmik Khosrovian, Takaharu Yaguchi, Hiroaki Yoshimura et al.
AssembleFlow: Rigid Flow Matching with Inertial Frames for Molecular Assembly
Hongyu Guo, Yoshua Bengio, Shengchao Liu
Don't flatten, tokenize! Unlocking the key to SoftMoE's efficacy in deep RL
Ghada Sokar, Johan S Obando Ceron, Aaron Courville et al.
Dynamic Low-Rank Sparse Adaptation for Large Language Models
Weizhong Huang, Yuxin Zhang, Xiawu Zheng et al.
Conflict-Averse Gradient Aggregation for Constrained Multi-Objective Reinforcement Learning
Dohyeong Kim, Mineui Hong, Jeongho Park et al.
CAKE: Cascading and Adaptive KV Cache Eviction with Layer Preferences
Ziran Qin, Yuchen Cao, Mingbao Lin et al.
Identifiability for Gaussian Processes with Holomorphic Kernels
Ameer Qaqish, Didong Li
Neural Dueling Bandits: Preference-Based Optimization with Human Feedback
Arun Verma, Zhongxiang Dai, Xiaoqiang Lin et al.
U-Nets as Belief Propagation: Efficient Classification, Denoising, and Diffusion in Generative Hierarchical Models
Song Mei
Chemistry-Inspired Diffusion with Non-Differentiable Guidance
Yuchen Shen, Chenhao Zhang, Sijie Fu et al.
Langevin Soft Actor-Critic: Efficient Exploration through Uncertainty-Driven Critic Learning
Haque Ishfaq, Guangyuan Wang, Sami Islam et al.
Bootstrapping Language Models with DPO Implicit Rewards
Changyu Chen, Zichen Liu, Chao Du et al.
Hessian Free Efficient Single Loop Iterative Differentiation Methods for Bi-Level Optimization Problems
Peiran Yu, Junyi Li, Heng Huang
Is Your Video Language Model a Reliable Judge?
Ming Liu, Wensheng Zhang
Controllable Satellite-to-Street-View Synthesis with Precise Pose Alignment and Zero-Shot Environmental Control
Xianghui Ze, Zhenbo Song, Qiwei Wang et al.
R-Sparse: Rank-Aware Activation Sparsity for Efficient LLM Inference
Zhenyu Zhang, Zechun Liu, Yuandong Tian et al.
Param$\Delta$ for Direct Mixing: Post-Train Large Language Model At Zero Cost
Sheng Cao, Mingrui Wu, Karthik Prasad et al.
Dualformer: Controllable Fast and Slow Thinking by Learning with Randomized Reasoning Traces
Andy (DiJia) Su, Sainbayar Sukhbaatar, Michael Rabbat et al.
CraftRTL: High-quality Synthetic Data Generation for Verilog Code Models with Correct-by-Construction Non-Textual Representations and Targeted Code Repair
Mingjie Liu, Yun-Da Tsai, Wenfei Zhou et al.
Latent Bayesian Optimization via Autoregressive Normalizing Flows
Seunghun Lee, Jinyoung Park, Jaewon Chu et al.
Locality Sensitive Avatars From Video
Chunjin Song, Zhijie Wu, Shih-Yang Su et al.
Conditional Diffusion with Ordinal Regression: Longitudinal Data Generation for Neurodegenerative Disease Studies
Hyuna Cho, Ziquan Wei, Seungjoo Lee et al.
Vision-RWKV: Efficient and Scalable Visual Perception with RWKV-Like Architectures
Yuchen Duan, Weiyun Wang, Zhe Chen et al.
CREIMBO: Cross-Regional Ensemble Interactions in Multi-view Brain Observations
Noga Mudrik, Ryan Ly, Oliver Ruebel et al.
Understanding Model Calibration - A gentle introduction and visual exploration of calibration and the expected calibration error (ECE)
Maja Pavlovic
$\sigma$-zero: Gradient-based Optimization of $\ell_0$-norm Adversarial Examples
Antonio Emanuele Cinà, Francesco Villani, Maura Pintor et al.
Is uniform expressivity too restrictive? Towards efficient expressivity of GNNs
Sammy Khalife, Josué Tonelli-Cueto
Tuning Frequency Bias of State Space Models
Annan Yu, Dongwei Lyu, Soon Hoe Lim et al.
Denoising Task Difficulty-based Curriculum for Training Diffusion Models
Jin-Young Kim, Hyojun Go, Soonwoo Kwon et al.
Interpreting the Second-Order Effects of Neurons in CLIP
Yossi Gandelsman, Alexei Efros, Jacob Steinhardt
Interpreting and Editing Vision-Language Representations to Mitigate Hallucinations
Nick Jiang, Anish Kachinthaya, Suzanne Petryk et al.
Knowledge Entropy Decay during Language Model Pretraining Hinders New Knowledge Acquisition
Jiyeon Kim, Hyunji Lee, Hyowon Cho et al.
Model-agnostic meta-learners for estimating heterogeneous treatment effects over time
Dennis Frauen, Konstantin Hess, Stefan Feuerriegel
Interpretable Causal Representation Learning for Biological Data in the Pathway Space
Jesus de la Fuente Cedeño, Robert Lehmann, Carlos Ruiz-Arenas et al.
MCNC: Manifold-Constrained Reparameterization for Neural Compression
Chayne Thrash, Reed Andreas, Ali Abbasi et al.
Rewarding Progress: Scaling Automated Process Verifiers for LLM Reasoning
Amrith Setlur, Chirag Nagpal, Adam Fisch et al.
Adaptive Rank Allocation: Speeding Up Modern Transformers with RaNA Adapters
Roberto Garcia, Jerry Liu, Daniel Sorvisto et al.
UniDetox: Universal Detoxification of Large Language Models via Dataset Distillation
Huimin LU, Masaru Isonuma, Junichiro Mori et al.