Most Cited ICML 2024 "environment adaptation" Papers
2,635 papers found • Page 1 of 14
Conference
Position: The No Free Lunch Theorem, Kolmogorov Complexity, and the Role of Inductive Biases in Machine Learning
Micah Goldblum, Marc Finzi, Keefer Rowan et al.
Time Weaver: A Conditional Time Series Generation Model
Sai Shankar Narasimhan, Shubhankar Agarwal, Oguzhan Akcin et al.
Cascade-CLIP: Cascaded Vision-Language Embeddings Alignment for Zero-Shot Semantic Segmentation
Yunheng Li, Zhong-Yu Li, Quan-Sheng Zeng et al.
Data-efficient Large Vision Models through Sequential Autoregression
Zhiwei Hao, Jianyuan Guo, Chengcheng Wang et al.
Minimum-Norm Interpolation Under Covariate Shift
Neil Mallinar, Austin Zane, Spencer Frei et al.
Understanding Inter-Concept Relationships in Concept-Based Models
Naveen Raman, Mateo Espinosa Zarlenga, Mateja Jamnik
Learning from Integral Losses in Physics Informed Neural Networks
Ehsan Saleh, Saba Ghaffari, Timothy Bretl et al.
Faster Sampling via Stochastic Gradient Proximal Sampler
Xunpeng Huang, Difan Zou, Hanze Dong et al.
Prompt-based Visual Alignment for Zero-shot Policy Transfer
Haihan Gao, Rui Zhang, Qi Yi et al.
Deconstructing the Goldilocks Zone of Neural Network Initialization
Artem Vysogorets, Anna Dawid, Julia Kempe
Two Tales of Single-Phase Contrastive Hebbian Learning
Rasmus Kjær Høier, Christopher Zach
Regularizing with Pseudo-Negatives for Continual Self-Supervised Learning
Sungmin Cha, Kyunghyun Cho, Taesup Moon
Noise-Adaptive Confidence Sets for Linear Bandits and Application to Bayesian Optimization
Kwang-Sung Jun, Jungtaek Kim
Exploration by Optimization with Hybrid Regularizers: Logarithmic Regret with Adversarial Robustness in Partial Monitoring
Taira Tsuchiya, Shinji Ito, Junya Honda
Bounded and Uniform Energy-based Out-of-distribution Detection for Graphs
Shenzhi Yang, Bin Liang, An Liu et al.
Latent Noise Segmentation: How Neural Noise Leads to the Emergence of Segmentation and Grouping
Ben Lonnqvist, Zhengqing Wu, Michael Herzog
Foundation Policies with Hilbert Representations
Seohong Park, Tobias Kreiman, Sergey Levine
PDHG-Unrolled Learning-to-Optimize Method for Large-Scale Linear Programming
Bingheng Li, Linxin Yang, Yupeng Chen et al.
Path-Guided Particle-based Sampling
Mingzhou Fan, Ruida Zhou, Chao Tian et al.
Graph Neural Networks with a Distribution of Parametrized Graphs
See Hian Lee, Feng Ji, Kelin Xia et al.
Attribution-based Explanations that Provide Recourse Cannot be Robust
Hidde Fokkema, Rianne de Heide, Tim van Erven
AI Control: Improving Safety Despite Intentional Subversion
Ryan Greenblatt, Buck Shlegeris, Kshitij Sachan et al.
Density-Softmax: Efficient Test-time Model for Uncertainty Estimation and Robustness under Distribution Shifts
Ha Manh Bui, Anqi Liu
One Size Fits All for Semantic Shifts: Adaptive Prompt Tuning for Continual Learning
Doyoung Kim, Susik Yoon, Dongmin Park et al.
Matroid Semi-Bandits in Sublinear Time
Ruo-Chun Tzeng, Naoto Ohsaka, Kaito Ariu
Complexity Matters: Feature Learning in the Presence of Spurious Correlations
GuanWen Qiu, Da Kuang, Surbhi Goel
Bridging Data Gaps in Diffusion Models with Adversarial Noise-Based Transfer Learning
Xiyu Wang, Baijiong Lin, Daochang Liu et al.
Bring Your Own (Non-Robust) Algorithm to Solve Robust MDPs by Estimating The Worst Kernel
Uri Gadot, Kaixin Wang, Navdeep Kumar et al.
Online Linear Regression in Dynamic Environments via Discounting
Andrew Jacobsen, Ashok Cutkosky
RoSA: Accurate Parameter-Efficient Fine-Tuning via Robust Adaptation
Mahdi Nikdan, Soroush Tabesh, Elvir Crnčević et al.
Simplicity Bias of Two-Layer Networks beyond Linearly Separable Data
Nikita Tsoy, Nikola Konstantinov
Optimal Acceleration for Minimax and Fixed-Point Problems is Not Unique
TaeHo Yoon, Jaeyeon (Jay) Kim, Jaewook Suh et al.
Enhancing Sufficient Dimension Reduction via Hellinger Correlation
Seungbeom Hong, Ilmun Kim, Jun Song
Contrastive Preference Optimization: Pushing the Boundaries of LLM Performance in Machine Translation
Haoran Xu, Amr Sharaf, Yunmo Chen et al.
IM-Unpack: Training and Inference with Arbitrarily Low Precision Integers
Zhanpeng Zeng, Karthikeyan Sankaralingam, Vikas Singh
How Uniform Random Weights Induce Non-uniform Bias: Typical Interpolating Neural Networks Generalize with Narrow Teachers
Gon Buzaglo, Itamar Harel, Mor Shpigel Nacson et al.
Robust Stable Spiking Neural Networks
Ding Jianhao, Zhiyu Pan, Yujia Liu et al.
Exponential Spectral Pursuit: An Effective Initialization Method for Sparse Phase Retrieval
Mengchu Xu, Zhang Yuxuan, Jian Wang
Agnostic Learning of Mixed Linear Regressions with EM and AM Algorithms
Avishek Ghosh, Arya Mazumdar
Position: On the Societal Impact of Open Foundation Models
Sayash Kapoor, Rishi Bommasani, Kevin Klyman et al.
DIDI: Diffusion-Guided Diversity for Offline Behavioral Generation
Jinxin Liu, Xinghong Guo, Zifeng Zhuang et al.
High-Dimensional Bayesian Optimization via Semi-Supervised Learning with Optimized Unlabeled Data Sampling
Yuxuan Yin, Yu Wang, Peng Li
Automated Statistical Model Discovery with Language Models
Michael Li, Emily Fox, Noah Goodman
Curated LLM: Synergy of LLMs and Data Curation for tabular augmentation in low-data regimes
Nabeel Seedat, Nicolas Huynh, Boris van Breugel et al.
Robust Classification via a Single Diffusion Model
Huanran Chen, Yinpeng Dong, Zhengyi Wang et al.
Pursuing Overall Welfare in Federated Learning through Sequential Decision Making
Seok-Ju Hahn, Gi-Soo Kim, Junghye Lee
Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts
Jiang-Xin Shi, Tong Wei, Zhi Zhou et al.
Enabling Few-Shot Learning with PID Control: A Layer Adaptive Optimizer
Le Yu, Xinde Li, Pengfei Zhang et al.
Non-clairvoyant Scheduling with Partial Predictions
Ziyad Benomar, Vianney Perchet
Solving Poisson Equations using Neural Walk-on-Spheres
Hong Chul Nam, Julius Berner, Anima Anandkumar
REST: Efficient and Accelerated EEG Seizure Analysis through Residual State Updates
Arshia Afzal, Grigorios Chrysos, Volkan Cevher et al.
Trust Regions for Explanations via Black-Box Probabilistic Certification
Amit Dhurandhar, Swagatam Haldar, Dennis Wei et al.
Position: Compositional Generative Modeling: A Single Model is Not All You Need
Yilun Du, Leslie Kaelbling
On Mechanistic Knowledge Localization in Text-to-Image Generative Models
Samyadeep Basu, Keivan Rezaei, Priyatham Kattakinda et al.
Enhancing Size Generalization in Graph Neural Networks through Disentangled Representation Learning
Zheng Huang, Qihui Yang, Dawei Zhou et al.
Tuning-free Estimation and Inference of Cumulative Distribution Function under Local Differential Privacy
Yi Liu, Qirui Hu, Linglong Kong
Prodigy: An Expeditiously Adaptive Parameter-Free Learner
Konstantin Mishchenko, Aaron Defazio
Rich-Observation Reinforcement Learning with Continuous Latent Dynamics
Yuda Song, Lili Wu, Dylan Foster et al.
Neuroexplicit Diffusion Models for Inpainting of Optical Flow Fields
Tom Fischer, Pascal Peter, Joachim Weickert et al.
Centralized Selection with Preferences in the Presence of Biases
L. Elisa Celis, Amit Kumar, Nisheeth K. Vishnoi et al.
On Gradient-like Explanation under a Black-box Setting: When Black-box Explanations Become as Good as White-box
Yi Cai, Gerhard Wunder
CodeIt: Self-Improving Language Models with Prioritized Hindsight Replay
Natasha Butt, Blazej Manczak, Auke Wiggers et al.
On Computational Limits of Modern Hopfield Models: A Fine-Grained Complexity Analysis
Jerry Yao-Chieh Hu, Thomas Lin, Zhao Song et al.
Memory Efficient Neural Processes via Constant Memory Attention Block
Leo Feng, Frederick Tung, Hossein Hajimirsadeghi et al.
ESM All-Atom: Multi-Scale Protein Language Model for Unified Molecular Modeling
Kangjie Zheng, Siyu Long, Tianyu Lu et al.
Fast Adversarial Attacks on Language Models In One GPU Minute
Vinu Sankar Sadasivan, Shoumik Saha, Gaurang Sriramanan et al.
Value-Evolutionary-Based Reinforcement Learning
Pengyi Li, Jianye Hao, Hongyao Tang et al.
Integrating Multimodal Data for Joint Generative Modeling of Complex Dynamics
Manuel Brenner, Florian Hess, Georgia Koppe et al.
A Dynamic Algorithm for Weighted Submodular Cover Problem
Kiarash Banihashem, Samira Goudarzi, MohammadTaghi Hajiaghayi et al.
Individualized Privacy Accounting via Subsampling with Applications in Combinatorial Optimization
Badih Ghazi, Pritish Kamath, Ravi Kumar et al.
SelfIE: Self-Interpretation of Large Language Model Embeddings
Haozhe Chen, Carl Vondrick, Chengzhi Mao
Non-Asymptotic Analysis for Single-Loop (Natural) Actor-Critic with Compatible Function Approximation
Yudan Wang, Yue Wang, Yi Zhou et al.
Subsampling is not Magic: Why Large Batch Sizes Work for Differentially Private Stochastic Optimisation
Ossi Räisä, Joonas Jälkö, Antti Honkela
Random Scaling and Momentum for Non-smooth Non-convex Optimization
Qinzi Zhang, Ashok Cutkosky
Generative Flows on Discrete State-Spaces: Enabling Multimodal Flows with Applications to Protein Co-Design
Andrew Campbell, Jason Yim, Regina Barzilay et al.
Allocation Requires Prediction Only if Inequality Is Low
Ali Shirali, Rediet Abebe, Moritz Hardt
Position: The Causal Revolution Needs Scientific Pragmatism
Joshua Loftus
Dirichlet Flow Matching with Applications to DNA Sequence Design
Hannes Stärk, Bowen Jing, Chenyu Wang et al.
Detecting and Identifying Selection Structure in Sequential Data
Yujia Zheng, Zeyu Tang, Yiwen Qiu et al.
Conditional Common Entropy for Instrumental Variable Testing and Partial Identification
Ziwei Jiang, Murat Kocaoglu
Discovering Features with Synergistic Interactions in Multiple Views
Chohee Kim, M van der Schaar, Changhee Lee
Discovering Symmetry Breaking in Physical Systems with Relaxed Group Convolution
Rui Wang, Elyssa Hofgard, Han Gao et al.
Generalized Smooth Variational Inequalities: Methods with Adaptive Stepsizes
Daniil Vankov, Angelia Nedich, Lalitha Sankar
Test-Time Model Adaptation with Only Forward Passes
Shuaicheng Niu, Chunyan Miao, Guohao Chen et al.
How do Large Language Models Navigate Conflicts between Honesty and Helpfulness?
Ryan Liu, Theodore R Sumers, Ishita Dasgupta et al.
Stochastic Optimization with Arbitrary Recurrent Data Sampling
William Powell, Hanbaek Lyu
Multi-layer Rehearsal Feature Augmentation for Class-Incremental Learning
Bowen Zheng, Da-Wei Zhou, Han-Jia Ye et al.
Neural Collapse meets Differential Privacy: Curious behaviors of NoisyGD with Near-Perfect Representation Learning
Chendi Wang, Yuqing Zhu, Weijie Su et al.
Double Momentum Method for Lower-Level Constrained Bilevel Optimization
Wanli Shi, Yi Chang, Bin Gu
Emergent Equivariance in Deep Ensembles
Jan Gerken, Pan Kessel
Diffusion Model-Augmented Behavioral Cloning
Shang-Fu Chen, Hsiang-Chun Wang, Ming-Hao Hsu et al.
Near-Optimal Regret in Linear MDPs with Aggregate Bandit Feedback
Asaf Cassel, Haipeng Luo, Aviv Rosenberg et al.
Double Stochasticity Gazes Faster: Snap-Shot Decentralized Stochastic Gradient Tracking Methods
Hao Di, Haishan Ye, Xiangyu Chang et al.
A Federated Stochastic Multi-level Compositional Minimax Algorithm for Deep AUC Maximization
Xinwen Zhang, Ali Payani, Myungjin Lee et al.
GeoMFormer: A General Architecture for Geometric Molecular Representation Learning
Tianlang Chen, Shengjie Luo, Di He et al.
An Effective Dynamic Gradient Calibration Method for Continual Learning
Weichen Lin, Jiaxiang Chen, Ruomin Huang et al.
Recurrent Early Exits for Federated Learning with Heterogeneous Clients
Royson Lee, Javier Fernandez-Marques, Xu Hu et al.
GPTSwarm: Language Agents as Optimizable Graphs
Mingchen Zhuge, Wenyi Wang, Louis Kirsch et al.
CompeteAI: Understanding the Competition Dynamics of Large Language Model-based Agents
Qinlin Zhao, Jindong Wang, Yixuan Zhang et al.
Covert Malicious Finetuning: Challenges in Safeguarding LLM Adaptation
Danny Halawi, Alexander Wei, Eric Wallace et al.
Symbolic Music Generation with Non-Differentiable Rule Guided Diffusion
Yujia Huang, Adishree Ghatare, Yuanzhe Liu et al.
Q-Probe: A Lightweight Approach to Reward Maximization for Language Models
Kenneth Li, Samy Jelassi, Hugh Zhang et al.
Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution
Aaron Lou, Chenlin Meng, Stefano Ermon
Weisfeiler-Leman at the margin: When more expressivity matters
Billy Franks, Christopher Morris, Ameya Velingker et al.
Learning with Complementary Labels Revisited: The Selected-Completely-at-Random Setting Is More Practical
Wei Wang, Takashi Ishida, Yu-Jie Zhang et al.
Mechanistic Design and Scaling of Hybrid Architectures
Michael Poli, Armin Thomas, Eric Nguyen et al.
Learning Universal Predictors
Jordi Grau-Moya, Tim Genewein, Marcus Hutter et al.
Adaptive Hierarchical Certification for Segmentation using Randomized Smoothing
Alaa Anani, Tobias Lorenz, Bernt Schiele et al.
Efficient Mixture Learning in Black-Box Variational Inference
Alexandra Hotti, Oskar Kviman, Ricky Molén et al.
Discovering Bias in Latent Space: An Unsupervised Debiasing Approach
Dyah Adila, Shuai Zhang, Boran Han et al.
AD3: Implicit Action is the Key for World Models to Distinguish the Diverse Visual Distractors
Yucen Wang, Shenghua Wan, Le Gan et al.
HALC: Object Hallucination Reduction via Adaptive Focal-Contrast Decoding
Zhaorun Chen, Zhuokai Zhao, HONGYIN LUO et al.
SparseTSF: Modeling Long-term Time Series Forecasting with *1k* Parameters
Shengsheng Lin, Weiwei Lin, Wentai Wu et al.
Longitudinal Targeted Minimum Loss-based Estimation with Temporal-Difference Heterogeneous Transformer
Toru Shirakawa, Yi Li, Yulun Wu et al.
GaussianPro: 3D Gaussian Splatting with Progressive Propagation
Kai Cheng, Xiaoxiao Long, Kaizhi Yang et al.
Representation Surgery: Theory and Practice of Affine Steering
Shashwat Singh, Shauli Ravfogel, Jonathan Herzig et al.
Towards an Understanding of Stepwise Inference in Transformers: A Synthetic Graph Navigation Model
Mikail Khona, Maya Okawa, Jan Hula et al.
In-context Convergence of Transformers
Yu Huang, Yuan Cheng, Yingbin LIANG
Statistically Optimal Generative Modeling with Maximum Deviation from the Empirical Distribution
Elen Vardanyan, Sona Hunanyan, Tigran Galstyan et al.
Reward-Free Kernel-Based Reinforcement Learning
Sattar Vakili, Farhang Nabiei, Da-shan Shiu et al.
An Improved Finite-time Analysis of Temporal Difference Learning with Deep Neural Networks
Zhifa Ke, Zaiwen Wen, Junyu Zhang
A3S: A General Active Clustering Method with Pairwise Constraints
Xun Deng, Junlong Liu, Han Zhong et al.
Unified Training of Universal Time Series Forecasting Transformers
Gerald Woo, Chenghao Liu, Akshat Kumar et al.
Privacy Attacks in Decentralized Learning
Abdellah El Mrini, Edwige Cyffers, Aurélien Bellet
Unsupervised Parameter-free Simplicial Representation Learning with Scattering Transforms
Hiren Madhu, Sravanthi Gurugubelli, Sundeep Prabhakar Chepuri
Federated Combinatorial Multi-Agent Multi-Armed Bandits
Fares Fourati, Mohamed-Slim Alouini, Vaneet Aggarwal
ReconBoost: Boosting Can Achieve Modality Reconcilement
Cong Hua, Qianqian Xu, Shilong Bao et al.
Privacy-Preserving Instructions for Aligning Large Language Models
Da Yu, Peter Kairouz, Sewoong Oh et al.
What is the Long-Run Distribution of Stochastic Gradient Descent? A Large Deviations Analysis
Waïss Azizian, Franck Iutzeler, Jérôme Malick et al.
Learning to Model the World With Language
Jessy Lin, Yuqing Du, Olivia Watkins et al.
Degeneration-free Policy Optimization: RL Fine-Tuning for Language Models without Degeneration
Youngsoo Jang, Geon-Hyeong Kim, Byoungjip Kim et al.
An Unsupervised Approach for Periodic Source Detection in Time Series
Berken Utku Demirel, Christian Holz
Accelerating Federated Learning with Quick Distributed Mean Estimation
Ran Ben Basat, Shay Vargaftik, Amit Portnoy et al.
WorkArena: How Capable are Web Agents at Solving Common Knowledge Work Tasks?
Alexandre Drouin, Maxime Gasse, Massimo Caccia et al.
Theoretical Analysis of Learned Database Operations under Distribution Shift through Distribution Learnability
Sepanta Zeighami, Cyrus Shahabi
Non-Vacuous Generalization Bounds for Large Language Models
Sanae Lotfi, Marc Finzi, Yilun Kuang et al.
GRATH: Gradual Self-Truthifying for Large Language Models
Weixin Chen, Dawn Song, Bo Li
Learning Low-dimensional Latent Dynamics from High-dimensional Observations: Non-asymptotics and Lower Bounds
Yuyang Zhang, Shahriar Talebi, Na Li
Orthogonal Bootstrap: Efficient Simulation of Input Uncertainty
Kaizhao Liu, Jose Blanchet, Lexing Ying et al.
BBox-Adapter: Lightweight Adapting for Black-Box Large Language Models
Haotian Sun, Yuchen Zhuang, Wei Wei et al.
Principled Preferential Bayesian Optimization
Wenjie Xu, Wenbin Wang, Yuning Jiang et al.
Membership Inference Attacks on Diffusion Models via Quantile Regression
Shuai Tang, Steven Wu, Sergul Aydore et al.
Improving Gradient-Guided Nested Sampling for Posterior Inference
Pablo Lemos, Nikolay Malkin, Will Handley et al.
Scaling Laws for Fine-Grained Mixture of Experts
Jan Ludziejewski, Jakub Krajewski, Kamil Adamczewski et al.
On the Recoverability of Causal Relations from Temporally Aggregated I.I.D. Data
Shunxing Fan, Mingming Gong, Kun Zhang
Unsupervised Evaluation of Code LLMs with Round-Trip Correctness
Miltiadis Allamanis, Sheena Panthaplackel, Pengcheng Yin
Differentially Private Bias-Term Fine-tuning of Foundation Models
Zhiqi Bu, Yu-Xiang Wang, Sheng Zha et al.
Causal Discovery via Conditional Independence Testing with Proxy Variables
Mingzhou Liu, Xinwei Sun, YU QIAO et al.
Bayesian Power Steering: An Effective Approach for Domain Adaptation of Diffusion Models
Ding Huang, Ting Li, Jian Huang
Stochastic Localization via Iterative Posterior Sampling
Louis Grenioux, Maxence Noble, Marylou Gabrié et al.
Audio Flamingo: A Novel Audio Language Model with Few-Shot Learning and Dialogue Abilities
Zhifeng Kong, ARUSHI GOEL, Rohan Badlani et al.
FlowMM: Generating Materials with Riemannian Flow Matching
Benjamin Kurt Miller, Ricky T. Q. Chen, Anuroop Sriram et al.
RL-CFR: Improving Action Abstraction for Imperfect Information Extensive-Form Games with Reinforcement Learning
Boning Li, Zhixuan Fang, Longbo Huang
Consistent Diffusion Meets Tweedie: Training Exact Ambient Diffusion Models with Noisy Data
Giannis Daras, Alexandros Dimakis, Constantinos Daskalakis
COPAL: Continual Pruning in Large Language Generative Models
Srikanth Malla, Joon Hee Choi, Chiho Choi
Peeking with PEAK: Sequential, Nonparametric Composite Hypothesis Tests for Means of Multiple Data Streams
Brian Cho, Kyra Gan, Nathan Kallus
Predictive Dynamic Fusion
Bing Cao, Yinan Xia, Yi Ding et al.
Generative Active Learning for Long-tailed Instance Segmentation
Muzhi Zhu, Chengxiang Fan, Hao Chen et al.
Improving Open-Ended Text Generation via Adaptive Decoding
Wenhong Zhu, Hongkun Hao, Zhiwei He et al.
Positional Knowledge is All You Need: Position-induced Transformer (PiT) for Operator Learning
Junfeng CHEN, Kailiang Wu
Position: Key Claims in LLM Research Have a Long Tail of Footnotes
Anna Rogers, Sasha Luccioni
Applying language models to algebraic topology: generating simplicial cycles using multi-labeling in Wu's formula
Kirill Brilliantov, Fedor Pavutnitskiy, Dmitrii A. Pasechniuk et al.
DisCo-Diff: Enhancing Continuous Diffusion Models with Discrete Latents
Yilun Xu, Gabriele Corso, Tommi Jaakkola et al.
DySLIM: Dynamics Stable Learning by Invariant Measure for Chaotic Systems
Yair Schiff, Zhong Yi Wan, Jeffrey Parker et al.
Stability and Generalization for Stochastic Recursive Momentum-based Algorithms for (Strongly-)Convex One to $K$-Level Stochastic Optimizations
Xiaokang Pan, Xingyu Li, Jin Liu et al.
Spike Distance Function as a Learning Objective for Spike Prediction
Kevin Doran, Marvin Seifert, Carola Yovanovich et al.
RLAIF vs. RLHF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Harrison Lee, Samrat Phatale, Hassan Mansoor et al.
Cross-view Masked Diffusion Transformers for Person Image Synthesis
Trung Pham, Kang Zhang, Chang Yoo
Sharp Rates in Dependent Learning Theory: Avoiding Sample Size Deflation for the Square Loss
Ingvar Ziemann, Stephen Tu, George Pappas et al.
Coarse-to-Fine Highlighting: Reducing Knowledge Hallucination in Large Language Models
Qitan Lv, Jie Wang, Hanzhu Chen et al.
Position: Why We Must Rethink Empirical Research in Machine Learning
Moritz Herrmann, F. Julian D. Lange, Katharina Eggensperger et al.
Improving Token-Based World Models with Parallel Observation Prediction
Lior Cohen, Kaixin Wang, Bingyi Kang et al.
Position: Tensor Networks are a Valuable Asset for Green AI
Eva Memmel, Clara Menzen, Jetze Schuurmans et al.
On Statistical Learning Theory for Distributional Inputs
Christian Fiedler, Pierre-François Massiani, Friedrich Solowjow et al.
Hybrid Inverse Reinforcement Learning
Juntao Ren, Gokul Swamy, Steven Wu et al.
OODRobustBench: a Benchmark and Large-Scale Analysis of Adversarial Robustness under Distribution Shift
Lin Li, Yifei Wang, Chawin Sitawarin et al.
Chain-of-Thought Predictive Control
Zhiwei Jia, Vineet Thumuluri, Fangchen Liu et al.
Align Your Steps: Optimizing Sampling Schedules in Diffusion Models
Amirmojtaba Sabour, Sanja Fidler, Karsten Kreis
Nesting Particle Filters for Experimental Design in Dynamical Systems
Sahel Iqbal, Adrien Corenflos, Simo Särkkä et al.
Adaptively Perturbed Mirror Descent for Learning in Games
Kenshi Abe, Kaito Ariu, Mitsuki Sakamoto et al.
The Entropy Enigma: Success and Failure of Entropy Minimization
Ori Press, Ravid Shwartz-Ziv, Yann LeCun et al.
Collapse-Aware Triplet Decoupling for Adversarially Robust Image Retrieval
Qiwei Tian, Chenhao Lin, Zhengyu Zhao et al.
A Rate-Distortion View of Uncertainty Quantification
Ifigeneia Apostolopoulou, Benjamin Eysenbach, Frank Nielsen et al.
Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Michal Nauman, Michał Bortkiewicz, Piotr Milos et al.
Learning to Continually Learn with the Bayesian Principle
Soochan Lee, Hyeonseong Jeon, Jaehyeon Son et al.
Incentivized Learning in Principal-Agent Bandit Games
Antoine Scheid, Daniil Tiapkin, Etienne Boursier et al.
FESSNC: Fast Exponentially Stable and Safe Neural Controller
Jingdong Zhang, Luan Yang, Qunxi Zhu et al.
Sample-specific Masks for Visual Reprogramming-based Prompting
Chengyi Cai, Zesheng Ye, Lei Feng et al.
Recovering the Pre-Fine-Tuning Weights of Generative Models
Eliahu Horwitz, Jonathan Kahana, Yedid Hoshen
Transferring Knowledge From Large Foundation Models to Small Downstream Models
Shikai Qiu, Boran Han, Danielle Robinson et al.
Agent Instructs Large Language Models to be General Zero-Shot Reasoners
Nicholas Crispino, Kyle Montgomery, Fankun Zeng et al.
Simultaneous identification of models and parameters of scientific simulators
Cornelius Schröder, Jakob Macke
Human Alignment of Large Language Models through Online Preference Optimisation
Daniele Calandriello, Zhaohan Guo, REMI MUNOS et al.
ILILT: Implicit Learning of Inverse Lithography Technologies
Haoyu Yang, Mark Ren
Confidence-aware Contrastive Learning for Selective Classification
Yu-Chang Wu, Shen-Huan Lyu, Haopu Shang et al.
Fewer Truncations Improve Language Modeling
Hantian Ding, Zijian Wang, Giovanni Paolini et al.
APT: Adaptive Pruning and Tuning Pretrained Language Models for Efficient Training and Inference
Bowen Zhao, Hannaneh Hajishirzi, Qingqing Cao
Lyapunov-stable Neural Control for State and Output Feedback: A Novel Formulation
Lujie Yang, Hongkai Dai, Zhouxing Shi et al.
Collaborative Heterogeneous Causal Inference Beyond Meta-analysis
Tianyu Guo, Sai Praneeth Karimireddy, Michael Jordan
DsDm: Model-Aware Dataset Selection with Datamodels
Logan Engstrom