Quanquan Gu

70
Papers
135
Total Citations

Papers (70)

How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?

ICLR 2024
85
citations

Semiparametric Differential Graph Models

NeurIPS 2016
24
citations

Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis

ICLR 2025arXiv
13
citations

Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment

ICML 2025
9
citations

CryoFM: A Flow-based Foundation Model for Cryo-EM Densities

ICLR 2025
4
citations

Speeding Up Latent Variable Gaussian Graphical Model Estimation via Nonconvex Optimization

NeurIPS 2017arXiv
0
citations

Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models

ICML 2024
0
citations

Borda Regret Minimization for Generalized Linear Dueling Bandits

ICML 2024
0
citations

Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption

ICML 2024
0
citations

Uncertainty-Aware Reward-Free Exploration with General Function Approximation

ICML 2024
0
citations

Feel-Good Thompson Sampling for Contextual Dueling Bandits

ICML 2024
0
citations

LLaVA-Critic: Learning to Evaluate Multimodal Models

CVPR 2025
0
citations

Diffusion Language Models Are Versatile Protein Learners

ICML 2024
0
citations

Protein Conformation Generation via Force-Guided SE(3) Diffusion Models

ICML 2024
0
citations

Position: TrustLLM: Trustworthiness in Large Language Models

ICML 2024
0
citations

High Dimensional EM Algorithm: Statistical Optimization and Asymptotic Normality

NeurIPS 2015
0
citations

Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks

NeurIPS 2021
0
citations

Do Wider Neural Networks Really Help Adversarial Robustness?

NeurIPS 2021
0
citations

Variance-Aware Off-Policy Evaluation with Linear Function Approximation

NeurIPS 2021
0
citations

Proxy Convexity: A Unified Framework for the Analysis of Neural Networks Trained by Gradient Descent

NeurIPS 2021
0
citations

Risk Bounds for Over-parameterized Maximum Margin Classification on Sub-Gaussian Mixtures

NeurIPS 2021
0
citations

Pure Exploration in Kernel and Neural Bandits

NeurIPS 2021
0
citations

Provably Efficient Reinforcement Learning with Linear Function Approximation under Adaptivity Constraints

NeurIPS 2021
0
citations

Uniform-PAC Bounds for Reinforcement Learning with Linear Function Approximation

NeurIPS 2021
0
citations

Nearly Minimax Optimal Reinforcement Learning for Discounted MDPs

NeurIPS 2021
0
citations

Iterative Teacher-Aware Learning

NeurIPS 2021arXiv
0
citations

Active Ranking without Strong Stochastic Transitivity

NeurIPS 2022
0
citations

A Simple and Provably Efficient Algorithm for Asynchronous Federated Contextual Linear Bandits

NeurIPS 2022
0
citations

Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime

NeurIPS 2022
0
citations

Towards Understanding the Mixture-of-Experts Layer in Deep Learning

NeurIPS 2022
0
citations

Benign Overfitting in Two-layer Convolutional Neural Networks

NeurIPS 2022
0
citations

The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift

NeurIPS 2022
0
citations

Learning Two-Player Markov Games: Neural Function Approximation and Correlated Equilibrium

NeurIPS 2022
0
citations

Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial Corruptions

NeurIPS 2022
0
citations

Computationally Efficient Horizon-Free Reinforcement Learning for Linear Mixture MDPs

NeurIPS 2022
0
citations

Robust Learning with Progressive Data Expansion Against Spurious Correlation

NeurIPS 2023
0
citations

Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU Networks on Nearly-orthogonal Data

NeurIPS 2023
0
citations

Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure

NeurIPS 2023
0
citations

Corruption-Robust Offline Reinforcement Learning with General Function Approximation

NeurIPS 2023
0
citations

Why Does Sharpness-Aware Minimization Generalize Better Than SGD?

NeurIPS 2023
0
citations

Towards a Lower Sample Complexity for Robust One-bit Compressed Sensing

ICML 2015
0
citations

On the Statistical Limits of Convex Relaxations

ICML 2016
0
citations

Towards Faster Rates and Oracle Property for Low-Rank Matrix Estimation

ICML 2016
0
citations

Uncertainty Assessment and False Discovery Rate Control in High-Dimensional Granger Causal Inference

ICML 2017
0
citations

Robust Gaussian Graphical Model Estimation with Arbitrary Corruption

ICML 2017
0
citations

A Unified Variance Reduction-Based Framework for Nonconvex Low-Rank Matrix Recovery

ICML 2017
0
citations

High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm

ICML 2017
0
citations

Covariate Adjusted Precision Matrix Estimation via Nonconvex Optimization

ICML 2018
0
citations

Continuous and Discrete-time Accelerated Stochastic Mirror Descent for Strongly Convex Functions

ICML 2018
0
citations

Fast and Sample Efficient Inductive Matrix Completion via Multi-Phase Procrustes Flow

ICML 2018
0
citations

A Primal-Dual Analysis of Global Optimality in Nonconvex Low-Rank Matrix Recovery

ICML 2018
0
citations

Stochastic Variance-Reduced Cubic Regularized Newton Methods

ICML 2018
0
citations

Stochastic Variance-Reduced Hamilton Monte Carlo Methods

ICML 2018
0
citations

On the Convergence and Robustness of Adversarial Training

ICML 2019
0
citations

Lower Bounds for Smooth Nonconvex Finite-Sum Optimization

ICML 2019
0
citations

Stochastic Nested Variance Reduction for Nonconvex Optimization

NeurIPS 2018
0
citations

Distributed Learning without Distress: Privacy-Preserving Empirical Risk Minimization

NeurIPS 2018
0
citations

Global Convergence of Langevin Dynamics Based Algorithms for Nonconvex Optimization

NeurIPS 2018
0
citations

Third-order Smoothness Helps: Faster Stochastic Optimization Algorithms for Finding Local Minima

NeurIPS 2018
0
citations

Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks

NeurIPS 2019
0
citations

An Improved Analysis of Training Over-parameterized Deep Neural Networks

NeurIPS 2019
0
citations

Algorithm-Dependent Generalization Bounds for Overparameterized Deep Residual Networks

NeurIPS 2019
0
citations

Layer-Dependent Importance Sampling for Training Deep and Large Graph Convolutional Networks

NeurIPS 2019
0
citations

Stochastic Gradient Hamiltonian Monte Carlo Methods with Recursive Variance Reduction

NeurIPS 2019
0
citations

Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks

NeurIPS 2019
0
citations

Agnostic Learning of a Single Neuron with Gradient Descent

NeurIPS 2020
0
citations

A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks

NeurIPS 2020
0
citations

A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods

NeurIPS 2020arXiv
0
citations

Reward-Free Model-Based Reinforcement Learning with Linear Function Approximation

NeurIPS 2021
0
citations

The Benefits of Implicit Regularization from SGD in Least Squares Problems

NeurIPS 2021
0
citations