Quanquan Gu
70
Papers
135
Total Citations
Papers (70)
How Many Pretraining Tasks Are Needed for In-Context Learning of Linear Regression?
ICLR 2024
85
citations
Semiparametric Differential Graph Models
NeurIPS 2016
24
citations
Convergence of Score-Based Discrete Diffusion Models: A Discrete-Time Analysis
ICLR 2025arXiv
13
citations
Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment
ICML 2025
9
citations
CryoFM: A Flow-based Foundation Model for Cryo-EM Densities
ICLR 2025
4
citations
Speeding Up Latent Variable Gaussian Graphical Model Estimation via Nonconvex Optimization
NeurIPS 2017arXiv
0
citations
Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models
ICML 2024
0
citations
Borda Regret Minimization for Generalized Linear Dueling Bandits
ICML 2024
0
citations
Towards Robust Model-Based Reinforcement Learning Against Adversarial Corruption
ICML 2024
0
citations
Uncertainty-Aware Reward-Free Exploration with General Function Approximation
ICML 2024
0
citations
Feel-Good Thompson Sampling for Contextual Dueling Bandits
ICML 2024
0
citations
LLaVA-Critic: Learning to Evaluate Multimodal Models
CVPR 2025
0
citations
Diffusion Language Models Are Versatile Protein Learners
ICML 2024
0
citations
Protein Conformation Generation via Force-Guided SE(3) Diffusion Models
ICML 2024
0
citations
Position: TrustLLM: Trustworthiness in Large Language Models
ICML 2024
0
citations
High Dimensional EM Algorithm: Statistical Optimization and Asymptotic Normality
NeurIPS 2015
0
citations
Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks
NeurIPS 2021
0
citations
Do Wider Neural Networks Really Help Adversarial Robustness?
NeurIPS 2021
0
citations
Variance-Aware Off-Policy Evaluation with Linear Function Approximation
NeurIPS 2021
0
citations
Proxy Convexity: A Unified Framework for the Analysis of Neural Networks Trained by Gradient Descent
NeurIPS 2021
0
citations
Risk Bounds for Over-parameterized Maximum Margin Classification on Sub-Gaussian Mixtures
NeurIPS 2021
0
citations
Pure Exploration in Kernel and Neural Bandits
NeurIPS 2021
0
citations
Provably Efficient Reinforcement Learning with Linear Function Approximation under Adaptivity Constraints
NeurIPS 2021
0
citations
Uniform-PAC Bounds for Reinforcement Learning with Linear Function Approximation
NeurIPS 2021
0
citations
Nearly Minimax Optimal Reinforcement Learning for Discounted MDPs
NeurIPS 2021
0
citations
Iterative Teacher-Aware Learning
NeurIPS 2021arXiv
0
citations
Active Ranking without Strong Stochastic Transitivity
NeurIPS 2022
0
citations
A Simple and Provably Efficient Algorithm for Asynchronous Federated Contextual Linear Bandits
NeurIPS 2022
0
citations
Risk Bounds of Multi-Pass SGD for Least Squares in the Interpolation Regime
NeurIPS 2022
0
citations
Towards Understanding the Mixture-of-Experts Layer in Deep Learning
NeurIPS 2022
0
citations
Benign Overfitting in Two-layer Convolutional Neural Networks
NeurIPS 2022
0
citations
The Power and Limitation of Pretraining-Finetuning for Linear Regression under Covariate Shift
NeurIPS 2022
0
citations
Learning Two-Player Markov Games: Neural Function Approximation and Correlated Equilibrium
NeurIPS 2022
0
citations
Nearly Optimal Algorithms for Linear Contextual Bandits with Adversarial Corruptions
NeurIPS 2022
0
citations
Computationally Efficient Horizon-Free Reinforcement Learning for Linear Mixture MDPs
NeurIPS 2022
0
citations
Robust Learning with Progressive Data Expansion Against Spurious Correlation
NeurIPS 2023
0
citations
Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU Networks on Nearly-orthogonal Data
NeurIPS 2023
0
citations
Optimal Extragradient-Based Algorithms for Stochastic Variational Inequalities with Separable Structure
NeurIPS 2023
0
citations
Corruption-Robust Offline Reinforcement Learning with General Function Approximation
NeurIPS 2023
0
citations
Why Does Sharpness-Aware Minimization Generalize Better Than SGD?
NeurIPS 2023
0
citations
Towards a Lower Sample Complexity for Robust One-bit Compressed Sensing
ICML 2015
0
citations
On the Statistical Limits of Convex Relaxations
ICML 2016
0
citations
Towards Faster Rates and Oracle Property for Low-Rank Matrix Estimation
ICML 2016
0
citations
Uncertainty Assessment and False Discovery Rate Control in High-Dimensional Granger Causal Inference
ICML 2017
0
citations
Robust Gaussian Graphical Model Estimation with Arbitrary Corruption
ICML 2017
0
citations
A Unified Variance Reduction-Based Framework for Nonconvex Low-Rank Matrix Recovery
ICML 2017
0
citations
High-Dimensional Variance-Reduced Stochastic Gradient Expectation-Maximization Algorithm
ICML 2017
0
citations
Covariate Adjusted Precision Matrix Estimation via Nonconvex Optimization
ICML 2018
0
citations
Continuous and Discrete-time Accelerated Stochastic Mirror Descent for Strongly Convex Functions
ICML 2018
0
citations
Fast and Sample Efficient Inductive Matrix Completion via Multi-Phase Procrustes Flow
ICML 2018
0
citations
A Primal-Dual Analysis of Global Optimality in Nonconvex Low-Rank Matrix Recovery
ICML 2018
0
citations
Stochastic Variance-Reduced Cubic Regularized Newton Methods
ICML 2018
0
citations
Stochastic Variance-Reduced Hamilton Monte Carlo Methods
ICML 2018
0
citations
On the Convergence and Robustness of Adversarial Training
ICML 2019
0
citations
Lower Bounds for Smooth Nonconvex Finite-Sum Optimization
ICML 2019
0
citations
Stochastic Nested Variance Reduction for Nonconvex Optimization
NeurIPS 2018
0
citations
Distributed Learning without Distress: Privacy-Preserving Empirical Risk Minimization
NeurIPS 2018
0
citations
Global Convergence of Langevin Dynamics Based Algorithms for Nonconvex Optimization
NeurIPS 2018
0
citations
Third-order Smoothness Helps: Faster Stochastic Optimization Algorithms for Finding Local Minima
NeurIPS 2018
0
citations
Tight Sample Complexity of Learning One-hidden-layer Convolutional Neural Networks
NeurIPS 2019
0
citations
An Improved Analysis of Training Over-parameterized Deep Neural Networks
NeurIPS 2019
0
citations
Algorithm-Dependent Generalization Bounds for Overparameterized Deep Residual Networks
NeurIPS 2019
0
citations
Layer-Dependent Importance Sampling for Training Deep and Large Graph Convolutional Networks
NeurIPS 2019
0
citations
Stochastic Gradient Hamiltonian Monte Carlo Methods with Recursive Variance Reduction
NeurIPS 2019
0
citations
Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks
NeurIPS 2019
0
citations
Agnostic Learning of a Single Neuron with Gradient Descent
NeurIPS 2020
0
citations
A Generalized Neural Tangent Kernel Analysis for Two-layer Neural Networks
NeurIPS 2020
0
citations
A Finite-Time Analysis of Two Time-Scale Actor-Critic Methods
NeurIPS 2020arXiv
0
citations
Reward-Free Model-Based Reinforcement Learning with Linear Function Approximation
NeurIPS 2021
0
citations
The Benefits of Implicit Regularization from SGD in Least Squares Problems
NeurIPS 2021
0
citations