Tengyu Ma
39
Papers
1,288
Total Citations
Papers (39)
Matrix Completion has No Spurious Local Minimum
NeurIPS 2016arXiv
617
citations
Large Language Models as Tool Makers
ICLR 2024
262
citations
Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
ICLR 2024
222
citations
On the Optimization Landscape of Tensor Decompositions
NeurIPS 2017arXiv
94
citations
Sum-of-Squares Lower Bounds for Sparse PCA
NeurIPS 2015arXiv
74
citations
A Non-generative Framework and Convex Relaxations for Unsupervised Learning
NeurIPS 2016arXiv
19
citations
Robust and On-the-fly Dataset Denoising for Image Classification
ECCV 2020
0
citations
Trash to Treasure: Low-Light Object Detection via Decomposition-and-Aggregation
AAAI 2024arXiv
0
citations
Rethinking Reconstruction and Denoising in the Dark: New Perspective, General Architecture and Beyond
CVPR 2025
0
citations
Linguistic Calibration of Long-Form Generations
ICML 2024
0
citations
Toward Fast, Flexible, and Robust Low-Light Image Enhancement
CVPR 2022arXiv
0
citations
Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss
NeurIPS 2021
0
citations
Safe Reinforcement Learning by Imagining the Near Future
NeurIPS 2021
0
citations
Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning
NeurIPS 2021
0
citations
Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration
NeurIPS 2021
0
citations
Learning Barrier Certificates: Towards Safe Reinforcement Learning with Zero Training-time Violations
NeurIPS 2021
0
citations
Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve Optimism, Embrace Virtual Curvature
NeurIPS 2021
0
citations
Label Noise SGD Provably Prefers Flat Global Minimizers
NeurIPS 2021
0
citations
Iterative Feature Matching: Toward Provable Domain Generalization with Logarithmic Environments
NeurIPS 2022
0
citations
Statistically Meaningful Approximation: a Case Study on Approximating Turing Machines with Transformers
NeurIPS 2022
0
citations
Beyond Separability: Analyzing the Linear Transferability of Contrastive Representations to Related Subpopulations
NeurIPS 2022
0
citations
Sharpness Minimization Algorithms Do Not Only Minimize Sharpness To Achieve Better Generalization
NeurIPS 2023
0
citations
What is the Inductive Bias of Flatness Regularization? A Study of Deep Matrix Factorization Models
NeurIPS 2023
0
citations
Data Selection for Language Models via Importance Resampling
NeurIPS 2023
0
citations
Beyond NTK with Vanilla Gradient Descent: A Mean-Field Analysis of Neural Networks with Polynomial Width, Samples, and Time
NeurIPS 2023
0
citations
DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
NeurIPS 2023
0
citations
Online Learning of Eigenvectors
ICML 2015
0
citations
Provable Algorithms for Inference in Topic Models
ICML 2016
0
citations
Generalization and Equilibrium in Generative Adversarial Nets (GANs)
ICML 2017
0
citations
Data-dependent Sample Complexity of Deep Neural Networks via Lipschitz Augmentation
NeurIPS 2019
0
citations
Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss
NeurIPS 2019
0
citations
Regularization Matters: Generalization and Optimization of Neural Nets v.s. their Induced Kernel
NeurIPS 2019
0
citations
Towards Explaining the Regularization Effect of Initial Large Learning Rate in Training Neural Networks
NeurIPS 2019
0
citations
Verified Uncertainty Calibration
NeurIPS 2019
0
citations
Federated Accelerated Stochastic Gradient Descent
NeurIPS 2020
0
citations
Model-based Adversarial Meta-Reinforcement Learning
NeurIPS 2020
0
citations
MOPO: Model-based Offline Policy Optimization
NeurIPS 2020
0
citations
Self-training Avoids Using Spurious Features Under Domain Shift
NeurIPS 2020
0
citations
Beyond Lazy Training for Over-parameterized Tensor Decomposition
NeurIPS 2020
0
citations