Tengyu Ma

39
Papers
1,288
Total Citations

Papers (39)

Matrix Completion has No Spurious Local Minimum

NeurIPS 2016arXiv
617
citations

Large Language Models as Tool Makers

ICLR 2024
262
citations

Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training

ICLR 2024
222
citations

On the Optimization Landscape of Tensor Decompositions

NeurIPS 2017arXiv
94
citations

Sum-of-Squares Lower Bounds for Sparse PCA

NeurIPS 2015arXiv
74
citations

A Non-generative Framework and Convex Relaxations for Unsupervised Learning

NeurIPS 2016arXiv
19
citations

Robust and On-the-fly Dataset Denoising for Image Classification

ECCV 2020
0
citations

Trash to Treasure: Low-Light Object Detection via Decomposition-and-Aggregation

AAAI 2024arXiv
0
citations

Rethinking Reconstruction and Denoising in the Dark: New Perspective, General Architecture and Beyond

CVPR 2025
0
citations

Linguistic Calibration of Long-Form Generations

ICML 2024
0
citations

Toward Fast, Flexible, and Robust Low-Light Image Enhancement

CVPR 2022arXiv
0
citations

Provable Guarantees for Self-Supervised Deep Learning with Spectral Contrastive Loss

NeurIPS 2021
0
citations

Safe Reinforcement Learning by Imagining the Near Future

NeurIPS 2021
0
citations

Why Do Pretrained Language Models Help in Downstream Tasks? An Analysis of Head and Prompt Tuning

NeurIPS 2021
0
citations

Calibrating Predictions to Decisions: A Novel Approach to Multi-Class Calibration

NeurIPS 2021
0
citations

Learning Barrier Certificates: Towards Safe Reinforcement Learning with Zero Training-time Violations

NeurIPS 2021
0
citations

Provable Model-based Nonlinear Bandit and Reinforcement Learning: Shelve Optimism, Embrace Virtual Curvature

NeurIPS 2021
0
citations

Label Noise SGD Provably Prefers Flat Global Minimizers

NeurIPS 2021
0
citations

Iterative Feature Matching: Toward Provable Domain Generalization with Logarithmic Environments

NeurIPS 2022
0
citations

Statistically Meaningful Approximation: a Case Study on Approximating Turing Machines with Transformers

NeurIPS 2022
0
citations

Beyond Separability: Analyzing the Linear Transferability of Contrastive Representations to Related Subpopulations

NeurIPS 2022
0
citations

Sharpness Minimization Algorithms Do Not Only Minimize Sharpness To Achieve Better Generalization

NeurIPS 2023
0
citations

What is the Inductive Bias of Flatness Regularization? A Study of Deep Matrix Factorization Models

NeurIPS 2023
0
citations

Data Selection for Language Models via Importance Resampling

NeurIPS 2023
0
citations

Beyond NTK with Vanilla Gradient Descent: A Mean-Field Analysis of Neural Networks with Polynomial Width, Samples, and Time

NeurIPS 2023
0
citations

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

NeurIPS 2023
0
citations

Online Learning of Eigenvectors

ICML 2015
0
citations

Provable Algorithms for Inference in Topic Models

ICML 2016
0
citations

Generalization and Equilibrium in Generative Adversarial Nets (GANs)

ICML 2017
0
citations

Data-dependent Sample Complexity of Deep Neural Networks via Lipschitz Augmentation

NeurIPS 2019
0
citations

Learning Imbalanced Datasets with Label-Distribution-Aware Margin Loss

NeurIPS 2019
0
citations

Regularization Matters: Generalization and Optimization of Neural Nets v.s. their Induced Kernel

NeurIPS 2019
0
citations

Towards Explaining the Regularization Effect of Initial Large Learning Rate in Training Neural Networks

NeurIPS 2019
0
citations

Verified Uncertainty Calibration

NeurIPS 2019
0
citations

Federated Accelerated Stochastic Gradient Descent

NeurIPS 2020
0
citations

Model-based Adversarial Meta-Reinforcement Learning

NeurIPS 2020
0
citations

MOPO: Model-based Offline Policy Optimization

NeurIPS 2020
0
citations

Self-training Avoids Using Spurious Features Under Domain Shift

NeurIPS 2020
0
citations

Beyond Lazy Training for Over-parameterized Tensor Decomposition

NeurIPS 2020
0
citations