Percy Liang
53
Papers
1,451
Total Citations
Papers (53)
Certified Defenses for Data Poisoning Attacks
NeurIPS 2017arXiv
830
citations
Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training
ICLR 2024
222
citations
Large Language Models as Analogical Reasoners
ICLR 2024
131
citations
On the Learnability of Watermarks for Language Models
ICLR 2024
68
citations
BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments
ICLR 2025
49
citations
Benchmarking and Improving Generator-Validator Consistency of Language Models
ICLR 2024
41
citations
Unsupervised Risk Estimation Using Only Conditional Independence Structure
NeurIPS 2016arXiv
35
citations
Learning Overcomplete HMMs
NeurIPS 2017arXiv
23
citations
Model Equality Testing: Which Model is this API Serving?
ICLR 2025
16
citations
Establishing Best Practices in Building Rigorous Agentic Benchmarks
NeurIPS 2025
12
citations
Unsupervised Transformation Learning via Convex Relaxations
NeurIPS 2017arXiv
12
citations
Reliable and Efficient Amortized Model-based Evaluation
ICML 2025
10
citations
Position: In-House Evaluation Is Not Enough. Towards Robust Third-Party Evaluation and Flaw Disclosure for General-Purpose AI
ICML 2025
2
citations
Calibrated Structured Prediction
NeurIPS 2015
0
citations
On-the-Job Learning with Bayesian Decision Theory
NeurIPS 2015
0
citations
MLAgentBench: Evaluating Language Agents on Machine Learning Experimentation
ICML 2024
0
citations
Position: A Safe Harbor for AI Evaluation and Red Teaming
ICML 2024
0
citations
Position: On the Societal Impact of Open Foundation Models
ICML 2024
0
citations
Learning with Relaxed Supervision
NeurIPS 2015
0
citations
Prismatic VLMs: Investigating the Design Space of Visually-Conditioned Language Models
ICML 2024
0
citations
Estimating Mixture Models via Mixtures of Polynomials
NeurIPS 2015
0
citations
Cheaply Estimating Inference Efficiency Metrics for Autoregressive Transformer Models
NeurIPS 2023
0
citations
DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining
NeurIPS 2023
0
citations
Holistic Evaluation of Text-to-Image Models
NeurIPS 2023
0
citations
Reified Context Models
ICML 2015
0
citations
Learning Fast-Mixing Models for Structured Prediction
ICML 2015
0
citations
Estimation from Indirect Supervision with Linear Moments
ICML 2016
0
citations
Understanding Black-box Predictions via Influence Functions
ICML 2017
0
citations
Developing Bug-Free Machine Learning Systems With Formal Mathematics
ICML 2017
0
citations
World of Bits: An Open-Domain Platform for Web-Based Agents
ICML 2017
0
citations
Convexified Convolutional Neural Networks
ICML 2017
0
citations
Fairness Without Demographics in Repeated Loss Minimization
ICML 2018
0
citations
On the Relationship between Data Efficiency and Error for Uncertainty Sampling
ICML 2018
0
citations
Semidefinite relaxations for certifying robustness to adversarial examples
NeurIPS 2018
0
citations
Uncertainty Sampling is Preconditioned Stochastic Gradient Descent on Zero-One Loss
NeurIPS 2018
0
citations
A Retrieve-and-Edit Framework for Predicting Structured Outputs
NeurIPS 2018
0
citations
Unlabeled Data Improves Adversarial Robustness
NeurIPS 2019
0
citations
SPoC: Search-based Pseudocode to Code
NeurIPS 2019
0
citations
On the Accuracy of Influence Functions for Measuring Group Effects
NeurIPS 2019
0
citations
Verified Uncertainty Calibration
NeurIPS 2019
0
citations
Enabling certification of verification-agnostic networks via memory-efficient semidefinite programming
NeurIPS 2020
0
citations
Picking on the Same Person: Does Algorithmic Monoculture lead to Outcome Homogenization?
NeurIPS 2022
0
citations
Diffusion-LM Improves Controllable Text Generation
NeurIPS 2022
0
citations
Improving Self-Supervised Learning by Characterizing Idealized Representations
NeurIPS 2022
0
citations
Insights into Pre-training via Simpler Synthetic Tasks
NeurIPS 2022
0
citations
Decentralized Training of Foundation Models in Heterogeneous Environments
NeurIPS 2022
0
citations
What Can Transformers Learn In-Context? A Case Study of Simple Function Classes
NeurIPS 2022
0
citations
Deep Bidirectional Language-Knowledge Graph Pretraining
NeurIPS 2022
0
citations
PRODIGY: Enabling In-context Learning Over Graphs
NeurIPS 2023
0
citations
Lexinvariant Language Models
NeurIPS 2023
0
citations
AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback
NeurIPS 2023
0
citations
Data Selection for Language Models via Importance Resampling
NeurIPS 2023
0
citations
Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes
NeurIPS 2023
0
citations