Percy Liang

53
Papers
1,451
Total Citations

Papers (53)

Certified Defenses for Data Poisoning Attacks

NeurIPS 2017arXiv
830
citations

Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training

ICLR 2024
222
citations

Large Language Models as Analogical Reasoners

ICLR 2024
131
citations

On the Learnability of Watermarks for Language Models

ICLR 2024
68
citations

BioDiscoveryAgent: An AI Agent for Designing Genetic Perturbation Experiments

ICLR 2025
49
citations

Benchmarking and Improving Generator-Validator Consistency of Language Models

ICLR 2024
41
citations

Unsupervised Risk Estimation Using Only Conditional Independence Structure

NeurIPS 2016arXiv
35
citations

Learning Overcomplete HMMs

NeurIPS 2017arXiv
23
citations

Model Equality Testing: Which Model is this API Serving?

ICLR 2025
16
citations

Establishing Best Practices in Building Rigorous Agentic Benchmarks

NeurIPS 2025
12
citations

Unsupervised Transformation Learning via Convex Relaxations

NeurIPS 2017arXiv
12
citations

Reliable and Efficient Amortized Model-based Evaluation

ICML 2025
10
citations

Position: In-House Evaluation Is Not Enough. Towards Robust Third-Party Evaluation and Flaw Disclosure for General-Purpose AI

ICML 2025
2
citations

Calibrated Structured Prediction

NeurIPS 2015
0
citations

On-the-Job Learning with Bayesian Decision Theory

NeurIPS 2015
0
citations

MLAgentBench: Evaluating Language Agents on Machine Learning Experimentation

ICML 2024
0
citations

Position: A Safe Harbor for AI Evaluation and Red Teaming

ICML 2024
0
citations

Position: On the Societal Impact of Open Foundation Models

ICML 2024
0
citations

Learning with Relaxed Supervision

NeurIPS 2015
0
citations

Prismatic VLMs: Investigating the Design Space of Visually-Conditioned Language Models

ICML 2024
0
citations

Estimating Mixture Models via Mixtures of Polynomials

NeurIPS 2015
0
citations

Cheaply Estimating Inference Efficiency Metrics for Autoregressive Transformer Models

NeurIPS 2023
0
citations

DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining

NeurIPS 2023
0
citations

Holistic Evaluation of Text-to-Image Models

NeurIPS 2023
0
citations

Reified Context Models

ICML 2015
0
citations

Learning Fast-Mixing Models for Structured Prediction

ICML 2015
0
citations

Estimation from Indirect Supervision with Linear Moments

ICML 2016
0
citations

Understanding Black-box Predictions via Influence Functions

ICML 2017
0
citations

Developing Bug-Free Machine Learning Systems With Formal Mathematics

ICML 2017
0
citations

World of Bits: An Open-Domain Platform for Web-Based Agents

ICML 2017
0
citations

Convexified Convolutional Neural Networks

ICML 2017
0
citations

Fairness Without Demographics in Repeated Loss Minimization

ICML 2018
0
citations

On the Relationship between Data Efficiency and Error for Uncertainty Sampling

ICML 2018
0
citations

Semidefinite relaxations for certifying robustness to adversarial examples

NeurIPS 2018
0
citations

Uncertainty Sampling is Preconditioned Stochastic Gradient Descent on Zero-One Loss

NeurIPS 2018
0
citations

A Retrieve-and-Edit Framework for Predicting Structured Outputs

NeurIPS 2018
0
citations

Unlabeled Data Improves Adversarial Robustness

NeurIPS 2019
0
citations

SPoC: Search-based Pseudocode to Code

NeurIPS 2019
0
citations

On the Accuracy of Influence Functions for Measuring Group Effects

NeurIPS 2019
0
citations

Verified Uncertainty Calibration

NeurIPS 2019
0
citations

Enabling certification of verification-agnostic networks via memory-efficient semidefinite programming

NeurIPS 2020
0
citations

Picking on the Same Person: Does Algorithmic Monoculture lead to Outcome Homogenization?

NeurIPS 2022
0
citations

Diffusion-LM Improves Controllable Text Generation

NeurIPS 2022
0
citations

Improving Self-Supervised Learning by Characterizing Idealized Representations

NeurIPS 2022
0
citations

Insights into Pre-training via Simpler Synthetic Tasks

NeurIPS 2022
0
citations

Decentralized Training of Foundation Models in Heterogeneous Environments

NeurIPS 2022
0
citations

What Can Transformers Learn In-Context? A Case Study of Simple Function Classes

NeurIPS 2022
0
citations

Deep Bidirectional Language-Knowledge Graph Pretraining

NeurIPS 2022
0
citations

PRODIGY: Enabling In-context Learning Over Graphs

NeurIPS 2023
0
citations

Lexinvariant Language Models

NeurIPS 2023
0
citations

AlpacaFarm: A Simulation Framework for Methods that Learn from Human Feedback

NeurIPS 2023
0
citations

Data Selection for Language Models via Importance Resampling

NeurIPS 2023
0
citations

Ecosystem-level Analysis of Deployed Machine Learning Reveals Homogeneous Outcomes

NeurIPS 2023
0
citations