Percy Liang

ORCID: 0000-0002-0458-6139
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Topic Modeling
  • Natural Language Processing Techniques
  • Machine Learning and Algorithms
  • Multimodal Machine Learning Applications
  • Adversarial Robustness in Machine Learning
  • Machine Learning and Data Classification
  • Domain Adaptation and Few-Shot Learning
  • Software Engineering Research
  • Speech and dialogue systems
  • Explainable Artificial Intelligence (XAI)
  • Anomaly Detection Techniques and Applications
  • Speech Recognition and Synthesis
  • Stochastic Gradient Optimization Techniques
  • Software Testing and Debugging Techniques
  • Reinforcement Learning in Robotics
  • Advanced Neural Network Applications
  • Algorithms and Data Compression
  • Gaussian Processes and Bayesian Inference
  • Ethics and Social Impacts of AI
  • Bayesian Methods and Mixture Models
  • Artificial Intelligence in Healthcare and Education
  • Advanced Bandit Algorithms Research
  • Biomedical Text Mining and Ontologies
  • Statistical Methods and Inference
  • Semantic Web and Ontologies

Stanford University
2015-2024

University of Washington
2018-2023

Laboratoire d'Informatique de Paris-Nord
2014-2023

New York Academy of Sciences
2023

John Wiley & Sons (Germany)
2023

Moss Landing Marine Laboratories
2023

Hudson Institute
2023

Microsoft (United States)
2005-2020

Semantic Designs (United States)
2020

University of Iowa
2019

We present the Stanford Question Answering Dataset (SQuAD), a new reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on set Wikipedia articles, where answer to each question is segment text from corresponding passage. analyze understand types reasoning required questions, leaning heavily dependency and constituency trees. build strong logistic regression model, which achieves an F1 score 51.0%, significant improvement over simple baseline (20%). However,...

10.18653/v1/d16-1264 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2016-01-01

Extractive reading comprehension systems can often locate the correct answer to a question in context document, but they also tend make unreliable guesses on questions for which is not stated context. Existing datasets either focus exclusively answerable questions, or use automatically generated unanswerable that are easy identify. To address these weaknesses, we present SQuADRUn, new dataset combines existing Stanford Question Answering Dataset (SQuAD) with over 50,000 written adversarially...

10.18653/v1/p18-2124 article EN cc-by 2018-01-01

Xiang Lisa Li, Percy Liang. Proceedings of the 59th Annual Meeting Association for Computational Linguistics and 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2021.

10.18653/v1/2021.acl-long.353 article EN cc-by 2021-01-01

In this paper, we train a semantic parser that scales up to Freebase. Instead of relying on annotated logical forms, which is especially expensive obtain at large scale, learn from question-answer pairs. The main challenge in setting narrowing down the huge number possible predicates for given question. We tackle problem two ways: First, build coarse mapping phrases using knowledge base and text corpus. Second, use bridging operation generate additional based neighboring predicates. On...

10.18653/v1/d13-1160 article EN Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2013-01-01

Standard accuracy metrics indicate that reading comprehension systems are making rapid progress, but the extent to which these truly understand language remains unclear. To reward with real understanding abilities, we propose an adversarial evaluation scheme for Stanford Question Answering Dataset (SQuAD). Our method tests whether can answer questions about paragraphs contain adversarially inserted sentences, automatically generated distract computer without changing correct or misleading...

10.18653/v1/d17-1215 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2017-01-01

How can we explain the predictions of a black-box model? In this paper, use influence functions -- classic technique from robust statistics to trace model's prediction through learning algorithm and back its training data, thereby identifying points most responsible for given prediction. To scale up modern machine settings, develop simple, efficient implementation that requires only oracle access gradients Hessian-vector products. We show even on non-convex non-differentiable models where...

10.48550/arxiv.1703.04730 preprint EN other-oa arXiv (Cornell University) 2017-01-01

We present the Stanford Question Answering Dataset (SQuAD), a new reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on set Wikipedia articles, where answer to each question is segment text from corresponding passage. analyze understand types reasoning required questions, leaning heavily dependency and constituency trees. build strong logistic regression model, which achieves an F1 score 51.0%, significant improvement over simple baseline (20%). However,...

10.48550/arxiv.1606.05250 preprint EN other-oa arXiv (Cornell University) 2016-01-01

Scaling up language models has been shown to predictably improve performance and sample efficiency on a wide range of downstream tasks. This paper instead discusses an unpredictable phenomenon that we refer as emergent abilities large models. We consider ability be if it is not present in smaller but larger Thus, cannot predicted simply by extrapolating the The existence such emergence implies additional scaling could further expand capabilities

10.48550/arxiv.2206.07682 preprint EN cc-by arXiv (Cornell University) 2022-01-01

We present QuAC, a dataset for Question Answering in Context that contains 14K information-seeking QA dialogs (100K questions total). The involve two crowd workers: (1) student who poses sequence of freeform to learn as much possible about hidden Wikipedia text, and (2) teacher answers the by providing short excerpts from text. QuAC introduces challenges not found existing machine comprehension datasets: its are often more open-ended, unanswerable, or only meaningful within dialog context,...

10.18653/v1/d18-1241 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2018-01-01

Suppose we want to build a system that answers natural language question by representing its semantics as logical forxm and computing the answer given structured database of facts. The core part such is semantic parser maps questions forms. Semantic parsers are typically trained from examples annotated with their target forms, but this type annotation expensive. Our goal instead learn question–answer pairs, where form modeled latent variable. We develop new formalism, dependency-based...

10.1162/coli_a_00127 article EN Computational Linguistics 2012-08-22

A central challenge in semantic parsing is handling the myriad ways which knowledge base predicates can be expressed.Traditionally, parsers are trained primarily from text paired with information.Our goal to exploit much larger amounts of raw not tied any base.In this paper, we turn on its head.Given an input utterance, first use a simple method deterministically generate set candidate logical forms canonical realization natural language for each.Then, paraphrase model choose that best...

10.3115/v1/p14-1133 article EN cc-by 2014-01-01

Juncen Li, Robin Jia, He He, Percy Liang. Proceedings of the 2018 Conference North American Chapter Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers). 2018.

10.18653/v1/n18-1169 preprint EN cc-by 2018-01-01

Modeling crisp logical regularities is crucial in semantic parsing, making it difficult for neural models with no task-specific prior knowledge to achieve good results.In this paper, we introduce data recombination, a novel framework injecting such into model.From the training data, induce highprecision synchronous context-free grammar, which captures important conditional independence properties commonly found parsing.We then train sequence-to-sequence recurrent network (RNN) model...

10.18653/v1/p16-1002 article EN cc-by Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2016-01-01

Panupong Pasupat, Percy Liang. Proceedings of the 53rd Annual Meeting Association for Computational Linguistics and 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2015.

10.3115/v1/p15-1142 preprint EN cc-by 2015-01-01

Many applications of machine learning require a model to make accurate pre-dictions on test examples that are distributionally different from training ones, while task-specific labels scarce during training. An effective approach this challenge is pre-train related tasks where data abundant, and then fine-tune it downstream task interest. While pre-training has been in many language vision domains, remains an open question how effectively use graph datasets. In paper, we develop new strategy...

10.48550/arxiv.1905.12265 preprint EN other-oa arXiv (Cornell University) 2019-01-01

John Hewitt, Percy Liang. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and 9th International Joint (EMNLP-IJCNLP). 2019.

10.18653/v1/d19-1275 article EN cc-by 2019-01-01

We present an unsupervised approach to symmetric word alignment in which two simple asymmetric models are trained jointly maximize a combination of data likelihood and agreement between the models. Compared standard practice intersecting predictions independently-trained models, joint training provides 32% reduction AER. Moreover, efficient pair HMM aligners 29% AER over symmetrized IBM model 4 predictions.

10.3115/1220835.1220849 article EN 2006-01-01

Believable proxies of human behavior can empower interactive applications ranging from immersive environments to rehearsal spaces for interpersonal communication prototyping tools. In this paper, we introduce generative agents: computational software agents that simulate believable behavior. Generative wake up, cook breakfast, and head work; artists paint, while authors write; they form opinions, notice each other, initiate conversations; remember reflect on days past as plan the next day....

10.1145/3586183.3606763 article EN 2023-10-21

We demonstrate, theoretically and empirically, that adversarial robustness can significantly benefit from semisupervised learning. Theoretically, we revisit the simple Gaussian model of Schmidt et al. shows a sample complexity gap between standard robust classification. prove unlabeled data bridges this gap: learning procedure (self-training) achieves high accuracy using same number labels required for achieving accuracy. Empirically, augment CIFAR-10 with 500K images sourced 80 Million Tiny...

10.48550/arxiv.1905.13736 preprint EN other-oa arXiv (Cornell University) 2019-01-01

How do we build a semantic parser in new domain starting with zero training examples? We introduce methodology for this setting: First, use simple grammar to generate logical forms paired canonical utterances. The are meant cover the desired set of compositional operators, and utterances capture meaning (although clumsily). then crowdsourcing paraphrase these into natural resulting data is used train parser. further study role compositionality paraphrases. Finally, test our on seven domains...

10.3115/v1/p15-1129 article EN 2015-01-01

Machine learning systems trained on user-provided data are susceptible to poisoning attacks, whereby malicious users inject false training with the aim of corrupting learned model. While recent work has proposed a number attacks and defenses, little is understood about worst-case loss defense in face determined attacker. We address this by constructing approximate upper bounds across broad family for defenders that first perform outlier removal followed empirical risk minimization. Our...

10.48550/arxiv.1706.03691 preprint EN other-oa arXiv (Cornell University) 2017-01-01

While neural networks have achieved high accuracy on standard image classification benchmarks, their drops to nearly zero in the presence of small adversarial perturbations test inputs. Defenses based regularization and training been proposed, but often followed by new, stronger attacks that defeat these defenses. Can we somehow end this arms race? In work, study problem for with one hidden layer. We first propose a method semidefinite relaxation outputs certificate given network input, no...

10.48550/arxiv.1801.09344 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Overparameterized neural networks can be highly accurate on average an i.i.d. test set yet consistently fail atypical groups of the data (e.g., by learning spurious correlations that hold but not in such groups). Distributionally robust optimization (DRO) allows us to learn models instead minimize worst-case training loss over a pre-defined groups. However, we find naively applying group DRO overparameterized fails: these perfectly fit data, and any model with vanishing also already has...

10.48550/arxiv.1911.08731 preprint EN other-oa arXiv (Cornell University) 2019-01-01

A central problem in grounded language acquisition is learning the correspondences between a rich world state and stream of text which references that state. To deal with high degree ambiguity present this setting, we generative model simultaneously segments into utterances maps each utterance to meaning representation We show our generalizes across three domains increasing difficulty---Robocup sportscasting, weather forecasts (a new domain), NFL recaps.

10.3115/1687878.1687893 article EN 2009-01-01
Coming Soon ...