Jeremiah Zhe Liu

ORCID: 0000-0002-7410-4502
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Adversarial Robustness in Machine Learning
  • Gaussian Processes and Bayesian Inference
  • Air Quality Monitoring and Forecasting
  • Machine Learning and Data Classification
  • Air Quality and Health Impacts
  • Statistical Methods and Inference
  • Topic Modeling
  • Domain Adaptation and Few-Shot Learning
  • Anomaly Detection Techniques and Applications
  • Fault Detection and Control Systems
  • Natural Language Processing Techniques
  • Climate Change and Health Impacts
  • Advanced Neural Network Applications
  • Text Readability and Simplification
  • Explainable Artificial Intelligence (XAI)
  • Medical Imaging Techniques and Applications
  • Hate Speech and Cyberbullying Detection
  • Neural Networks and Applications
  • Software Engineering Research
  • Seismic Imaging and Inversion Techniques
  • Child Nutrition and Water Access
  • Advanced Multi-Objective Optimization Algorithms
  • Advanced Statistical Methods and Models
  • Speech and dialogue systems
  • Machine Learning in Healthcare

Harvard University
2015-2024

Google (United States)
2020-2024

UC San Diego Health System
2022-2023

Harvard University Press
2019-2023

University of California, Santa Cruz
2023

California University of Pennsylvania
2020

Boston Children's Hospital
2017

Chinese Research Academy of Environmental Sciences
2014

Overconfidence and underconfidence in machine learning classifiers is measured by calibration: the degree to which probabilities predicted for each class match accuracy of classifier on that prediction. How one measures calibration remains a challenge: expected error, most popular metric, has numerous flaws we outline, there no clear empirical understanding how its choices affect conclusions practice, what recommendations are counteract flaws. In this paper, perform comprehensive study...

10.48550/arxiv.1904.01685 preprint EN other-oa arXiv (Cornell University) 2019-01-01

Bayesian neural networks (BNN) and deep ensembles are principled approaches to estimate the predictive uncertainty of a learning model. However their practicality in real-time, industrial-scale applications limited due heavy memory inference cost. This motivates us study high-quality estimation that require only single network (DNN). By formalizing quantification as minimax problem, we first identify input distance awareness, i.e., model's ability quantify testing example from training data...

10.48550/arxiv.2006.10108 preprint EN cc-by arXiv (Cornell University) 2020-01-01

Mahalanobis distance (MD) is a simple and popular post-processing method for detecting out-of-distribution (OOD) inputs in neural networks. We analyze its failure modes near-OOD detection propose fix called relative (RMD) which improves performance more robust to hyperparameter choice. On wide selection of challenging vision, language, biology OOD benchmarks (CIFAR-100 vs CIFAR-10, CLINC intent detection, Genomics OOD), we show that RMD meaningfully upon MD (by up 15% AUROC on genomics OOD).

10.48550/arxiv.2106.09022 preprint EN other-oa arXiv (Cornell University) 2021-01-01

Recent approaches to efficiently ensemble neural networks have shown that strong robustness and uncertainty performance can be achieved with a negligible gain in parameters over the original network. However, these methods still require multiple forward passes for prediction, leading significant computational cost. In this work, we show surprising result: benefits of using predictions `for free' under single model's pass. particular, that, multi-input multi-output (MIMO) configuration, one...

10.48550/arxiv.2010.06610 preprint EN other-oa arXiv (Cornell University) 2020-01-01

To examine whether a commercial digital health application could support influenza surveillance in China.

10.2105/ajph.2017.303767 article EN American Journal of Public Health 2017-05-18

Ensemble learning is a standard approach to building machine systems that capture complex phenomena in real-world data. An important aspect of these the complete and valid quantification model uncertainty. We introduce Bayesian nonparametric ensemble (BNE) augments an existing account for different sources BNE model's prediction distribution functions using machinery. It has theoretical guarantee it robustly estimates uncertainty patterns data distribution, can decompose its overall...

10.48550/arxiv.1911.04061 preprint EN cc-by arXiv (Cornell University) 2019-01-01

Traditional (unstructured) pruning methods for a Transformer model focus on regularizing the individual weights by penalizing them toward zero. In this work, we explore spectral-normalized identity priors (SNIP), structured approach which penalizes an entire residual module in mapping. Our method identifies and discards unimportant non-linear mappings connections applying thresholding operator function norm, is applicable to any including single attention head, blocks, or feed-forward...

10.18653/v1/2020.findings-emnlp.64 article EN cc-by 2020-01-01

Accurate estimation of predictive uncertainty in modern neural networks is critical to achieve well calibrated predictions and detect out-of-distribution (OOD) inputs. The most promising approaches have been predominantly focused on improving model (e.g. deep ensembles Bayesian networks) post-processing techniques for OOD detection ODIN Mahalanobis distance). However, there has relatively little investigation into how the parametrization probabilities discriminative classifiers affects...

10.48550/arxiv.2007.05134 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Abstract Uncertainty has been a central concept in psychological theories of anxiety. However, this plagued by divergent connotations and operationalizations. The lack consensus hinders the current search for cognitive biological mechanisms anxiety, jeopardizes theory creation comparison, restrains translation basic research into improved diagnoses interventions. Drawing upon uncertainty decomposition Bayesian Decision Theory, we propose well‐defined conceptual structure clinical sciences,...

10.1111/bjop.12693 article EN British Journal of Psychology 2024-01-12

Structural variants (SVs) play a causal role in numerous diseases but can be difficult to detect and accurately genotype (determine zygosity) with short-read genome sequencing data (SRS). Improving SV genotyping accuracy SRS data, particularly for the many SVs first detected long-read sequencing, will improve our understanding of genetic variation.

10.1093/bioinformatics/btae129 article EN cc-by Bioinformatics 2024-03-01

Accurate image reconstruction is at the heart of diagnostics in medical imaging. Supervised deep learning-based approaches have been investigated for solving inverse problems including reconstruction. However, these trained models encounter unseen data distributions that are widely shifted from training during deployment. Therefore, it essential to assess whether a given input falls within distribution. Current uncertainty estimation focus on providing an map radiologists, rather than...

10.1109/jbhi.2024.3404883 article EN IEEE Journal of Biomedical and Health Informatics 2024-05-24

Gene-environment and nutrition-environment studies often involve testing of high-dimensional interactions between two sets variables, each having potentially complex nonlinear main effects on an outcome. Construction a valid powerful hypothesis test for such interaction is challenging, due to the difficulty in constructing efficient unbiased estimator complex, effects. In this work we address problem by proposing Cross-validated Ensemble Kernels (CVEK) that learns space appropriate functions...

10.1080/01621459.2021.1962889 article EN Journal of the American Statistical Association 2021-08-04

Content moderation is often performed by a collaboration between humans and machine learning models. However, it not well understood how to design the collaborative process so as maximize combined moderator-model system performance. This work presents rigorous study of this problem, focusing on an approach that incorporates model uncertainty into process. First, we introduce principled metrics describe performance under capacity constraints human moderator, quantifying efficiently utilizes...

10.18653/v1/2021.woah-1.5 article EN cc-by 2021-01-01

Pre-trained seq2seq models excel at graph semantic parsing with rich annotated data, but generalize worse to out-of-distribution (OOD) and long-tail examples. In comparison, symbolic parsers under-perform on population-level metrics, exhibit unique strength in OOD tail generalization. this work, we study compositionality-aware approach neural-symbolic inference informed by model confidence, performing fine-grained reasoning subgraph level (i.e., nodes edges) precisely targeting components...

10.18653/v1/2022.emnlp-main.314 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2022-01-01
Coming Soon ...