Chuanhai Liu

ORCID: 0000-0003-1919-7280
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Statistical Methods and Bayesian Inference
  • Statistical Methods and Inference
  • Bayesian Modeling and Causal Inference
  • Bayesian Methods and Mixture Models
  • Advanced Statistical Methods and Models
  • Gaussian Processes and Bayesian Inference
  • Statistical Distribution Estimation and Applications
  • Markov Chains and Monte Carlo Methods
  • Machine Learning and Algorithms
  • Multi-Criteria Decision Making
  • Advanced Statistical Process Monitoring
  • Neural Networks and Applications
  • Optimal Experimental Design Methods
  • Mathematical Approximation and Integration
  • Gene expression and cancer classification
  • Blind Source Separation Techniques
  • Survey Sampling and Estimation Techniques
  • Statistical Methods in Clinical Trials
  • Genetic Associations and Epidemiology
  • Complex Network Analysis Techniques
  • Statistics Education and Methodologies
  • Probabilistic and Robust Engineering Design
  • Algorithms and Data Compression
  • Statistical Mechanics and Entropy
  • Soil Geostatistics and Mapping

Chinese University of Hong Kong
2024-2025

Purdue University West Lafayette
2013-2024

Ministry of Education of the People's Republic of China
2024

Hong Kong Science and Technology Parks Corporation
2024

University of Hong Kong
2024

University of Illinois Chicago
2014

Los Alamos National Laboratory
2010

Indiana University – Purdue University Indianapolis
2010

Harvard University Press
2009

Google (United States)
2006

A generalisation of the ecm algorithm (Meng & Rubin, 1993), which is itself an extension em (Dempster, Laird 1977), can be obtained by replacing some CM-steps ECM, maximise constrained expected complete-data loglikelihood function, with steps that correspondingly actual likelihood function. This algorithm, we call ecme for Expectation/Conditional Maximisation Either, shares both and their stable monotone convergence basic simplicity implementation relative to competing faster converging...

10.1093/biomet/81.4.633 article EN Biometrika 1994-01-01

Linear mixed-effects models are frequently used to analyze repeated measures data, because they model flexibly the within-subject correlation often present in this type of data. The most popular linear for a continuous response assumes normal distributions random effects and errors, making it sensitive outliers. Such outliers more problematic than fixed-effects models, may occur effects, or both, them harder be detected practice. Motivated by real dataset from an orthodontic study, we...

10.1198/10618600152628059 article EN Journal of Computational and Graphical Statistics 2001-06-01

AbstractThe Wang–Landau (WL) algorithm is an adaptive Markov chain Monte Carlo used to calculate the spectral density for a physical system. A remarkable feature of WL that it not trapped by local energy minima, which very important systems with rugged landscapes. This has led many successful applications in statistical physics and biophysics; however, there does exist rigorous theory support its convergence, estimates produced can reach only limited accuracy. In this article we propose...

10.1198/016214506000001202 article EN Journal of the American Statistical Association 2007-02-04

Posterior probabilistic statistical inference without priors is an important but so far elusive goal. Fisher's fiducial inference, Dempster–Shafer theory of belief functions, and Bayesian with default are attempts to achieve this goal but, date, none has given a completely satisfactory picture. This article presents new framework for based on inferential models (IMs), which not only provides data-dependent measures uncertainty about the unknown parameter, also does automatic long-run...

10.1080/01621459.2012.747960 article EN Journal of the American Statistical Association 2013-03-01

Abstract We present a method of analyzing series independent cross-sectional surveys in which some questions are not answered and respondents do answer the posed. The is also applicable to single survey different asked or sampling methods used strata clusters. Our involves multiply imputing missing items by adding existing imputation designed for hierarchical regression model that allows covariates at individual levels. Information from weights exploited including analysis variables on were...

10.1080/01621459.1998.10473737 article EN Journal of the American Statistical Association 1998-09-01

A new method for multinomial inference is proposed by representing the cell probabilities as unordered segments on unit interval and following Dempster-Shafer (DS) theory. The resulting DS posterior then strengthened to improve symmetry learning properties with final model being characterized a Dirichlet distribution. In addition computational simplicity, has desirable invariance related category permutations, refinements, coarsenings. Furthermore, relative amongst certain cells depends only...

10.48550/arxiv.2410.05512 preprint EN arXiv (Cornell University) 2024-10-07

Abstract Big mechanically-active culture systems (BigMACS) are promising to stimulate, control, and pattern cell tissue behaviours with less soluble factor requirements. However, it remains challenging predict if how distributed mechanical forces impact single-cell tissue. In this study, we introduce a tissue-scale finite element analysis (FEA) framework able correlate sub-cellular quantitative histology centimetre-scale biomechanics. Our is relevant diverse BigMACS, including media...

10.1088/1758-5090/adcd9f article EN cc-by Biofabrication 2025-04-16

Summary. Many chemical and environmental data sets are complicated by the existence of fully missing values or censored known to lie below detection thresholds. For example, week‐long samples airborne particulate matter were obtained at Alert, NWT, Canada, between 1980 1991, where some concentrations 24 constituents coarsened in sense being either limits. To facilitate scientific analysis, it is appealing create complete filling so that standard complete‐data methods can be applied. We...

10.1111/j.0006-341x.2001.00022.x article EN Biometrics 2001-03-01

Abstract The multivariate t distribution and other normal/independent distributions, such as the slash contaminated distribution, are used for robust regression with complete or incomplete data. Most previous work focused on method of maximum likelihood estimation linear using distributions. This article considers Bayesian models distributions fully observed predictor variables possible missing values from outcome variables. A monotone data augmentation algorithm posterior simulation...

10.1080/01621459.1996.10476991 article EN Journal of the American Statistical Association 1996-09-01

The Dempster--Shafer (DS) theory is a powerful tool for probabilistic reasoning based on formal calculus combining evidence. DS has been widely used in computer science and engineering applications, but yet to reach the statistical mainstream, perhaps because belief functions do not satisfy long-run frequency properties. Recently, two of authors proposed an extension DS, called weak (WB) approach, that can incorporate desirable properties into framework by systematically enlarging focal...

10.1214/10-sts322 article EN Statistical Science 2010-02-01

The inferential models (IM) framework provides prior-free, frequency-calibrated, posterior probabilistic inference. key is the use of random sets to predict unobservable auxiliary variables connected observable data and unknown parameters. When nuisance parameters are present, a marginalization step can reduce dimension variable which, in turn, leads more efficient For regular problems, exact be achieved, we give conditions for marginal IM validity. We show that our approach inference...

10.1080/01621459.2014.985827 article EN Journal of the American Statistical Association 2014-11-15

10.1006/jmva.1997.1703 article EN publisher-specific-oa Journal of Multivariate Analysis 1997-11-01

Summary The inferential model (IM) framework provides valid prior-free probabilistic inference by focusing on predicting unobserved auxiliary variables. But, efficient IM-based can be challenging when the variable is of higher dimension than parameter. Here we show that features are often fully observed and, in such cases, a simultaneous reduction and information aggregation achieved conditioning. This proposed conditioning strategy leads to IM casts new light Fisher's notions sufficiency,...

10.1111/rssb.12070 article EN Journal of the Royal Statistical Society Series B (Statistical Methodology) 2014-05-15

In this study, an abbreviated introduction to hierarchical statistical models for quantifying and explaining variations in the utilization of medical care is presented. The illustrative example was derived from analysis interstate variation coronary angiography Medicare patients with a recent acute myocardial infarction. model distinguished within-from between-states variation: former modeled via separate logistic regression each state, age sex as independent variables, while latter...

10.1097/00005650-199305001-00008 article EN Medical Care 1993-05-01

AbstractThis article describes a fast, statistically principled method for monitoring streams of network counts, which have long-term trends, rough cyclical patterns, outliers, and missing data. The key step is to build reference (predictive) model the counts that captures their complex, salient features but has just few parameters can be kept up-to-date as flow by, without requiring access past This justifies using negative binomial distribution with capture trends patterns moment...

10.1198/016214505000000943 article EN Journal of the American Statistical Association 2006-02-15

Multivariate ordinal data arise in many applications. This article proposes a new, efficient method for Bayesian inference multivariate probit models using Markov chain Monte Carlo techniques. The key idea is the novel use of parameter expansion to sample correlation matrices. A nice feature approach that performed straightforward Gibbs sampling. methods model selection are also discussed. Our motivated by study how women make decisions on taking medication reduce risk breast cancer....

10.1198/004017008000000064 article EN Technometrics 2008-05-01

The work of A. P. Dempster in 1960s extending Fisher's fiducial infer- ence for parametric inference using multivalued mapping, and that G. Shafer 1970s on the assessment combination evidence led to what is now known as Dempster-Shafer (DS) theory belief functions. However, application DS has been limited due, perhaps, its computational diffi- culty, non-uniqueness, lack frequency properties. In this paper, we return Dempster's original approach constructing functions ence, called basic...

10.5705/ss.2011.022a article EN Statistica Sinica 2011-04-01

A central focus of data science is the transformation empirical evidence into knowledge. As such, key insights and scientific attitudes deep thinkers like Fisher, Popper, Tukey are expected to inspire exciting new advances in machine learning artificial intelligence years come. Along these lines, present paper a novel {\em typicality principle} which states, roughly, that if observed sufficiently ``atypical'' certain sense relative posited theory, then theory unwarranted. This emphasis on...

10.48550/arxiv.2501.14860 preprint EN arXiv (Cornell University) 2025-01-24

From a model-building perspective, we propose paradigm shift for fitting over-parameterized models. Philosophically, the mindset is to fit models future observations rather than observed sample. Technically, given an imputation method generate observations, these by optimizing approximation of desired expected loss function based on its sample counterpart and adaptive duality function. The required also developed using same estimation technique with m-out-of-n bootstrap approach. We...

10.1080/01621459.2025.2455192 article EN Journal of the American Statistical Association 2025-01-31
Coming Soon ...