Covariate Powered Cross-Weighted Multiple Testing

Overfitting Multiple comparisons problem Statistical power False Discovery Rate
DOI: 10.1111/rssb.12411 Publication Date: 2021-08-04T14:32:31Z
ABSTRACT
A fundamental task in the analysis of datasets with many variables is screening for associations. This can be cast as a multiple testing task, where objective achieving high detection power while controlling type I error. We consider $m$ hypothesis tests represented by pairs $((P_i, X_i))_{1\leq i \leq m}$ p-values $P_i$ and covariates $X_i$, such that $P_i \perp X_i$ if $H_i$ null. Here, we show how to use information potentially available about heterogeneities among hypotheses increase compared conventional procedures only $P_i$. To this end, upgrade existing weighted through Independent Hypothesis Weighting (IHW) framework data-driven weights are calculated function covariates. Finite sample guarantees, e.g., false discovery rate (FDR) control, derived from cross-weighting, data-splitting approach enables learning weight-covariate without overfitting long partitioned into independent folds, arbitrary within-fold dependence. IHW has increased methods do not covariate information. key implication rejection common setups should proceed according ranking p-values, but an alternative implied covariate-weighted p-values.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (78)
CITATIONS (39)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....