论文标题
通过整数编程中随机测试中二元结果错误分类的敏感性分析
Sensitivity Analysis for Binary Outcome Misclassification in Randomization Tests via Integer Programming
论文作者
论文摘要
进行随机测试是在随机实验中测试因果无效假设的常见方法。随机测试的普及主要是因为它们的统计有效性仅取决于随机设计,并且不需要对结果变量的分布或建模假设。但是,随机测试仍然可能遭受其他偏见来源,其中结果错误分类是重要的。我们在随机测试中提出了一种无模型和有限型敏感性分析方法,用于二进制结果错误分类。我们框架中的核心数量是``警告准确性'',定义为阈值,以使基于测量结果的随机测试结果可能会根据真实结果而有所不同,如果结果测量的准确性没有超过该阈值。我们在不超过构度的情况下显示了对随机分析的范围,以表明该概念的范围是如何扩大了序列化的范围。可以通过对随机设计的大规模整数程序进行自适应重新安装大规模数据集,以有效地计算精度。
Conducting a randomization test is a common method for testing causal null hypotheses in randomized experiments. The popularity of randomization tests is largely because their statistical validity only depends on the randomization design, and no distributional or modeling assumption on the outcome variable is needed. However, randomization tests may still suffer from other sources of bias, among which outcome misclassification is a significant one. We propose a model-free and finite-population sensitivity analysis approach for binary outcome misclassification in randomization tests. A central quantity in our framework is ``warning accuracy," defined as the threshold such that a randomization test result based on the measured outcomes may differ from that based on the true outcomes if the outcome measurement accuracy did not surpass that threshold. We show how learning the warning accuracy and related concepts can amplify analyses of randomization tests subject to outcome misclassification without adding additional assumptions. We show that the warning accuracy can be computed efficiently for large data sets by adaptively reformulating a large-scale integer program with respect to the randomization design. We apply the proposed approach to the Prostate Cancer Prevention Trial (PCPT). We also developed an open-source R package for implementation of our approach.