Dry Runs and "PWRD" Aggregation: Two New Methods for Extracting Power from Careful Observation of a Randomized Controlled Trial's Context
Lycurgus, Timothy
2022
Abstract
Well-conducted field experiments, broadly construed to contain both randomized controlled trials and quasi-experiments, involve extensive planning with substantive deliberation. Such deliberation has the potential to fuel and strengthen the analysis stage of the study. Each field experiment is unique in its own manner, from the subgroups on which effects are expected to concentrate to the design of the study itself. Reliance on off-the-shelf methods to analyze field experiments may exclude this potentially valuable information that, if handled properly, would provide a greater opportunity to detect an effect. In this dissertation, we propose two novel methods that look to extract information unique to a specific study and translate it into additional power. We demonstrate these methods on a large-scale education intervention aimed at correcting the stalled reading trajectories of early elementary students. The first method, Power-maximizing Weighting for Repeated-measurements and Delayed-effects (i.e. PWRD) aggregation, converts the theory of change behind a class of education interventions into a test statistic that maximizes the asymptotic relative efficiency over standard methods, thus providing greater power. The scheme emphasizes cohorts and years-of-follow-up on which effects are expected to accrue with appropriate attention paid to the relative precision of estimates within cohorts. We find through a simulation study mirroring the design of the reading intervention that this method provides stark gains in power over methods common to education research. While PWRD aggregation increases power, confidence interval estimation is more difficult. To alleviate this problem, we partition our parameter space into three regions: equivalence, superiority, and inferiority. In the first, we employ PWRD aggregation to provide the greatest opportunity to detect an effect. In the latter two regions, we employ a method common to the domain in which the research occurs such that when we are able to detect an effect, interpretation of the point estimate and confidence interval proceeds in a standard fashion. The second method we propose is a dry run simulation scheme that creates a pseudo-experiment replicating the initial randomized trial in a manner that preserves blinding to impact estimates. This procedure, which uses real rather than synthetic data, provides a sandbox in which various models may be tested to discover the model specification that most precisely estimates an artificially imposed treatment effect. The dry run method, similar in motivation to cross-validation and uniformity trials, allows the statistician advising field experiments to estimate expected losses for each of a variety of methods, enabling them to elect a novel or unfamiliar method if it demonstrably outperforms methods more familiar to the broader team. When applied to the reading intervention that motivated dry runs, results from this method challenged preconceived notions about covariate choice, suggesting we control for covariates beyond pre-test scores.Deep Blue DOI
Subjects
causal inference power analysis model selection hypothesis testing randomized trials
Types
Thesis
Metadata
Show full item recordCollections
Remediation of Harmful Language
The University of Michigan Library aims to describe its collections in a way that respects the people and communities who create, use, and are represented in them. We encourage you to Contact Us anonymously if you encounter harmful or problematic language in catalog records or finding aids. More information about our policies and practices is available at Remediation of Harmful Language.
Accessibility
If you are unable to use this file in its current format, please select the Contact Us link and we can modify it to make it more accessible to you.