Robust estimation of optimal dynamic treatment regimes for sequential treatment decisions
A dynamic treatment regime is a list of sequential decision rules for assigning treatment based on a patient's history. Q- and A-learning are two main approaches for estimating the optimal regime, i.e., that yielding the most beneficial outcome in the patient population, using data from a clinical trial or observational study. Q-learning requires postulated regression models for the outcome, while A-learning involves models for that part of the outcome regression representing treatment contrasts and for treatment assignment. We propose an alternative to Q- and A-learning that maximizes a doubly robust augmented inverse probability weighted estimator for population mean outcome over a restricted class of regimes. Simulations demonstrate the method's performance and robustness to model misspecification, which is a key concern. Copyright 2013, Oxford University Press.
Year of publication: |
2013
|
---|---|
Authors: | Zhang, Baqun ; Tsiatis, Anastasios A. ; Laber, Eric B. ; Davidian, Marie |
Published in: |
Biometrika. - Biometrika Trust, ISSN 0006-3444. - Vol. 100.2013, 3, p. 681-694
|
Publisher: |
Biometrika Trust |
Saved in:
Online Resource
Saved in favorites
Similar items by person
-
A Robust Method for Estimating Optimal Treatment Regimes
Zhang, Baqun, (2012)
-
Assessing the Causal Effect of Organ Transplantation on the Distribution of Residual Lifetime
Vock, David M., (2013)
-
Discussion of “Combining biomarkers to optimize patient treatment recommendation”
Laber, Eric B., (2014)
- More ...