Causal Inference: What If - Harvard University

2y ago
30 Views
4 Downloads
5.92 MB
311 Pages
Last View : 14d ago
Last Download : 3m ago
Upload by : Lilly Kaiser
Transcription

Causal Inference: What IfMiguel A. Hernán, James M. RobinsNovember 23, 2020

iiCausal InferenceSuggested citation: Hernán MA, Robins JM (2020). Causal Inference:What If. Boca Raton: Chapman & Hall/CRC.This book cannot be purchased at this time, The draft is only available online at inference-book/

ContentsIntroduction: Towards less casual causal inferencesIviiCausal inference without models11 A definition of causal effect1.1 Individual causal effects . . .1.2 Average causal effects . . . .1.3 Measures of causal effect . . .1.4 Random variability . . . . . .1.5 Causation versus association.33478102 Randomized experiments2.1 Randomization . . . . . . . .2.2 Conditional randomization .2.3 Standardization . . . . . . . .2.4 Inverse probability weighting.13131719203 Observational studies3.1 Identifiability conditions . . . . . . . . . . . . . . . . . . . . . .3.2 Exchangeability . . . . . . . . . . . . . . . . . . . . . . . . . . .3.3 Positivity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .3.4 Consistency: First, define the counterfactual outcome . . . . .3.5 Consistency: Second, link counterfactuals to the observed data3.6 The target trial . . . . . . . . . . . . . . . . . . . . . . . . . . .252527303135374 Effect modification4.1 Definition of effect modification . . . . . . . .4.2 Stratification to identify effect modification .4.3 Why care about effect modification . . . . . .4.4 Stratification as a form of adjustment . . . .4.5 Matching as another form of adjustment . . .4.6 Effect modification and adjustment methods.414143454749505 Interaction5.1 Interaction requires a joint intervention . . . . .5.2 Identifying interaction . . . . . . . . . . . . . . .5.3 Counterfactual response types and interaction . .5.4 Sufficient causes . . . . . . . . . . . . . . . . . .5.5 Sufficient cause interaction . . . . . . . . . . . .5.6 Counterfactuals or sufficient-component causes? .55555658606365.

ivCausal Inference6 Graphical representation of causal effects6.1 Causal diagrams . . . . . . . . . . . . . . . . .6.2 Causal diagrams and marginal independence .6.3 Causal diagrams and conditional independence6.4 Positivity and consistency in causal diagrams .6.5 A structural classification of bias . . . . . . . .6.6 The structure of effect modification . . . . . . .696971737578807 Confounding7.1 The structure of confounding . . . . . .7.2 Confounding and exchangeability . . . .7.3 Confounding and the backdoor criterion7.4 Confounding and confounders . . . . . .7.5 Single-world intervention graphs . . . .7.6 Confounding adjustment . . . . . . . . .838385879093948 Selection bias8.1 The structure of selection bias8.2 Examples of selection bias . . .8.3 Selection bias and confounding8.4 Selection bias and censoring . .8.5 How to adjust for selection bias8.6 Selection without bias . . . . .99991011031051071109 Measurement bias9.1 Measurement error . . . . . . . . . .9.2 The structure of measurement error9.3 Mismeasured confounders . . . . . .9.4 Intention-to-treat effect: the effect of9.5 Per-protocol effect . . . . . . . . . .a. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .misclassified treatment. . . . . . . . . . . . . .11311311411611711910 Random variability10.1 Identification versus estimation .10.2 Estimation of causal effects . . .10.3 The myth of the super-population10.4 The conditionality “principle” . .10.5 The curse of dimensionality . . .123123126128130133II.Causal inference with models13711 Why model?11.1 Data cannot speak for themselves . . . . . . . . . .11.2 Parametric estimators of the conditional mean . . .11.3 Nonparametric estimators of the conditional mean11.4 Smoothing . . . . . . . . . . . . . . . . . . . . . . .11.5 The bias-variance trade-off . . . . . . . . . . . . . .13913914114214314512 IP weighting and marginal structural models12.1 The causal question . . . . . . . . . . . . . . . . . .12.2 Estimating IP weights via modeling . . . . . . . . .12.3 Stabilized IP weights . . . . . . . . . . . . . . . . .12.4 Marginal structural models . . . . . . . . . . . . . .12.5 Effect modification and marginal structural models.149149150153155157

CONTENTSv12.6 Censoring and missing data . . . . . . . . . . . . . . . . . . . . 15813 Standardization and the parametric g-formula13.1 Standardization as an alternative to IP weighting . . . . . . . .13.2 Estimating the mean outcome via modeling . . . . . . . . . . .13.3 Standardizing the mean outcome to the confounder distribution13.4 IP weighting or standardization? . . . . . . . . . . . . . . . . .13.5 How seriously do we take our estimates? . . . . . . . . . . . . .16116116316416516714 G-estimation of structural nested models14.1 The causal question revisited . . . . . . .14.2 Exchangeability revisited . . . . . . . . . .14.3 Structural nested mean models . . . . . .14.4 Rank preservation . . . . . . . . . . . . . .14.5 G-estimation . . . . . . . . . . . . . . . . .14.6 Structural nested models with two or more.17117117217317517717915 Outcome regression and propensity scores15.1 Outcome regression . . . . . . . . . . . . . . . . . . . . .15.2 Propensity scores . . . . . . . . . . . . . . . . . . . . . .15.3 Propensity stratification and standardization . . . . . . .15.4 Propensity matching . . . . . . . . . . . . . . . . . . . .15.5 Propensity models, structural models, predictive models.18318318518618818916 Instrumental variable estimation16.1 The three instrumental conditions . . . . . . . . . . . .16.2 The usual IV estimand . . . . . . . . . . . . . . . . . .16.3 A fourth identifying condition: homogeneity . . . . . .16.4 An alternative fourth condition: monotonicity . . . . .16.5 The three instrumental conditions revisited . . . . . .16.6 Instrumental variable estimation versus other methods.19319319619820020420617 Causal survival analysis17.1 Hazards and risks . . . . . . . . . . . . . .17.2 From hazards to risks . . . . . . . . . . . .17.3 Why censoring matters . . . . . . . . . . .17.4 IP weighting of marginal structural models17.5 The parametric g-formula . . . . . . . . .17.6 G-estimation of structural nested models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .parameters.20920921121421621721918 Variable selection for causal inference18.1 The different goals of variable selection . . .18.2 Variables that induce or amplify bias . . . .18.3 Causal inference and machine learning . . .18.4 Doubly robust machine learning estimators .18.5 Variable selection is a difficult problem . . .223223225228229230IIICausal inference from complex longitudinal data 23319 Time-varying treatments23519.1 The causal effect of time-varying treatments . . . . . . . . . . . 23519.2 Treatment strategies . . . . . . . . . . . . . . . . . . . . . . . . 23619.3 Sequentially randomized experiments . . . . . . . . . . . . . . . 237

viCausal Inference19.4 Sequential exchangeability . . . . . . . . . . . . . . . . . . . . . 24019.5 Identifiability under some but not all treatment strategies . . . 24119.6 Time-varying confounding and time-varying confounders . . . . 24520 Treatment-confounder feedback20.1 The elements of treatment-confounder feedback20.2 The bias of traditional methods . . . . . . . . .20.3 Why traditional methods fail . . . . . . . . . .20.4 Why traditional methods cannot be fixed . . . .20.5 Adjusting for past treatment . . . . . . . . . . .24724724925125325421 G-methods for time-varying treatments21.1 The g-formula for time-varying treatments . . . . . . .21.2 IP weighting for time-varying treatments . . . . . . . .21.3 A doubly robust estimator for time-varying treatments21.4 G-estimation for time-varying treatments . . . . . . . .21.5 Censoring is a time-varying treatment . . . . . . . . .257257261266268275.22 Target trial emulation27722.1 The target trial (revisited) . . . . . . . . . . . . . . . . . . . . . 27722.2 Causal effects in randomized trials . . . . . . . . . . . . . . . . 27822.3 Causal effects in observational analyses that emulate a target trial28122.4 Time zero . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28322.5 A unified analysis for causal inference . . . . . . . . . . . . . . . 284References288

INTRODUCTION: TOWARDS LESS CASUAL CAUSAL INFERENCESCausal Inference is an admittedly pretentious title for a book. A complexscientific task, causal inference relies on triangulating evidence from multiplesources and on the application of a variety of methodological approaches. Nobook can possibly provide a comprehensive description of all methodologies forcausal inference across the sciences. The authors of any Causal Inference bookwill have to choose which aspects of causal inference methodology they wantto emphasize.The title of this introduction reflects our own choices: a book that helpsscientists–especially health and social scientists–generate and analyze datato make causal inferences that are explicit about both the causal question andthe assumptions underlying the data analysis. Unfortunately, the scientificliterature is plagued by studies in which the causal question is not explicitlystated and the investigators’ unverifiable assumptions are not declared. Thiscasual attitude towards causal inference has led to a great deal of confusion.For example, it is not uncommon to find studies in which the effect estimatesare hard to interpret because the data analysis methods cannot appropriatelyanswer the causal question (were it explicitly stated) under the investigators’assumptions (were they declared).In this book, we stress the need to take the causal question seriously enoughto articulate it, and to delineate the separate roles of data and assumptions forcausal inference. Once these foundations are in place, causal inferences becomenecessarily less casual, which helps prevent confusion. The book describesvarious data analysis approaches to estimate the causal effect of interest undera particular set of assumptions when data are collected on each individual ina population. A key message of the book is that causal inference cannot bereduced to a collection of recipes for data analysis.The book is divided in three parts of increasing difficulty: Part I is aboutcausal inference without models (i.e., nonparametric identification of causal effects), Part II is about causal inference with models (i.e., estimation of causaleffects with parametric models), and Part III is about causal inference fromcomplex longitudinal data (i.e., estimation of causal effects of time-varyingtreatments). Throughout the text, we have interspersed Fine Points and Technical points that elaborate on certain topics mentioned in the main text. FinePoints are designed to be accessible to all readers while Technical Points aredesigned for readers with intermediate training in statistics. The book providesa cohesive presentation of concepts and methods for causal inference that arecurrently scattered across journals in several disciplines. We expect that itwill be of interest to all professionals that make causal inferences, includingepidemiologists, statisticians, psychologists, economists, sociologists, politicalscientists, computer scientists. . .This is not a philosophy book. We remain agnostic about metaphysicalconcepts like causality and cause. Instead, we focus on the identification andestimation of causal effects in populations, that is, numerical quantities thatmeasure changes in the distribution of an outcome under different interventions. For example, we discuss how to estimate the risk of death in patients

viiiCausal Inferencewith serious heart failure if they received a heart transplant versus if they didnot. Through actionable causal inference, we want to help decision makersmake better decisions.We are grateful to many people who have made this book possible. StephenCole, Sander Greenland, Jay Kaufman, Eleanor Murray, Sonja Swanson, TylerVanderWeele, and Jan Vandenbroucke provided detailed comments. GoodarzDanaei, Kosuke Kawai, Martin Lajous, and Kathleen Wirth helped create theNHEFS dataset. The sample code in Part II was developed by Roger Logan inSAS, Eleanor Murray and Roger Logan in Stata, Joy Shi and Sean McGrath inR, and James Fiedler in Python. Roger Logan has also been our LaTeX wizard.Randall Chaput helped create the figures in Chapters 1 and 2. Josh McKibledesigned the book cover. Rob Calver, our patient publisher, encouraged us towrite the book and supported our decision to make it freely available online.In addition, multiple colleagues have helped us improve the book by detecting typos and identifying unclear passages. We especially thank Kafui AdjayeGbewonyo, Álvaro Alonso, Katherine Almendinger, Ingelise Andersen, JuanJosé Beunza, Karen Biala, Joanne Brady, Alex Breskin, Shan Cai, Yu-HanChiu, Alexis Dinno, James Fiedler, Birgitte Frederiksen, Tadayoshi Fushiki,Leticia Grize, Dominik Hangartner, Michael Hudgens, John Jackson, MarshallJoffe, Luke Keele, Laura Khan, Dae Hyun Kim, Lauren Kunz, Martín Lajous,Angeliki Lambrou, Wen Wei Loh, Haidong Lu, Mohammad Ali Mansournia,Giovanni Marchetti, Lauren McCarl, Shira Mitchell, Louis Mittel, Hannah Oh,Ibironke Olofin, Robert Paige, Jeremy Pertman, Melinda Power, Bruce Psaty,Brian Sauer, Tomohiro Shinozaki, Ian Shrier, Yan Song, Øystein Sørensen,Etsuji Suzuki, Denis Talbot, Mohammad Tavakkoli, Sarah Taubman, EvanThacker, Kun-Hsing Yu, Vera Zietemann, Jessica Young, and Dorith Zimmermann.

Part ICausal inference without models

Chapter 1A DEFINITION OF CAUSAL EFFECTAs a human being, you are already innately familiar with causal inference’s fundamental concepts. Throughsheer existence, you know what a causal effect is, understand the difference between association and causation, andyou have used this knowledge consistently throughout your life. Had you not, you’d be dead. Without basic causalconcepts, you would not have survived long enough to read this chapter, let alone learn to read. As a toddler, youwould have jumped right into the swimming pool after seeing those who did were later able to reach the jam jar.As a teenager, you would have skied down the most dangerous slopes after seeing those who did won the next skirace. As a parent, you would have refused to give antibiotics to your sick child after observing that those childrenwho took their medicines were not at the park the next day.Since you already understand the definition of causal effect and the difference between association and causation, do not expect to gain deep conceptual insights from this chapter. Rather, the purpose of this chapter isto introduce mathematical notation that formalizes the causal intuition that you already possess. Make sure thatyou can match your causal intuition with the mathematical notation introduced here. This notation is necessaryto precisely define causal concepts, and will be used it throughout the book.1.1 Individual causal effectsCapital letters represent randomvariables. Lower case letters denoteparticular values of a random variable.Zeus is a patient waiting for a heart transplant. On January 1, he receives anew heart. Five days later, he dies. Imagine that we can somehow know–perhaps by divine revelation–that had Zeus not received a heart transplanton January 1, he would have been alive five days later. Equipped with thisinformation most would agree that the transplant caused Zeus’s death. Theheart transplant intervention had a causal effect on Zeus’s five-day survival.Another patient, Hera, also received a heart transplant on January 1. Fivedays later she was alive. Imagine we can somehow know that, had Hera notreceived the heart on January 1, she would still have been alive five days later.Hence the transplant did not have a causal effect on Hera’s five-day survival.These two vignettes illustrate how humans reason about causal effects:We compare (usually only mentally) the outcome when an action is takenversus the outcome when the action is withheld. If the two outcomes differ,we say that the action has a causal effect, causative or preventive, on theoutcome. Otherwise, we say that the action has no causal effect on theoutcome. Epidemiologists, statisticians, economists, and other social scientistsoften refer to the action as an intervention, an exposure, or a treatment.To make our causal intuition amenable to mathematical and statisticalanalysis we will introduce some notation. Consider a dichotomous treatmentvariable (1: treated, 0: untreated) and a dichotomous outcome variable (1: death, 0: survival). In this book we refer to variables such as and that may have different values for different individuals as random variables.Let 1 (read under treatment 1) be the outcome variable that wouldhave been observed under the treatment value 1, and 0 (read undertreatment 0) the outcome variable that would have been observed underthe treatment value 0. 1 and 0 are also random variables. Zeus

4Sometimes we abbreviate the expression “individual has outcome 1” by writing 1. Technically, when refers to a specificindividual, such as Zeus, is nota random variable because we areassuming that individual counterfactual outcomes are deterministic(see Technical Point 1.2).Causal effect for individual : 1 6 0Consistency:if , then A definition of causal effecthas 1 1 and 0 0 because he died when treated but would havesurvived if untreated, while Hera has 1 0 and 0 0 because shesurvived when treated and would also have survived if untreated.We can now provide a formal definition of a causal effect for an individual : The treatment has a causal effect on an individual’s outcome if 1 6 0 for the individual. Thus, the treatment has a causal effect onZeus’s outcome because 1 1 6 0 0 , but not on Hera’s outcomebecause 1 0 0 . The variables 1 and 0 are referred toas potential outcomes or as counterfactual outcomes. Some authors prefer theterm “potential outcomes” to emphasize that, depending on the treatment thatis received, either of these two outcomes can be potentially observed. Otherauthors prefer the term “counterfactual outcomes” to emphasize that theseoutcomes represent situations that may not actually occur (that is, counterto-the-fact situations).For each individual, one of the counterfactual outcomes–the one that corresponds to the treatment value that the individual did receive–is actuallyfactual. For example, because Zeus was actually treated ( 1), his counterfactual outcome under treatment 1 1 is equal to his observed (actual)outcome 1. That is, an individual with observed treatment equal to ,has observed outcome equal to his counterfactual outcome . This equalitycan be succinctly expressed as where denotes the counterfactual evaluated at the value corresponding to the individual’s observed treatment . The equality is referred to as consistency.Individual causal effects are defined as a contrast of the values of counterfactual outcomes, but only one of those outcomes is observed for each individual–the one corresponding to the treatment value actually experienced by the individual. All other counterfactual outcomes remain unobserved. Because ofmissing data, individual effects cannot be identified, that is, they cannot beexpressed as a function of the observed data (See Fine Point 2.1 for a possibleexception.)1.2 Average causal effectsWe needed three pieces of information to define an individual causal effect: anoutcome of interest, the actions 1 and 0 to be compared, and theindividual whose counterfactual outcomes 0 and 1 are to be compared.However, because identifying individual causal effects is generally not possible,we now turn our attention to an aggregated causal effect: the average causaleffect in a population of individuals. To define it, we need three pieces ofinformation: an outcome of interest, the actions 1 and 0 to becompared, and a well-defined population of individuals whose outcomes 0and 1 are to be compared.Take Zeus’s extended family as our population of interest. Table 1.1 showsthe counterfactual outcomes under both treatment ( 1) and no treatment( 0) for all 20 members of our population. Focus on the last column: theoutcome 1 that would have been observed for each individual if they hadreceived the treatment (a heart transplant). Half of the members of the population (10 out of 20) would have died if they had received a heart transplant.That is, the proportion of individuals that would have developed the outcomehad all population individuals received 1 is Pr[ 1 1] 10 20 0 5.Similarly, from the other column of Table 1.1, we can conclude that half of

1.2 Average causal effects5Fine Point 1.1Interference. Our definition of a counterfactual outcome implicitly assumes that an individual’s counterfactual outcomeunder treatment value does not depend on other individuals’ treatment values. For example, we implicitly assumedthat Zeus would die if he received a heart transplant, regardless of whether Hera also received a heart transplant. Thatis, Hera’s treatment value did not interfere with Zeus’s outcome. On the other hand, suppose that Hera’s gettinga new heart upsets Zeus to the extent that he would not survive his own heart transplant, even though he wouldhave survived had Hera not been transplanted. In this scenario, Hera’s treatment interferes with Zeus’s outcome.Interference between individuals is common in studies that deal with contagious agents or educational programs, inwhich an individual’s outcome is influenced by their social interaction with other population members.In the presence of interference, the counterfactual for an individual is not well defined because an individual’soutcome depends on other individuals’ treatment values. When there is interference, “the causal effect of heart transplanton Zeus’s outcome” is not well defined. Rather, one needs to refer to “the causal effect of heart transplant on Zeus’soutcome when Hera does not get a new heart” or “the causal effect of heart transplant on Zeus’s outcome when Heradoes get a new heart.” If other relatives and friends’ treatment also interfere with Zeus’s outcome, then one may needto refer to the causal effect of heart transplant on Zeus’s outcome when “no relative or friend gets a new heart,” “whenonly Hera gets a new heart,” etc. because the causal effect of treatment on Zeus’s outcome may differ for each particularallocation of hearts. The assumption of no interference was labeled “no interaction between units” by Cox (1958), andis included in the “stable-unit-treatment-value assumption (SUTVA)” described by Rubin (1980). See Halloran andStruchiner (1995), Sobel (2006), Rosenbaum (2007), and Hudgens and Halloran (2009) for a more detailed discussionof the role of interference in the definition of causal effects. Unless otherwise specified, we will assume no interferencethroughout this book.Table 1.1 rsephoneHermesHebeDionysus 001000100110110001111 110000001101111111000the members of the population (10 out of 20) would have died if they had notreceived a heart transplant. That is, the proportion of individuals that wouldhave developed the outcome had all population individuals received 0 isPr[ 0 1] 10 20 0 5. We have computed the counterfactual risk undertreatment to be 0 5 by counting the number of deaths (10) and dividing themby the total number of individuals (20), which is the same as computing theaverage of the counterfactual outcomes across all individuals in the population.To see the equivalence between risk and average for a dichotomous outcome,use the data in Table 1.1 to compute the average of 1 .We are now ready to provide a formal definition of the average causal effectin the population: An average causal effect of treatment on outcome is present if Pr[ 1 1] 6 Pr[ 0 1] in the population of interest.Under this definition, treatment does not have an average causal effect onoutcome in our population because both the risk of death under treatmentPr[ 1 1] and the risk of death under no treatment Pr[ 0 1] are0 5. It does not matter whether all or none of the individuals receive a hearttransplant: Half of them would die in either case. When, like here, the averagecausal effect in the population is null, we say that the null hypothesis of noaverage causal effect is true. Because the risk equals the average and becausethe letter E is usually employed to represent the population average or mean(also referred to as ‘E’xpectation), we can rewrite the definition of a non-nullaverage causal effect in the population as E[ 1 ] 6 E[ 0 ] so that thedefinition applies to both dichotomous and nondichotomous outcomes.The presence of an “average causal effect of heart transplant ” is definedby a contrast that involves the two actions “receiving a heart transplant ( 1)” and “not receiving a heart transplant ( 0).” When more than twoactions are possible (i.e., the treatment is not dichotomous), the particular

6A definition of causal effectFine Point 1.2Multiple versions of treatment. Our definition of a counterfactual outcome under treatment value also implicitlyassumes that there is only one version of treatment value . For example, we said that Zeus would die if hereceived a heart transplant. This statement implicitly assumes that all heart transplants are performed by the samesurgeon using the same procedure and equipment. That is, there is only one version of the treatment “heart transplant.”If there were multiple versions of treatment (e.g., surgeons with different skills), then it is possible that Zeus wouldsurvive if his transplant were performed by Asclepios, and would die if his transplant were performed by Hygieia. Inthe presence of multiple versions of treatment, the counterfactual for an individual is not well defined because anindividual’s outcome depends on the version of treatment . When there are multiple versions of treatment, “the causaleffect of heart transplant on Zeus’s outcome” is not well defined. Rather, one needs to refer to “the causal effect ofheart transplant on Zeus’s outcome when Asclepios performs the surgery” or “the causal effect of heart transplant onZeus’s outcome when Hygieia performs the surgery.” If other components of treatment (e.g., procedure, place) are alsorelevant to the outcome, then one may need to refer to “the causal effect of heart transplant on Zeus’s outcome whenAsclepios performs the surgery using his rod at the temple of Kos” because the causal effect of treatment on Zeus’soutcome may differ for each particular version of treatment.Like the assumption of no interference (see Fine Point 1.1), the assumption of no multiple versions of treatment isincluded in the “stable-unit-treatment-value assumption (SUTVA)” described by Rubin (1980). Robins and Greenland(2000) made the point that if the versions of a particular treatment (e.g., heart transplant) had the same causal effecton the outcome (survival), then the counterfactual 1 would be well-defined. VanderWeele (2009) formalized thispoint as the assumption of “treatment variation irrelevance,” i.e., the assumption that multiple versions of treatment may exist but they all result in the same outcome . We return to this issue in Chapter 3 but, unless otherwisespecified, we will assume treatment variation irrelevance throughout this book.Average causal effect in population:E[ 1 ] 6 E[ 0 ]contrast of interest needs to be specified. For example, “the causal effect ofaspirin” is meaningless unless we specify that the contrast of interest is, say,“taking, while alive, 150 mg of aspirin by mouth (or nasogastric tube if need be)daily for 5 years” versus “not taking aspirin.” This causal effect is well definedeven if counterfactual outcomes under other interventions are not well definedor do not exist (e.g., “taking, while alive, 500 mg of aspirin by absorptionthrough the skin daily for 5 years”).Absence of an average causal effect does not imply absence of individualeffects. Table 1.1 shows that treatment has an individual causal effect on12 members (including Zeus) of the population because, for each of these 12individuals, the value of their counterfactual outcomes 1and 0 differ.¡ 1 Of the 12 , 6 were harmedincluding Zeus 0 1 ,¡ 1 by treatment, and 6 were helped 0 1 . This equality is not an accident:The average causal effect E[ 1 ] E[ 0 ] is always equal to the averageE[ 1 0 ] of the individual causal effects 1 0 , as a differenceof averages is equal to the average of the differences. When there is no causaleffect for any individual in the population, i.e., 1 0 for all individuals

causal inference across the sciences. The authors of any Causal Inference book will have to choose which aspects of causal inference methodology they want to emphasize. The title of this introduction reflects our own choices: a book that helps scientists–especial

Related Documents:

Chapter 1 (pp. 1 -7 & 24-33) of J. Pearl, M. Glymour, and N.P. Jewell, Causal Inference in Statistics: A Primer, Wiley, 2016. Correlation Is Not Causation The gold rule of causal analysis: no causal claim can be established purely by a statistical method. . Every causal inf

CAUSAL INFERENCE AT THE TIPPING POINT Causal inference is now on the cusp of broad adoption in business. There are three main factors that are driving the emergence of causal inference as an accelerating category now. First, the limitations of current AI are becoming evident. The hype for A

Causal inference with graphical models – in small and big data 1 Outline Association is not causation How adjustment can help or harm Counterfactuals - individual-level causal effect - average causal effect Causal graphs - Graph structure, joint distribution, conditional independencies - how to esti

So a causal effect of X on Y was established, but we want more! X M Y The directed acyclic graph (DAG) above encodes assumptions. Nodes are variables, directed arrows depict causal pathways Here M is caused by X, and Y is caused by both M and X. DAGs can be useful for causal inference: clarify the assumptions taken and facilitate the discussion.

For Causal Inference We Need: 6/7/2018 dechter, class 8, 276-18 1. A working definition of “causation” 2. A method by which to formally articulate causal assumptions—that is, to create causal models 3. A method by which to link the structure of a causal model to features of data 4.

Causal inference primer 2. Causality from non-experimental data 3. Text as a control 4. Double machine learning . Text and Causal Inference: A Review of Using Text to Remove Confounding from Causal Estimates (ACL 2020) . chern.handout.pdf. Double ML Text Examples. What We’v

Statistics and Causal Inference PAUL W. HOLLAND* Problems involving causal inference have dogged at the heels of statistics since its earliest days. Correlation does not imply causation, and yet causal conclusions drawn from a carefully designed experiment are often valid. What can a statistical model say about causation? This question is ad-

American Revolution were the same white guys who controlled it after the American Revolution. And this leads us to the second, and more important way that as a revolution, the American one falls a bit short. So, if you've ever studied American history, you're probably familiar with the greatest line in the Declaration of Independence: “We hold these truths to be self-evident, that all men .