Learning Out Loud - (LOL)

Unpacked: Double Machine Learning

You’ve all heard about Machine Learning. Some say it’s amaying, I say it’s meh. I would rather double it and give it to the next person which, happend to be you (see the pun I just did ?). You guessed it today we will talk about Double Machine Learning.

NB: For now we will focus on how Double ML is used to compute ATE. CATE may come later.

Introduction

Before diving into this subject we need to talk about causation and causal inference.

We will start with a very basic example:

Say we have two populations PDP_D and PCP_C where people in PDP_D were given a drug to increase their heights and people in PCP_C were not. A few weeks lates, we measure the average height (HDH_D resp. HCH_C) of each population and find out that HDH_D > HCH_C. Wait what ? Does this mean that the drug works ? Let’s rewind a bit.

We travel back in time a few weeks before people in PDP_D were given the drug, measure their initial height HDH_D’, and compare it to HCH_C. We make a shocking discovery: HDH_D’ > HCH_C.

Our initial thought was equivalent to the following computation:

E[YT=1]E[YT=0] E[Y | T = 1] - E[Y | T = 0]

If we expand this calcul we get

=E[PDT=1]E[PCT=0]=E[PDT=1]E[PCT=0]+E[PCT=1]E[PCT=1]=E[PDPCT=1]ATT+E[PCT=1]E[PCT=0]Bias \begin{align*} &= E[P_D | T = 1] - E[P_C | T = 0] \\ &= E[P_D | T = 1] - E[P_C | T = 0] + E[P_C | T = 1] - E[P_C | T = 1] \\ &= \underbrace{E[P_D - P_C | T = 1]}_\text{ATT} + \underbrace{E[P_C | T = 1] - E[P_C | T = 0]}_\text{Bias} \end{align*}