Page 56 - Read Online
P. 56

Page 249                          Su et al. Intell Robot 2022;2(3):244­74  I http://dx.doi.org/10.20517/ir.2022.17

                                  U  S
                                                            U

                          U  X     S      U Y        X              Y         X             Y

                                                                                     S
                           X              Y                 S                    do(S=s)


                                   (a)                      (b)                      (c)

               Figure 3. (a) An example causal graph based on Markovian assumption; (b) a causal graph based on semi-Markovian assumption; and (c)
               a causal graph after performing an intervention on   .


               pointing to variable   . Figure 3(c) shows the causal diagram after the intervention     (  ). The mathematical
               meaning of     (  ) is defined as the substitution of equation    =       (    (  ),      ) with    =   . For another
               endogenous variable    which is affected by the intervention, its post-intervention distribution under     (  )
               is denoted by   (  |    (  )) or   (      ) for short. Intuitively,   (  |   =   ) represents the population distribution
               of    condition on observing attribute    value of individuals is   , while   (  |    (   =   )) (i.e.,   (  |    (  )))
               represents the population distribution of    if everyone in the population had their    value fixed at   . This
               post-intervention distribution   (  |    (  )) is considered a counterfactual distribution since the intervention
                   (  ) forces    to take a certain value different from the one it would take in the actual world. For exam-
               ple, if    represents sex (   , male;    , female) and    represents the hiring decision (   , hired;    , not hired),
                                                                                      +
                                    +
                                            −
                                                                                               −
                 (      |   =    ,    =    ) involves two worlds: a real world that a male applicant has been hired and a counterfac-
                         +
                               +
                   −
               tual world where the same applicant is female. Such expression means that, when a job-hunter whose gender
               is male has been observed to be hired, what is the probability that the same job-hunter would still be hired if
               this job-hunter were female.
               Causality-based fairness notions aim to tell whether the outcome of a decision made by the AI decision model
               is discriminative, which are expressed by interventional and counterfactual probability distributions. The ap-
               plication of the causality-based fairness notions not only requires a dataset D as input but also relies on a
               causal graph G. Causal approaches aim to limit the causal effects of sensitive attributes on decisions, which
               are computed by interventional and counterfactual probabilities. However, since these probabilities cannot be
               observed, they fail to be uniquely assessed from D and G in some cases, which is called the unidentifiable issue.
               In other words, if two variables have different causal effect measures resulting from different causal models
               that all agree with the observational distribution, the causal effects are unidentifiable.




               4. CAUSALITY-BASED FAIRNESS NOTIONS
               Pearl defined the causality as three rungs: correlation, intervention, and counterfactual [25] . The first rung
               (correlation) reflects the ability of observation, which aims to discover the patterns in the environment. The
               secondrung(intervention)reflectstheabilityofaction,whichreferstothepredictionoftheresultsofdeliberate
               changes to the environment. The third rung (counterfactual) refers to the ability to imagine the counterfactual
               world and speculate on the causes of the observed phenomena. The second and third rungs aim to expose
               the root causes of the patterns that we observe. Thus far, many fairness metrics have been proposed and all of
               them can be placed in such causal rungs. Figure 4 presents categorization of fairness notions. Early definitions
               of fairness are based on statistical correlations, all of which can be found at the first rung. All causality-based
               fairness notions can be found at the second and third rungs, each of which considers the mechanism of data
               generation. Thus, causality-based fairness notions can better reveal the causes of discrimination than statistics-
               based ones and have been attracting more and more attention.
   51   52   53   54   55   56   57   58   59   60   61