Page 67 - Read Online
P. 67

Su et al. Intell Robot 2022;2(3):244­74  I http://dx.doi.org/10.20517/ir.2022.17    Page 260

               based on statistical correlation and half-sibling regression, which leverages the statistical dependency between
               gender-biased word vectors and gender-definition word vectors to learn the counterfactual gender informa-
               tion of an individual through causal inference. The learned spurious gender information is then subtracted
               from the gender-biased word vectors to remove the gender bias. Lu et al. [80]  proposed a method called CDA to
               eliminate gender bias through counterfactual data augmentation. The main idea of CDA is to augment the cor-
               pus by exchanging gender word pairs in the corpus and constructing matching gender word pairs with causal
               interventions. As such, CDA breaks associations between gendered and gender-neutral words and alleviates
               the problem that gender bias increases as loss decreases when training with gradient descent.


               There exists a certain degree of bias and fairness issues in word embedding, machine translation, sentiment
               analysis, language models, and dialog generation in NLP. At present, most studies only focus on a single bias
               (such as gender bias), and there is a lack of research results on other biases or eliminating multiple biases at the
               same time. Therefore, how should we analyze and evaluate the mechanism and impact of multi-bias in word
               embedding and machine learning algorithms? Establishing effective techniques for eliminating various biases
               in word embedding and machine learning algorithms requires further research which needs to be carried out
               for fair NLP.

               6.4. Fair medical
               Electronic health records (EHRe) contain large amounts of clinical information about heterogeneous patients
               and their responses to treatments. It is possible for machine learning techniques to efficiently leverage the full
               extent of EHRs to help physicians make predictions for patients, thus greatly improving the quality of care
               and reducing costs. However, because of discrimination implicitly embedded in EHRs, the automated systems
               may introduce or even aggravate the nursing gap between underrepresented groups and disadvantaged ones.
               The prior works on eliminating discrimination for clinical predictive models mostly focus on statistics-based
               fairness-enhancing approaches [81,82] . In addition, they do not provide an effective evaluation of fairness to
               individuals, and the fairness metrics they used are difficult to verify. Some recent studies focus on assessing
               fairness for clinical predictive models from a causal perspective [83,84] . For example, Pfohl et al. [83] proposed a
               counterfactual fairness notion to extend fairness to the individual level and leveraged variational autoencoder
               technology to eliminate discrimination against certain patients.

               6.5. Causal analysis packages
               This section introduces some representative packages or software for causal analysis, which are helpful for us
               to develop causality-based fairness-enhancing approaches. These packages can be roughly divided into two
               categories: one for discovering potential causal structure in data and the other for making causal inferences.
               Table 2 summarizes typical packages or software for causal analysis.

               TETRAD  [85]  is a full-featured software for causal analysis after considerable development where it can be used
               to discover the causal structure behind the dataset, estimate the causal effects, simulate the causal models,
               etc. TETRAD can accept different types of data as input, e.g., discrete data, continuous data, time series data,
               etc. The users can choose the appropriate well-tested causal discovery algorithms it integrates to search causal
               structure, as well as input prior causal knowledge to limit the search. In addition, TETRAD can parameterize
               the causal model and simulate the data according to the existing causal diagram. Causal-learn package [86]  is
               the Python version of TETRAD. It provides the implementation of the latest causal discovery methods rang-
               ing from constraint-based methods, score-based methods, and constrained functional causal models-based
               methods to permutation-based methods. In addition, there are many packages for causal discovery [87–89] .
               Tigramite [88]  focuses on searching causal structure from observational time series data. In addition to provid-
               ing classic causal discovery algorithms, gCastle [89]  provides many gradient-based causal discovery approaches.


               CausalML [90]  is a Python package which encapsulates many causal learning and causal inference approaches.
   62   63   64   65   66   67   68   69   70   71   72