Page 8 - Read Online
P. 8

Page 4 of 12                 Sathyan et al. Complex Eng Syst 2022;2:18  I http://dx.doi.org/10.20517/ces.2022.41


                              [2]
               recommendations . This in turn helps the patients to trust the model results and system recommendations.
               This can also increase the probability that the patient will accept and follow the recommended medical treat-
               ment. Moreover, XAI can decrease the probability of error in the diagnostic process since it helps clinicians to
               focus on the relevant data and help them to better understand the model recommendations.


               XAI is an evolving field. As mentioned before, at this current stage, even state-of-the-art XAI algorithms
               have disadvantages. In literature, we can find approaches that aim to improve some aspects. One of the main
               challenges of using XAI in healthcare environments is the need to remain neutral regarding preferences. We
               canfindabonafideapproachcalledscientificexplanationinAI(sXAI)thatcanbeusedinthefieldofmedicine
               and healthcare [19] . An additional approach based on integrated Electronic Medical Records (EMR) medical
               systemsisdescribedin [20] . Theapproachfocusesonexplainabilityandinteroperabilityfromthehumanaspect.
               Ensemble of machine learning (ML) can also increase the level of interpretability, as can be seen in [21] . In [21] ,
               theauthoruseensembleofMLforlogicdrivingofanthropometricmeasurementsinfluencingbodymassindex
               (BMI). Additional evidence for the implementations of several XAI models is mentioned in [22] . The paper
               showshowintegratingXAImodelshelpstoincreasethepersuasiveandcoherencelevelsinthedecisionmaking
               of clinicians and medical professionals teams. The usage of XAI has shown an improvement in transparency
               and reliability in the field of neuroscience field [23] .


               In this paper, we apply some XAI concepts to a use case applicable to the medical field. Our work focus on XAI
               implementation for breast cancer diagnostics. Our research uses the commonly researched UCI breast cancer
               dataset. We focus on breast cancer since it is the most common type of cancer amongst women [24] . The usage
               of XAI for diagnostics and prediction of breast cancer can impact and help a large number of patients. The
               UCI breast cancer dataset includes 569 data points [25] . Each data point consists of 32 attributes that include
               the ID number, the diagnosis, and 30 features used as predictors in this work. The 30 predictors include the
               mean, standard deviation and the mean of 3 largest values of 10 features: (1) radius (mean of distances from
               center to points on the perimeter); (2) texture (standard deviation of gray-scale values); (3) perimeter; (4) area;
               (5) smoothness; (6) compactness; (7) concavity; (8) concave points; (9) symmetry; and (10) fractal dimension.



               3. METHODOLOGY
               3.1. LIME
               LIME is one of the methodologies that is used to explain the predictions made by machine learning classifier
               models [26] . It can explain individual predictions made by text classifiers as well as classifiers that are modeled
               on tabular data.

               In this work, we are focusing on using LIME to explain decisions made by a neural network classifier that
               works on tabular dataset. The process of LIME to explain individual predictions are as follows:
               1. For each instance that needs to be explained, LIME perturbs the observation    times.
               2. For tabular data, the statistics for each variable in the data are evaluated.
               3. The permutations are then sampled from the variable distributions within the neighborhood of the original
                  data point for which an explanation is being sought.
               4. In our case, the original model is a neural network. The trained neural network model is used to predict
                  the outcome of all permuted observations.
               5. Calculate the distance from the perturbed points to the original observation and then convert it to a simi-
                  larity score.
               6. Select    features best describing the original model outcome for the perturbed data.
               7. Fit a simple model (linear model) on the perturbed data, explaining the original model outcome with the
                     features from the permuted data weighted by its similarity to the original observation.
               8. Extract the feature weights from the simple model and use these as explanations.
   3   4   5   6   7   8   9   10   11   12   13