Page 21 - Read Online
P. 21

Page 2 of 22                   Ernest et al. Complex Eng Syst 2023;3:4  I http://dx.doi.org/10.20517/ces.2022.54



               1. INTRODUCTION
               Artificial intelligence (AI) applications in reinforcement learning (RL) have garnered significant attention in
               recent years [1,2]  due to their wide ranging applicability to previously intractable problems. In particular, the
               success of DeepMind’s AlphaGo system [3]  ignited a revitalization of research and attention in the area, specif-
               ically with the introduction of new techniques combining RL with deep neural networks (DNN), known as
               Deep Reinforcement Learning (DRL). However, while improvements within the overarching field of RL and
               DRL continue to increase the scalability and performance of these approaches, verification and explainability
               efforts have not received comparable attention. There have been efforts to take highly performant DRL so-
               lutions and increase explainability and trustworthiness ex post facto. An example of this was DARPA’s XAI
                                                                                            [4]
               program, introduced to study and identify the importance and usage of explainability in AI . Their conclu-
               sions were that many DRL solutions were brittle, unverifiable, and opaque to human designers/operators that
               may want to audit, verify, or extract knowledge from what the agent learned.


               Fuzzy inference systems (FIS), function approximators that utilize Fuzzy Logic and inference rules to map
                              [5]
               inputs to outputs , have several properties that lend themselves towards XAI, but have other potential draw-
               backs compared to DNNs, namely scalability. Fuzzy Logic-based systems have long been used in control
                                                                  [6]
               system development due to their approximation capabilities , ease of implementation from expert knowl-
                   [7]
                                                                                  [9]
                                           [8]
               edge , robustness to input noise , explainability and transparency to humans , and ability to be formally
               verified [10] . However, scalability issues with respect to the number of inputs limited the potential applications.
               Fuzzy trees were introduced in 2015 [11]  in order to mitigate scalability issues while also retaining explainability
               and approximation capabilities by combining multiple FISs arranged in a network or tree-like structure.
               Genetic Algorithms are a class of gradient-free search algorithms that evolve solutions by mutation and re-
               combination over a number of generations by evaluating their fitness against one or more metrics in a fitness
               function. GAs have long been used to great effect in many areas, but also in a large body of work related to
               optimization of FIS parameters [12] . Combining Fuzzy Trees with Genetic Algorithms led to Genetic Fuzzy
               Trees (GFTs) [11] , a powerful combination that uses an explainable and formally verifiable function approxima-
               tor with a gradient-free optimizer and has been applied to several complex use cases in both supervised [13]
               and reinforcement learning domains [14] . Thales’s GFT software toolkit includes both a Fuzzy Logic engine,
               Psion, and a state-of-the-art Genetic Algorithm-based optimization tool, EVE [15] . Its strengths include ease
               of utilization for finding low wall-time solutions and wide applicability due to the nature of gradient-free op-
               timization. Perhaps the most notable prior use-case was the Alpha system [14] , a super-human AI for beyond
               visual range air-to-air engagements within high-fidelity simulations against expert human pilots [14] .

               Another benefit of GFTs is that they can be verified using Formal Methods. Formal Methods are often defined
               as ”mathematically rigorous techniques for the development, specification, and verification of systems.” Many
               methods and techniques fall under the umbrella of Formal Methods including the boolean satisfiability prob-
               lem (SAT) [16] , satisfiability modulo theories (SMT), model checking, theorem proving, reachability analysis,
               etc. Formal Verification is the utilization of Formal Methods towards verifying the correctness of a system.
               In general, verification is about confidence in the correctness of a system, and formal verification extends
               traditional verification methods (e.g., Monte Carlo evaluation) towards definitive proofs of correctness. The
               adoption of formal verification has been slow in the world of AI and ML mainly due to the difficulty in proving
               properties of DNNs as their scale continues to increase.

               In this work, an AI agent using the GFT construct is created and then trained using reinforcement learning
               to play specific scenarios within StarCraft 2. Note that this study does not analyze an entire standard Starcraft
               2 match. Instead, the focus will be on specific control applications with a focus on explainability and formal
               verifiability, though an entire standard game of Starcraft 2 could certainly be studied through utilization of the
               GFT approach. This study is not aimed towards demonstrating a performance disparity between Fuzzy logic-
   16   17   18   19   20   21   22   23   24   25   26