Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

Author: Wojciech Samek

Publisher: Springer Nature

Published: 2019-09-10

Total Pages: 435

ISBN-13: 3030289540

DOWNLOAD EBOOK

The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner. The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.


Interpretable Machine Learning

Interpretable Machine Learning

Author: Christoph Molnar

Publisher: Lulu.com

Published: 2020

Total Pages: 320

ISBN-13: 0244768528

DOWNLOAD EBOOK

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.


Explainable and Interpretable Models in Computer Vision and Machine Learning

Explainable and Interpretable Models in Computer Vision and Machine Learning

Author: Hugo Jair Escalante

Publisher: Springer

Published: 2018-11-29

Total Pages: 305

ISBN-13: 3319981315

DOWNLOAD EBOOK

This book compiles leading research on the development of explainable and interpretable machine learning methods in the context of computer vision and machine learning. Research progress in computer vision and pattern recognition has led to a variety of modeling techniques with almost human-like performance. Although these models have obtained astounding results, they are limited in their explainability and interpretability: what is the rationale behind the decision made? what in the model structure explains its functioning? Hence, while good performance is a critical required characteristic for learning machines, explainability and interpretability capabilities are needed to take learning machines to the next step to include them in decision support systems involving human supervision. This book, written by leading international researchers, addresses key topics of explainability and interpretability, including the following: · Evaluation and Generalization in Interpretable Machine Learning · Explanation Methods in Deep Learning · Learning Functional Causal Models with Generative Neural Networks · Learning Interpreatable Rules for Multi-Label Classification · Structuring Neural Networks for More Explainable Predictions · Generating Post Hoc Rationales of Deep Visual Classification Decisions · Ensembling Visual Explanations · Explainable Deep Driving by Visualizing Causal Attention · Interdisciplinary Perspective on Algorithmic Job Candidate Search · Multimodal Personality Trait Analysis for Explainable Modeling of Job Interview Decisions · Inherent Explainability Pattern Theory-based Video Event Interpretations


Hands-On Explainable AI (XAI) with Python

Hands-On Explainable AI (XAI) with Python

Author: Denis Rothman

Publisher: Packt Publishing Ltd

Published: 2020-07-31

Total Pages: 455

ISBN-13: 1800202768

DOWNLOAD EBOOK

Resolve the black box models in your AI applications to make them fair, trustworthy, and secure. Familiarize yourself with the basic principles and tools to deploy Explainable AI (XAI) into your apps and reporting interfaces. Key FeaturesLearn explainable AI tools and techniques to process trustworthy AI resultsUnderstand how to detect, handle, and avoid common issues with AI ethics and biasIntegrate fair AI into popular apps and reporting tools to deliver business value using Python and associated toolsBook Description Effectively translating AI insights to business stakeholders requires careful planning, design, and visualization choices. Describing the problem, the model, and the relationships among variables and their findings are often subtle, surprising, and technically complex. Hands-On Explainable AI (XAI) with Python will see you work with specific hands-on machine learning Python projects that are strategically arranged to enhance your grasp on AI results analysis. You will be building models, interpreting results with visualizations, and integrating XAI reporting tools and different applications. You will build XAI solutions in Python, TensorFlow 2, Google Cloud’s XAI platform, Google Colaboratory, and other frameworks to open up the black box of machine learning models. The book will introduce you to several open-source XAI tools for Python that can be used throughout the machine learning project life cycle. You will learn how to explore machine learning model results, review key influencing variables and variable relationships, detect and handle bias and ethics issues, and integrate predictions using Python along with supporting the visualization of machine learning models into user explainable interfaces. By the end of this AI book, you will possess an in-depth understanding of the core concepts of XAI. What you will learnPlan for XAI through the different stages of the machine learning life cycleEstimate the strengths and weaknesses of popular open-source XAI applicationsExamine how to detect and handle bias issues in machine learning dataReview ethics considerations and tools to address common problems in machine learning dataShare XAI design and visualization best practicesIntegrate explainable AI results using Python modelsUse XAI toolkits for Python in machine learning life cycles to solve business problemsWho this book is for This book is not an introduction to Python programming or machine learning concepts. You must have some foundational knowledge and/or experience with machine learning libraries such as scikit-learn to make the most out of this book. Some of the potential readers of this book include: Professionals who already use Python for as data science, machine learning, research, and analysisData analysts and data scientists who want an introduction into explainable AI tools and techniquesAI Project managers who must face the contractual and legal obligations of AI Explainability for the acceptance phase of their applications


Explainable AI with Python

Explainable AI with Python

Author: Leonida Gianfagna

Publisher: Springer Nature

Published: 2021-04-28

Total Pages: 202

ISBN-13: 303068640X

DOWNLOAD EBOOK

This book provides a full presentation of the current concepts and available techniques to make “machine learning” systems more explainable. The approaches presented can be applied to almost all the current “machine learning” models: linear and logistic regression, deep learning neural networks, natural language processing and image recognition, among the others. Progress in Machine Learning is increasing the use of artificial agents to perform critical tasks previously handled by humans (healthcare, legal and finance, among others). While the principles that guide the design of these agents are understood, most of the current deep-learning models are "opaque" to human understanding. Explainable AI with Python fills the current gap in literature on this emerging topic by taking both a theoretical and a practical perspective, making the reader quickly capable of working with tools and code for Explainable AI. Beginning with examples of what Explainable AI (XAI) is and why it is needed in the field, the book details different approaches to XAI depending on specific context and need. Hands-on work on interpretable models with specific examples leveraging Python are then presented, showing how intrinsic interpretable models can be interpreted and how to produce “human understandable” explanations. Model-agnostic methods for XAI are shown to produce explanations without relying on ML models internals that are “opaque.” Using examples from Computer Vision, the authors then look at explainable models for Deep Learning and prospective methods for the future. Taking a practical perspective, the authors demonstrate how to effectively use ML and XAI in science. The final chapter explains Adversarial Machine Learning and how to do XAI with adversarial examples.


Interpretable Machine Learning with Python

Interpretable Machine Learning with Python

Author: Serg Masís

Publisher: Packt Publishing Ltd

Published: 2021-03-26

Total Pages: 737

ISBN-13: 1800206577

DOWNLOAD EBOOK

A deep and detailed dive into the key aspects and challenges of machine learning interpretability, complete with the know-how on how to overcome and leverage them to build fairer, safer, and more reliable models Key Features Learn how to extract easy-to-understand insights from any machine learning model Become well-versed with interpretability techniques to build fairer, safer, and more reliable models Mitigate risks in AI systems before they have broader implications by learning how to debug black-box models Book DescriptionDo you want to gain a deeper understanding of your models and better mitigate poor prediction risks associated with machine learning interpretation? If so, then Interpretable Machine Learning with Python deserves a place on your bookshelf. We’ll be starting off with the fundamentals of interpretability, its relevance in business, and exploring its key aspects and challenges. As you progress through the chapters, you'll then focus on how white-box models work, compare them to black-box and glass-box models, and examine their trade-off. You’ll also get you up to speed with a vast array of interpretation methods, also known as Explainable AI (XAI) methods, and how to apply them to different use cases, be it for classification or regression, for tabular, time-series, image or text. In addition to the step-by-step code, this book will also help you interpret model outcomes using examples. You’ll get hands-on with tuning models and training data for interpretability by reducing complexity, mitigating bias, placing guardrails, and enhancing reliability. The methods you’ll explore here range from state-of-the-art feature selection and dataset debiasing methods to monotonic constraints and adversarial retraining. By the end of this book, you'll be able to understand ML models better and enhance them through interpretability tuning. What you will learn Recognize the importance of interpretability in business Study models that are intrinsically interpretable such as linear models, decision trees, and Naïve Bayes Become well-versed in interpreting models with model-agnostic methods Visualize how an image classifier works and what it learns Understand how to mitigate the influence of bias in datasets Discover how to make models more reliable with adversarial robustness Use monotonic constraints to make fairer and safer models Who this book is for This book is primarily written for data scientists, machine learning developers, and data stewards who find themselves under increasing pressures to explain the workings of AI systems, their impacts on decision making, and how they identify and manage bias. It’s also a useful resource for self-taught ML enthusiasts and beginners who want to go deeper into the subject matter, though a solid grasp on the Python programming language and ML fundamentals is needed to follow along.


Learning Deep Learning

Learning Deep Learning

Author: Magnus Ekman

Publisher: Addison-Wesley Professional

Published: 2021-07-19

Total Pages: 1106

ISBN-13: 0137470290

DOWNLOAD EBOOK

NVIDIA's Full-Color Guide to Deep Learning: All You Need to Get Started and Get Results "To enable everyone to be part of this historic revolution requires the democratization of AI knowledge and resources. This book is timely and relevant towards accomplishing these lofty goals." -- From the foreword by Dr. Anima Anandkumar, Bren Professor, Caltech, and Director of ML Research, NVIDIA "Ekman uses a learning technique that in our experience has proven pivotal to success—asking the reader to think about using DL techniques in practice. His straightforward approach is refreshing, and he permits the reader to dream, just a bit, about where DL may yet take us." -- From the foreword by Dr. Craig Clawson, Director, NVIDIA Deep Learning Institute Deep learning (DL) is a key component of today's exciting advances in machine learning and artificial intelligence. Learning Deep Learning is a complete guide to DL. Illuminating both the core concepts and the hands-on programming techniques needed to succeed, this book is ideal for developers, data scientists, analysts, and others--including those with no prior machine learning or statistics experience. After introducing the essential building blocks of deep neural networks, such as artificial neurons and fully connected, convolutional, and recurrent layers, Magnus Ekman shows how to use them to build advanced architectures, including the Transformer. He describes how these concepts are used to build modern networks for computer vision and natural language processing (NLP), including Mask R-CNN, GPT, and BERT. And he explains how a natural language translator and a system generating natural language descriptions of images. Throughout, Ekman provides concise, well-annotated code examples using TensorFlow with Keras. Corresponding PyTorch examples are provided online, and the book thereby covers the two dominating Python libraries for DL used in industry and academia. He concludes with an introduction to neural architecture search (NAS), exploring important ethical issues and providing resources for further learning. Explore and master core concepts: perceptrons, gradient-based learning, sigmoid neurons, and back propagation See how DL frameworks make it easier to develop more complicated and useful neural networks Discover how convolutional neural networks (CNNs) revolutionize image classification and analysis Apply recurrent neural networks (RNNs) and long short-term memory (LSTM) to text and other variable-length sequences Master NLP with sequence-to-sequence networks and the Transformer architecture Build applications for natural language translation and image captioning NVIDIA's invention of the GPU sparked the PC gaming market. The company's pioneering work in accelerated computing--a supercharged form of computing at the intersection of computer graphics, high-performance computing, and AI--is reshaping trillion-dollar industries, such as transportation, healthcare, and manufacturing, and fueling the growth of many others. Register your book for convenient access to downloads, updates, and/or corrections as they become available. See inside book for details.


Rule Extraction from Support Vector Machines

Rule Extraction from Support Vector Machines

Author: Joachim Diederich

Publisher: Springer

Published: 2007-12-27

Total Pages: 267

ISBN-13: 3540753907

DOWNLOAD EBOOK

Support vector machines (SVMs) are one of the most active research areas in machine learning. SVMs have shown good performance in a number of applications, including text and image classification. However, the learning capability of SVMs comes at a cost – an inherent inability to explain in a comprehensible form, the process by which a learning result was reached. Hence, the situation is similar to neural networks, where the apparent lack of an explanation capability has led to various approaches aiming at extracting symbolic rules from neural networks. For SVMs to gain a wider degree of acceptance in fields such as medical diagnosis and security sensitive areas, it is desirable to offer an explanation capability. User explanation is often a legal requirement, because it is necessary to explain how a decision was reached or why it was made. This book provides an overview of the field and introduces a number of different approaches to extracting rules from support vector machines developed by key researchers. In addition, successful applications are outlined and future research opportunities are discussed. The book is an important reference for researchers and graduate students, and since it provides an introduction to the topic, it will be important in the classroom as well. Because of the significance of both SVMs and user explanation, the book is of relevance to data mining practitioners and data analysts.


Explainable AI in Healthcare and Medicine

Explainable AI in Healthcare and Medicine

Author: Arash Shaban-Nejad

Publisher: Springer Nature

Published: 2020-11-02

Total Pages: 351

ISBN-13: 3030533522

DOWNLOAD EBOOK

This book highlights the latest advances in the application of artificial intelligence and data science in health care and medicine. Featuring selected papers from the 2020 Health Intelligence Workshop, held as part of the Association for the Advancement of Artificial Intelligence (AAAI) Annual Conference, it offers an overview of the issues, challenges, and opportunities in the field, along with the latest research findings. Discussing a wide range of practical applications, it makes the emerging topics of digital health and explainable AI in health care and medicine accessible to a broad readership. The availability of explainable and interpretable models is a first step toward building a culture of transparency and accountability in health care. As such, this book provides information for scientists, researchers, students, industry professionals, public health agencies, and NGOs interested in the theory and practice of computational models of public and personalized health intelligence.


Embedded Systems and Artificial Intelligence

Embedded Systems and Artificial Intelligence

Author: Vikrant Bhateja

Publisher: Springer Nature

Published: 2020-04-07

Total Pages: 880

ISBN-13: 9811509476

DOWNLOAD EBOOK

This book gathers selected research papers presented at the First International Conference on Embedded Systems and Artificial Intelligence (ESAI 2019), held at Sidi Mohamed Ben Abdellah University, Fez, Morocco, on 2–3 May 2019. Highlighting the latest innovations in Computer Science, Artificial Intelligence, Information Technologies, and Embedded Systems, the respective papers will encourage and inspire researchers, industry professionals, and policymakers to put these methods into practice.