Explainable AI with Python

Explainable AI with Python
Author: Leonida Gianfagna
Publisher: Springer Nature
Total Pages: 202
Release: 2021-04-28
Genre: Computers
ISBN: 303068640X

This book provides a full presentation of the current concepts and available techniques to make “machine learning” systems more explainable. The approaches presented can be applied to almost all the current “machine learning” models: linear and logistic regression, deep learning neural networks, natural language processing and image recognition, among the others. Progress in Machine Learning is increasing the use of artificial agents to perform critical tasks previously handled by humans (healthcare, legal and finance, among others). While the principles that guide the design of these agents are understood, most of the current deep-learning models are "opaque" to human understanding. Explainable AI with Python fills the current gap in literature on this emerging topic by taking both a theoretical and a practical perspective, making the reader quickly capable of working with tools and code for Explainable AI. Beginning with examples of what Explainable AI (XAI) is and why it is needed in the field, the book details different approaches to XAI depending on specific context and need. Hands-on work on interpretable models with specific examples leveraging Python are then presented, showing how intrinsic interpretable models can be interpreted and how to produce “human understandable” explanations. Model-agnostic methods for XAI are shown to produce explanations without relying on ML models internals that are “opaque.” Using examples from Computer Vision, the authors then look at explainable models for Deep Learning and prospective methods for the future. Taking a practical perspective, the authors demonstrate how to effectively use ML and XAI in science. The final chapter explains Adversarial Machine Learning and how to do XAI with adversarial examples.

Hands-On Explainable AI (XAI) with Python

Hands-On Explainable AI (XAI) with Python
Author: Denis Rothman
Publisher: Packt Publishing Ltd
Total Pages: 455
Release: 2020-07-31
Genre: Computers
ISBN: 1800202768

Resolve the black box models in your AI applications to make them fair, trustworthy, and secure. Familiarize yourself with the basic principles and tools to deploy Explainable AI (XAI) into your apps and reporting interfaces. Key FeaturesLearn explainable AI tools and techniques to process trustworthy AI resultsUnderstand how to detect, handle, and avoid common issues with AI ethics and biasIntegrate fair AI into popular apps and reporting tools to deliver business value using Python and associated toolsBook Description Effectively translating AI insights to business stakeholders requires careful planning, design, and visualization choices. Describing the problem, the model, and the relationships among variables and their findings are often subtle, surprising, and technically complex. Hands-On Explainable AI (XAI) with Python will see you work with specific hands-on machine learning Python projects that are strategically arranged to enhance your grasp on AI results analysis. You will be building models, interpreting results with visualizations, and integrating XAI reporting tools and different applications. You will build XAI solutions in Python, TensorFlow 2, Google Cloud’s XAI platform, Google Colaboratory, and other frameworks to open up the black box of machine learning models. The book will introduce you to several open-source XAI tools for Python that can be used throughout the machine learning project life cycle. You will learn how to explore machine learning model results, review key influencing variables and variable relationships, detect and handle bias and ethics issues, and integrate predictions using Python along with supporting the visualization of machine learning models into user explainable interfaces. By the end of this AI book, you will possess an in-depth understanding of the core concepts of XAI. What you will learnPlan for XAI through the different stages of the machine learning life cycleEstimate the strengths and weaknesses of popular open-source XAI applicationsExamine how to detect and handle bias issues in machine learning dataReview ethics considerations and tools to address common problems in machine learning dataShare XAI design and visualization best practicesIntegrate explainable AI results using Python modelsUse XAI toolkits for Python in machine learning life cycles to solve business problemsWho this book is for This book is not an introduction to Python programming or machine learning concepts. You must have some foundational knowledge and/or experience with machine learning libraries such as scikit-learn to make the most out of this book. Some of the potential readers of this book include: Professionals who already use Python for as data science, machine learning, research, and analysisData analysts and data scientists who want an introduction into explainable AI tools and techniquesAI Project managers who must face the contractual and legal obligations of AI Explainability for the acceptance phase of their applications

Practical Explainable AI Using Python

Practical Explainable AI Using Python
Author: Pradeepta Mishra
Publisher: Apress
Total Pages: 344
Release: 2021-12-15
Genre: Computers
ISBN: 9781484271575

Learn the ins and outs of decisions, biases, and reliability of AI algorithms and how to make sense of these predictions. This book explores the so-called black-box models to boost the adaptability, interpretability, and explainability of the decisions made by AI algorithms using frameworks such as Python XAI libraries, TensorFlow 2.0+, Keras, and custom frameworks using Python wrappers. You'll begin with an introduction to model explainability and interpretability basics, ethical consideration, and biases in predictions generated by AI models. Next, you'll look at methods and systems to interpret linear, non-linear, and time-series models used in AI. The book will also cover topics ranging from interpreting to understanding how an AI algorithm makes a decision Further, you will learn the most complex ensemble models, explainability, and interpretability using frameworks such as Lime, SHAP, Skater, ELI5, etc. Moving forward, you will be introduced to model explainability for unstructured data, classification problems, and natural language processing–related tasks. Additionally, the book looks at counterfactual explanations for AI models. Practical Explainable AI Using Python shines the light on deep learning models, rule-based expert systems, and computer vision tasks using various XAI frameworks. What You'll Learn Review the different ways of making an AI model interpretable and explainable Examine the biasness and good ethical practices of AI models Quantify, visualize, and estimate reliability of AI models Design frameworks to unbox the black-box models Assess the fairness of AI models Understand the building blocks of trust in AI models Increase the level of AI adoption Who This Book Is For AI engineers, data scientists, and software developers involved in driving AI projects/ AI products.

Interpretable Machine Learning

Interpretable Machine Learning
Author: Christoph Molnar
Publisher: Lulu.com
Total Pages: 320
Release: 2020
Genre: Artificial intelligence
ISBN: 0244768528

This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.

Interpretable Machine Learning with Python

Interpretable Machine Learning with Python
Author: Serg Masís
Publisher: Packt Publishing Ltd
Total Pages: 737
Release: 2021-03-26
Genre: Computers
ISBN: 1800206577

A deep and detailed dive into the key aspects and challenges of machine learning interpretability, complete with the know-how on how to overcome and leverage them to build fairer, safer, and more reliable models Key Features Learn how to extract easy-to-understand insights from any machine learning model Become well-versed with interpretability techniques to build fairer, safer, and more reliable models Mitigate risks in AI systems before they have broader implications by learning how to debug black-box models Book DescriptionDo you want to gain a deeper understanding of your models and better mitigate poor prediction risks associated with machine learning interpretation? If so, then Interpretable Machine Learning with Python deserves a place on your bookshelf. We’ll be starting off with the fundamentals of interpretability, its relevance in business, and exploring its key aspects and challenges. As you progress through the chapters, you'll then focus on how white-box models work, compare them to black-box and glass-box models, and examine their trade-off. You’ll also get you up to speed with a vast array of interpretation methods, also known as Explainable AI (XAI) methods, and how to apply them to different use cases, be it for classification or regression, for tabular, time-series, image or text. In addition to the step-by-step code, this book will also help you interpret model outcomes using examples. You’ll get hands-on with tuning models and training data for interpretability by reducing complexity, mitigating bias, placing guardrails, and enhancing reliability. The methods you’ll explore here range from state-of-the-art feature selection and dataset debiasing methods to monotonic constraints and adversarial retraining. By the end of this book, you'll be able to understand ML models better and enhance them through interpretability tuning. What you will learn Recognize the importance of interpretability in business Study models that are intrinsically interpretable such as linear models, decision trees, and Naïve Bayes Become well-versed in interpreting models with model-agnostic methods Visualize how an image classifier works and what it learns Understand how to mitigate the influence of bias in datasets Discover how to make models more reliable with adversarial robustness Use monotonic constraints to make fairer and safer models Who this book is for This book is primarily written for data scientists, machine learning developers, and data stewards who find themselves under increasing pressures to explain the workings of AI systems, their impacts on decision making, and how they identify and manage bias. It’s also a useful resource for self-taught ML enthusiasts and beginners who want to go deeper into the subject matter, though a solid grasp on the Python programming language and ML fundamentals is needed to follow along.

Explanatory Model Analysis

Explanatory Model Analysis
Author: Przemyslaw Biecek
Publisher: CRC Press
Total Pages: 312
Release: 2021-02-15
Genre: Business & Economics
ISBN: 0429651376

Explanatory Model Analysis Explore, Explain and Examine Predictive Models is a set of methods and tools designed to build better predictive models and to monitor their behaviour in a changing environment. Today, the true bottleneck in predictive modelling is neither the lack of data, nor the lack of computational power, nor inadequate algorithms, nor the lack of flexible models. It is the lack of tools for model exploration (extraction of relationships learned by the model), model explanation (understanding the key factors influencing model decisions) and model examination (identification of model weaknesses and evaluation of model's performance). This book presents a collection of model agnostic methods that may be used for any black-box model together with real-world applications to classification and regression problems.

Explainable AI: Interpreting, Explaining and Visualizing Deep Learning

Explainable AI: Interpreting, Explaining and Visualizing Deep Learning
Author: Wojciech Samek
Publisher: Springer Nature
Total Pages: 435
Release: 2019-09-10
Genre: Computers
ISBN: 3030289540

The development of “intelligent” systems that can take decisions and perform autonomously might lead to faster and more consistent decisions. A limiting factor for a broader adoption of AI technology is the inherent risks that come with giving up human control and oversight to “intelligent” machines. For sensitive tasks involving critical infrastructures and affecting human well-being or health, it is crucial to limit the possibility of improper, non-robust and unsafe decisions and actions. Before deploying an AI system, we see a strong need to validate its behavior, and thus establish guarantees that it will continue to perform as expected when deployed in a real-world environment. In pursuit of that objective, ways for humans to verify the agreement between the AI decision structure and their own ground-truth knowledge have been explored. Explainable AI (XAI) has developed as a subfield of AI, focused on exposing complex AI models to humans in a systematic and interpretable manner. The 22 chapters included in this book provide a timely snapshot of algorithms, theory, and applications of interpretable and explainable AI and AI techniques that have been proposed recently reflecting the current discourse in this field and providing directions of future development. The book is organized in six parts: towards AI transparency; methods for interpreting AI systems; explaining the decisions of AI systems; evaluating interpretability and explanations; applications of explainable AI; and software for explainable AI.

Machine Learning for Time Series Forecasting with Python

Machine Learning for Time Series Forecasting with Python
Author: Francesca Lazzeri
Publisher: John Wiley & Sons
Total Pages: 224
Release: 2020-12-03
Genre: Computers
ISBN: 111968238X

Learn how to apply the principles of machine learning to time series modeling with this indispensable resource Machine Learning for Time Series Forecasting with Python is an incisive and straightforward examination of one of the most crucial elements of decision-making in finance, marketing, education, and healthcare: time series modeling. Despite the centrality of time series forecasting, few business analysts are familiar with the power or utility of applying machine learning to time series modeling. Author Francesca Lazzeri, a distinguished machine learning scientist and economist, corrects that deficiency by providing readers with comprehensive and approachable explanation and treatment of the application of machine learning to time series forecasting. Written for readers who have little to no experience in time series forecasting or machine learning, the book comprehensively covers all the topics necessary to: Understand time series forecasting concepts, such as stationarity, horizon, trend, and seasonality Prepare time series data for modeling Evaluate time series forecasting models’ performance and accuracy Understand when to use neural networks instead of traditional time series models in time series forecasting Machine Learning for Time Series Forecasting with Python is full real-world examples, resources and concrete strategies to help readers explore and transform data and develop usable, practical time series forecasts. Perfect for entry-level data scientists, business analysts, developers, and researchers, this book is an invaluable and indispensable guide to the fundamental and advanced concepts of machine learning applied to time series modeling.

T-Minus AI

T-Minus AI
Author: Michael Kanaan
Publisher: BenBella Books
Total Pages: 249
Release: 2020-08-25
Genre: Science
ISBN: 1950665135

Late in 2017, the global significance of the conversation about artificial intelligence (AI) changed forever. China put the world on alert when it released a plan to dominate all aspects of AI across the planet. Only weeks later, Vladimir Putin raised a Russian red flag in response by declaring AI the future for all humankind, and proclaiming that, "Whoever becomes the leader in this sphere will become the ruler of the world." The race was on. Consistent with their unique national agendas, countries throughout the world began plotting their paths and hurrying their pace. Now, not long after, the race has become a sprint. Despite everything at stake, to most of us AI remains shrouded by a cloud of mystery and misunderstanding. Hidden behind complicated and technical jargon and confused by fantastical depictions of science fiction, the modern realities of AI and its profound implications are hard to decipher, but crucial to recognize. In T-Minus AI: Humanity's Countdown to Artificial Intelligence and the New Pursuit of Global Power, author Michael Kanaan explains AI from a human-oriented perspective we can all finally understand. A recognized national expert and the U.S. Air Force's first Chairperson for Artificial Intelligence, Kanaan weaves a compelling new view on our history of innovation and technology to masterfully explain what each of us should know about modern computing, AI, and machine learning. Kanaan also dives into the global implications of AI by illuminating the cultural and national vulnerabilities already exposed and the pressing issues now squarely on the table. AI has already become China's all-purpose tool to impose its authoritarian influence around the world. Russia, playing catch up, is weaponizing AI through its military systems and now infamous, aggressive efforts to disrupt democracy by whatever disinformation means possible. America and like-minded nations are awakening to these new realities—and the paths they're electing to follow echo loudly the political foundations and, in most cases, the moral imperatives upon which they were formed. As we march toward a future far different than ever imagined, T-Minus AI is fascinating and crucially well-timed. It leaves the fiction behind, paints the alarming implications of AI for what they actually are, and calls for unified action to protect fundamental human rights and dignities for all.