|
![]() |
|||
|
||||
OverviewCreate interpretable AI models for transparent and explainable anomaly detection with this hands-on guide Purchase of the print or Kindle book includes a free PDF eBook Key Features Build auditable XAI models for replicability and regulatory compliance Derive critical insights from transparent anomaly detection models Strike the right balance between model accuracy and interpretability Book DescriptionDespite promising advances, the opaque nature of deep learning models makes it difficult to interpret them, which is a drawback in terms of their practical deployment and regulatory compliance. Deep Learning and XAI Techniques for Anomaly Detection shows you state-of-the-art methods that’ll help you to understand and address these challenges. By leveraging the Explainable AI (XAI) and deep learning techniques described in this book, you’ll discover how to successfully extract business-critical insights while ensuring fair and ethical analysis. This practical guide will provide you with tools and best practices to achieve transparency and interpretability with deep learning models, ultimately establishing trust in your anomaly detection applications. Throughout the chapters, you’ll get equipped with XAI and anomaly detection knowledge that’ll enable you to embark on a series of real-world projects. Whether you are building computer vision, natural language processing, or time series models, you’ll learn how to quantify and assess their explainability. By the end of this deep learning book, you’ll be able to build a variety of deep learning XAI models and perform validation to assess their explainability. What you will learn Explore deep learning frameworks for anomaly detection Mitigate bias to ensure unbiased and ethical analysis Increase your privacy and regulatory compliance awareness Build deep learning anomaly detectors in several domains Compare intrinsic and post hoc explainability methods Examine backpropagation and perturbation methods Conduct model-agnostic and model-specific explainability techniques Evaluate the explainability of your deep learning models Who this book is forThis book is for anyone who aspires to explore explainable deep learning anomaly detection, tenured data scientists or ML practitioners looking for Explainable AI (XAI) best practices, or business leaders looking to make decisions on trade-off between performance and interpretability of anomaly detection applications. A basic understanding of deep learning and anomaly detection–related topics using Python is recommended to get the most out of this book. Full Product DetailsAuthor: Cher Simon , Jeff BarrPublisher: Packt Publishing Limited Imprint: Packt Publishing Limited ISBN: 9781804617755ISBN 10: 180461775 Pages: 218 Publication Date: 31 January 2023 Audience: General/trade , General Format: Paperback Publisher's Status: Active Availability: Available To Order ![]() We have confirmation that this item is in stock with the supplier. It will be ordered in for you and dispatched immediately. Table of ContentsTable of Contents Understanding Deep Learning Anomaly Detection Understanding Explainable AI Natural Language Processing Anomaly Explainability Time Series Anomaly Explainability Computer Vision Anomaly Explainability Differentiating Intrinsic versus Post Hoc Explainability Backpropagation Versus Perturbation Explainability Model-Agnostic versus Model-Specific Explainability Explainability Evaluation SchemesReviewsAuthor InformationCher Simon is a principal solutions architect specializing in artificial intelligence, machine learning, and data analytics at AWS. Cher has 20 years of experience in architecting enterprise-scale, data-driven, and AI-powered industry solutions. Besides building cloud-native solutions in her day-to-day role with customers, Cher is also an avid writer and a frequent speaker at AWS conferences. Tab Content 6Author Website:Countries AvailableAll regions |