Open Source Explainability – Understanding Model Decisions Using Alibi


Explainable AI, or XAI, is a rapidly expanding field of research that aims to supply methods for understanding model predictions. We will start by providing a general introduction to the field of explainability, introduce the Alibi library and focus on how it helps you to understand trained models. We will then explore the collection of algorithms provided by Alibi and the types of insight they each provide, looking at a broad range of datasets and models, discussing the pros and cons of each. In particular, we'll look at methods that apply to any model. The focus will be on application to real-world datasets to show the practitioner that XAI can justify, explore and enhance their use of ML.

Background Knowledge:
Basic familiarity with python is desirable


Alex Athorne is a Research Engineer at Seldon, where he works on open-source libraries for explainability and drift detection. He studied mathematics at Warwick and went on to do a PhD at Imperial College London in dynamical systems. He's passionate about open-source development and writing about his experiences in ML.

Open Data Science




Open Data Science
One Broadway
Cambridge, MA 02142

Privacy Settings
We use cookies to enhance your experience while using our website. If you are using our Services via a browser you can restrict, block or remove cookies through your web browser settings. We also use content and scripts from third parties that may use tracking technologies. You can selectively provide your consent below to allow such third party embeds. For complete information about the cookies we use, data we collect and how we process them, please check our Privacy Policy
Consent to display content from - Youtube
Consent to display content from - Vimeo
Google Maps
Consent to display content from - Google