Interpretable Machine Learning

Software > Computer Software > Educational Software Duke University

Course Overview

As Artificial Intelligence (AI) becomes integrated into high-risk domains like healthcare, finance, and criminal justice, it is critical that those responsible for building these systems think outside the black box and develop systems that are not only accurate, but also transparent and trustworthy. This course is a comprehensive, hands-on guide to Interpretable Machine Learning, empowering you to develop AI solutions that are aligned with responsible AI principles. You will also gain an understanding of the emerging field of Mechanistic Interpretability and its use in understanding large language models. Through discussions, case studies, programming labs, and real-world examples, you will gain the following skills: 1. Describe interpretable machine learning and differentiate between interpretability and explainability. 2. Explain and implement regression models in Python. 3. Demonstrate knowledge of generalized models in Python. 4. Explain and implement decision trees in Python. 5. Demonstrate knowledge of decision rules in Python. 6. Define and explain neural network interpretable model approaches, including prototype-based networks, monotonic networks, and Kolmogorov-Arnold networks. 7. Explain foundational Mechanistic Interpretability concepts, including features and circuits 8. Describe the Superposition Hypothesis 9. Define Representation Learning and be able to analyze current research on scaling Representation Learning to LLMs. This course is ideal for data scientists or machine learning engineers who have a firm grasp of machine learning but have had little exposure to interpretability concepts. By mastering Interpretable Machine Learning approaches, you'll be equipped to create AI solutions that are not only powerful but also ethical and trustworthy, solving critical challenges in domains like healthcare, finance, and criminal justice. To succeed in this course, you should have an intermediate understanding of machine learning concepts like supervised learning and neural networks.

Course FAQs

What are the prerequisites for 'Interpretable Machine Learning'?

Prerequisites for this continuing education class are set by Duke University. Most professional development online classes benefit from some prior knowledge. Please check the provider's page for specific requirements.

Will I receive a certificate for this CE class?

Yes, upon successful completion, Duke University typically offers a shareable certificate to showcase your new skills and fulfill your continuing education requirements.

How long does this online course take to complete?

Completion times for online continuing education courses vary. The provider's website will have the most accurate estimate of the time commitment needed.