Stanford CS521 - AI Safety Seminar
Stanford,
Updated On 02 Feb, 19
Stanford,
Updated On 02 Feb, 19
4.1 ( 11 )
Carlos Guestrin, Stanford University
May 11, 2022
Machine learning (ML) and AI systems are becoming integral parts of every aspect of our lives. The definition, development and deployment of these systems are driven by (complex) human choices. And, as these AIs are making more and more decisions for us, and the underlying ML systems are becoming more and more complex, it is natural to ask the question: “How can you trust machine learning?”
In this talk, I’ll present a framework, anchored on three pillars: Clarity, Competence and Alignment. For each, I’ll describe algorithmic and human processes that can help drive towards more effective, impactful and trustworthy AIs. For Clarity, I’ll cover methods for making the predictions of machine learning more explainable. For Competence, I will focus on methods to evaluating and testing ML models with the rigor that we apply to complex software products. Finally, for Alignment, I’ll describe the complexities of aligning the behaviors of an AI with the values we want to reflect in the world, along with methods that can yield more aligned outcomes.Through this discussion, we will cover both fundamental concepts and actionable algorithms and tools that can lead to increased trust in ML.
Additional recommended reading:
https://arxiv.org/abs/1602.04938
https://homes.cs.washington.edu/~marcotcr/acl20_checklist.pdf
https://www.vox.com/2015/9/18/9348821/photography-race-bias
Learn more about the speaker: https://guestrin.su.domains/
Watch more videos from this seminar series here: https://www.youtube.com/playlist?list=PLoROMvodv4rNtnS3JSRRZzLWQo2dd6XNs
#trust #machinelearning
Sam
Sep 12, 2018
Excellent course helped me understand topic that i couldn't while attendinfg my college.
Dembe
March 29, 2019
Great course. Thank you very much.