How Not to Destroy the World With AI Stuart
Author : faustina-dinatale | Published Date : 2025-05-29
Description: How Not to Destroy the World With AI Stuart Russell University of California Berkeley In David Lodges Small World the protagonist causes consternation by asking a panel of eminent but contradictory literary theorists the following
Presentation Embed Code
Download Presentation
Download
Presentation The PPT/PDF document
"How Not to Destroy the World With AI Stuart" is the property of its rightful owner.
Permission is granted to download and print the materials on this website for personal, non-commercial use only,
and to display it on your personal computer provided you do not modify the materials and that you retain all
copyright notices contained in the materials. By downloading content from our website, you accept the terms of
this agreement.
Transcript:How Not to Destroy the World With AI Stuart:
How Not to Destroy the World With AI Stuart Russell University of California, Berkeley In David Lodge’s Small World,, the protagonist causes consternation by asking a panel of eminent but contradictory literary theorists the following question: “What if you were right?” None of the theorists seems to have considered this question before. Similar confusion can sometimes be evoked by asking AI researchers, “What if you succeed?” AI is fascinating, and intelligent computers are clearly more useful than unintelligent computers, so why worry? AIMA1e, 1994 Growth in PPL papers AI systems will eventually make better decisions than humans From: Superior Alien Civilization To: humanity@UN.org Subject: Contact Be warned: we shall arrive in 30-50 years From: humanity@UN.org To: Superior Alien Civilization Subject: Out of office: Re: Contact Humanity is currently out of the office. We will respond to your message when we return. Standard model for AI Righty-ho Also the standard model for control theory, statistics, operations research, economics King Midas problem: Cannot specify R correctly Smarter AI => worse outcome E.g., social media Optimizing clickthrough = learning what people want = modifying people to be more predictable Humans are intelligent to the extent that our actions can be expected to achieve our objectives Machines are intelligent to the extent that their actions can be expected to achieve their objectives Machines are beneficial to the extent that their actions can be expected to achieve our objectives How we got into this mess 1. Robot goal: satisfy human preferences* 2. Robot is uncertain about human preferences 3. Human behavior provides evidence of preferences New model: Provably beneficial AI => assistance game with human and machine players Smarter AI => better outcome Human behaviour Machine behaviour Human objective AIMA 1,2,3: objective given to machine Machine behaviour Human objective AIMA 1,2,3: objective given to machine Human behaviour Machine behaviour Human objective AIMA 4: objective is a latent variable Old: minimize loss with (typically) a uniform loss matrix Accidentally classify human as gorilla Spend millions fixing public relations disaster New: structured prior distribution over loss matrices Some examples safe to classify Say “don’t know” for others Use active learning to gain additional feedback from humans Example: image classification What does “fetch some coffee” mean? If there is so much uncertainty about preferences, how does the robot do anything useful? Answer: The instruction suggests coffee would have higher value than expected a priori, ceteris