cmsc 671 fall 2001

Post on 06-Jan-2016

24 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

DESCRIPTION

CMSC 671 Fall 2001. Class #11 – Tuesday, October 9. Today’s class. Philosophy of AI Can we build intelligent machines? If we do, how will we know they’re intelligent? Should we build intelligent machines? If we do, how should we treat them… …and how will they treat us?. Philosophy of AI. - PowerPoint PPT Presentation

TRANSCRIPT

1

CMSC 671CMSC 671Fall 2001Fall 2001

Class #11 – Tuesday, October 9

2

Today’s class

• Philosophy of AI– Can we build intelligent machines?

• If we do, how will we know they’re intelligent?

– Should we build intelligent machines?• If we do, how should we treat them…

• …and how will they treat us?

3

Philosophy of AI

Alan M. Turing, “Computing Machinery and Intelligence”

John R. Searle, “Minds, Brains, and Programs”

J. Storrs Hall, “Ethics for Machines”

(supplementary: Russell & Norvig Ch. 27)

4

Philosophical debates

• What is AI, really?– What does an intelligent system look like?

– Do we need, and can we have, emotions, consciousness, empathy, love?

• Can we ever achieve AI, even in principle?

• How will we know if we’ve done it?

• If we can do it, should we?

5

Turing test

• Basic test: – Interrogator in one room, human in another, system in a third

– Interrogator asks questions; human and system answer

– Interrogator tries to guess which is which

– If the system wins, it’s passed the Turing Test

• The system doesn’t have to tell the truth (obviously…)

6

Turing test objections

• Objections are basically of two forms:– “No computer will ever be able to pass this test”

– “Even if a computer passed this test, it wouldn’t be intelligent”

7

“Machines can’t think”

• Theological objections• “It’s simply not possible, that’s all”• Arguments from incompleteness theorems

– But people aren’t complete, are they?

• Machines can’t be conscious or feel emotions– Reductionism doesn’t really answer the question: why can’t

machines be conscious or feel emotions??

• Machines don’t have Human Quality X• Machines just do what we tell them to do

– Maybe people just do what their neurons tell them to do…

• Machines are digital; people are analog

8

“The Turing test isn’t meaningful”:Chinese Room argument

9

“The Turing test isn’t meaningful”

• Maybe so, but…

If we don’t use the Turing test, what measure should we use?

• Very much an open question…

10

Ethical concerns: Robot behavior

• How do we want our intelligent systems to behave?

• How can we ensure they do so?

• Asimov’s Three Laws of Robotics:1. A robot may not injure a human being or, through inaction, allow a

human being to come to harm.

2. A robot must obey orders given it by human beings except where such orders would conflict with the First Law.

3. A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.

11

Ethical concerns: Human behavior

• Is it morally justified to create intelligent systems with these constraints?– As a secondary question, would it be possible to do so?

• Should intelligent systems have free will? Can we prevent them from having free will??

• Will intelligent systems have consciousness? (Strong AI) – If they do, will it drive them insane to be constrained by artificial

ethics placed on them by humans?

• If intelligent systems develop their own ethics and morality, will we like what they come up with?

top related