1 4 questions (Revisited) What are our underlying assumptions about intelligence? What kinds of techniques will be useful for solving AI problems? At what.

Post on 21-Dec-2015

214 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

Transcript

1

4 questions (Revisited)

What are our underlying assumptions about intelligence?

What kinds of techniques will be useful for solving AI problems?

At what level of detail are we trying to model human intelligence?

How will we know when we have succeeded in building an intelligent program?

2

Symbolic vs. Subsymbolic AI

Subsymbolic AI: Model intelligence at a level similar to the neuron. Let such things as knowledge and planning emerge.

Symbolic AI: Model such things as knowledge and planning in data structures that make sense to the programmers that build them.

(blueberry (isa fruit) (shape round) (color purple) (size .4 inch))

3

Interest in Subsymbolic AI

40 50 60 70 80 90 00 10

4

Low-level (Sensory and Motor) Processing and the Resurgence of Subsymbolic Systems

•Computer vision

•Motor control

•Subsymbolic systems perform cognitive tasks

•Detect credit card fraud

•The backpropagation algorithm eliminated a formal weakness of earlier systems

•Neural networks learn.

5

Turing test

Experimenter

AI system

Control

6

Eliza, 1965

HUMAN: You are like my father in some ways. ELIZA: What resemblance do you see? HUMAN: You are not very aggressive, but I think you don’t want me to notice that. ELIZA: What makes you think I am not very aggressive? HUMAN: You don’t argue with me. ELIZA: Why do you think I don’t argue with you? HUMAN: You are afraid of me. ELIZA: Does it please you to believe I am afraid of you? HUMAN: My father is afraid of everybody. ELIZA: What else comes to mind when you think of your father? HUMAN: Bullies.

http://www-ai.ijs.si/eliza/

7

The Chinese Room

Set of rules, in English, for

transforming phrases

Chinese Writing is

given to the person

Correct Responses

She does not know Chinese

8

The Chinese Room Scenario

An individual is locked in a room and given a batch of Chinese writing. The person locked in the room does not understand Chinese.

Next she is given more Chinese writing and a set of rules (in English which she understands) on how to collate the first set of Chinese characters with the second set of Chinese characters.

If the person becomes good at manipulating the Chinese symbols and the rules are good enough, then to someone outside the room it appears that the person understands Chinese.

9

The Chinese Room (cont.)

Searle's, who developed the argument, point is that she doesn't really understand Chinese, she really only follows a set of rules.

Following this argument, a computer could never be truly intelligent, it is only manipulates symbols. The computer does not understand the semantic context.

Searle’s criteria is “intentionality,” the entity must be intentionally exhibiting the behavior, not simply following a set of rules.

Intentionality is as difficult to define as intelligence. Searle excludes ‘weak AI’ from his argument against the possibility of AI.

10

Philosophical extremes in AI

Weak AI vs. Strong AI Weak AI believes that machine intelligence need only mimic the behavior of

human intelligence Strong AI demands that machine intelligence must mimic the internal processes

of human intelligence, not just the external behavior

11

Newell and Simon Prediction

In 1997, Deep Blue beat Gary Kasparov.

12

Why Did They Get it Wrong?

They failed to understand at least three key things:

•The need for knowledge (lots of it)

•Scalability and the problem of complexity and exponential growth

•The need to perceive the world

13

Scalability

Solving hard problems requires search in a large space.

To play master-level chess requires searching about 8 ply deep. So about 358 nodes must be examined.

14

Exponential Growth

15

But Chess is Easy

•The rules are simple enough to fit on one page

•The branching factor is only 35.

16

A Harder One

John saw a boy and a girl with a red wagon with one blue and one white wheel dragging on the ground under a tree with huge branches.

17

How Bad is the Ambiguity?

•Kim (1) •Kim and Sue (1) •Kim and Sue or Lee (2) •Kim and Sue or Lee and Ann (5) •Kim and Sue or Lee and Ann or Jon (14) •Kim and Sue or Lee and Ann or Jon and Joe (42) •Kim and Sue or Lee and Ann or Jon and Joe or Zak (132) •Kim and Sue or Lee and Ann or Jon and Joe or Zak and Mel (469) •Kim and Sue or Lee and Ann or Jon and Joe or Zak and Mel or Guy (1430) •Kim and Sue or Lee and Ann or Jon and Joe or Zak and Mel or Guy and Jan (4862)

1

22)(

n

n

n

nnCat

18

The Differences Between Us and Them

Emotions

Understanding

Consciousness

19

Today: Computer as Artist

Two paintings done by Harold Cohen’s Aaron program:

20

Solving AI Problems

Define and analyse the problem– What knowledge is necessary?

Choose a problem-solving technique

e.g. Chess– What information do we need to represent in a

chess-playing program?

21

‘State Space’

Initial state

operators

Goal state(s)

r k b ki q b k r

p p p p p p pp

P P P P P P P P

K B KI Q B K RR

pp

k

q

KI

22

The Water Jugs Problem

2 jugs– 4 gallon – 3 gallon

How can you get exactly 2 gallons into the 4 gallon jug?

Possible operators:1. Empty jug

2. Fill jug from tap

3. Pour contents from one jug into another

43

top related