Top Banner
Abstract Neuron w 2 w n w 1 w 0 i 0 =1 o u t p u t y i 2 i n i 1 . . . i n p u t i n i i ii w net 0 y 1 if net > 0 0 otherwise {
56

Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Dec 20, 2015

Download

Documents

Welcome message from author
This document is posted to help you gain knowledge. Please leave a comment to let me know what you think about it! Share it to your friends and learn new things together.
Transcript
Page 1: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Abstract Neuron

w2 wnw1

w0

i0=1

o u t p u t y

i2 ini1. . .

i n p u t i

n

i

iiiwnet0

y 1 if net > 00 otherwise{

Page 2: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Computing with Abstract Neurons

McCollough-Pitts Neurons were initially used to model pattern classification

size = small AND shape = round AND color = green AND Location = on_tree => Unripe_fruit

linking classified patterns to behavior size = large OR motion = approaching => move_away size = small AND location = above => move_above

McCollough-Pitts Neurons can compute logical functions. AND, NOT, OR

Page 3: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Computing logical functions: the OR function

• Assume a binary threshold activation function.

• What should you set w01, w02 and w0b to be so that you can get the right answers for y0?

i1 i2 y0

0 0 0

0 1 1

1 0 1

1 1 1

x0 f

i1 w01

y0i2

b=1

w02

w0b

Page 4: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Many answers would work

y = f (w01i1 + w02i2 + w0bb)

recall the threshold function

the separation happens when w01i1 + w02i2 + w0bb = 0

move things around and you get

i2 = - (w01/w02)i1 - (w0bb/w02)

i2

i1

Page 5: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Decision Hyperplane

The two classes are therefore separated by the `decision' line which is defined by putting the activation equal to the threshold.

It turns out that it is possible to generalise this result to Threshold Units with n inputs.

In 3-D the two classes are separated by a decision-plane.

In n-D this becomes a decision-hyperplane.

Page 6: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Linearly separable patterns

Linearly Separable PatternsPERCEPTRON is an architecture which can solve this type of decision boundary problem. An "on" response in the output node represents one class, and an "off" response represents the other.

Page 7: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

The XOR function

i1 i2 y

0 0 0

0 1 1

1 0 1

1 1 0

Page 8: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

The Input Pattern Space

 

Page 9: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

The Decision planes

 

Page 10: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Multiple Layers

I1 I2

1.50.5

0.5

-11

1 11

1

y

Page 11: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Multiple Layers

I1 I2

1.50.5

0.5

-11

1 11

1

0 1

y

Page 12: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Multiple Layers

I1 I2

1.50.5

0.5

-11

1 11

1

1 1

y

Page 13: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Computing other relations

The 2/3 node is a useful function that activates its outputs (3) if any (2) of its 3 inputs are active

Such a node is also called a triangle node and will be useful for lots of representations.

Page 14: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Triangle nodes and McCullough-Pitts Neurons?

Object (B) Value (C)

Relation (A)

A B C

Page 15: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

“They all rose”

triangle nodes:

when two of the abstract neurons fire, the third also fires

model of spreading activation

Page 16: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Basic Ideas behind the model

Parallel activation streams. Top down and bottom up activation combine to

determine the best matching structure. Triangle nodes bind features of objects to values Mutual inhibition and competition between

structures Mental connections are active neural connections

Page 17: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

5 levels of Neural Theory of Language

Cognition and Language

Computation

Structured Connectionism

Computational Neurobiology

Biology

MidtermQuiz Finals

Neural Development

Triangle NodesNeural Net and learning

abst

ract

ion

Pyscholinguistic experiments

Page 18: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Psychological Studies

Eva Mok

CS182/CogSci110/Ling109

Spring 2006

Page 19: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Read the list

ORANGE BROWN GREENYELLOWBLUERED

Page 20: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Name the print color

XXXXXXXXXXXXXXXXXXXXXXXXXXXXXX

Page 21: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Name the print color

RED GREENBLUEBROWNORANGEYELLOW

Page 22: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

The Stroop Test

Form and meaning interact in comprehension, production and

learning

Page 23: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Top down and bottom up information Bottom-up: stimulus driving processing Top-down: knowledge and context driving

processing When are these information integrated?

Modular view: Staged serial processing Interaction view: Information is used as

soon as available

Page 24: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Tanenhaus et al. (1979) [also Swinney, 1979]

Word / non-word forced choice

Page 25: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Modeling the task with triangle nodes

Page 26: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Reaction times in milliseconds after: “They all rose”

flower 685 659

stood 677 623

desk(control)

711 652

0 delay 200ms. delay

(facilitation)

(facilitation) (facilitation)

(no facilitation)

Page 27: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

When is context integrated? Prime: spoken sentences ending in

homophones

They all rose vs. They bought a rose Probe: stood and flower

No offset: primes both stood and flower 200 ms offset: only primes appropriate sense

Modularity? Or weak contextual constraints?

Page 28: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Eye trackingcomputerEye camera

Scene camera

Allopenna, Magnuson & Tanenhaus (1998)

“Pick up the beaker”

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 29: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Do rhymes compete? Cohort (Marlsen-Wilson): onset similarity is primary because

of the incremental (serial) nature of speech Cat activates cap, cast, cattle, camera, etc. Rhymes won’t compete

NAM (Neighborhood Activation Model; Luce): global similarity is primary Cat activates bat, rat, cot, cast, etc. Rhymes among set of strong competitors

TRACE (McClelland & Elman): global similarity constrained by incremental nature of speech Cohorts and rhymes compete, but with different time

course

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 30: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

TRACE predicts different time course for cohorts and rhymes

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 31: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

TRACE predictions match eye-tracking data

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 32: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Natural contexts are used continuously Conclusion from this and other eye-tracking

studies: When constraints from natural contexts are

extremely predictive, they are integrated as quickly as we can measure

Suggests rapid, continuous interaction among Linguistic levels Nonlinguistic context

Even for processes assumed to be low-level and automatic

Constrains processing theories, also has implications for, e.g., learnability

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 33: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Eye movement paradigm More sensitive than conventional

paradigms More naturalistic Simultaneous measures of multiple items Transparently linkable to computational

model

Adapted from Jim Magnuson, “Interaction in language processing: Pragmatic constraints on lexical access”

Page 34: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Eye-tracker without headsets

http://www.bcs.rochester.edu/infanteyetrack/eyetrack.html

Page 35: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Recap: Goals of psycholinguistic studies Direct goal: finding out what affect

sentence processing Indirect goal: getting at how words,

syntax, concepts are represented in the brain

Modeling: testing out these hypotheses with computational models

Page 36: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Areas studied in psycholinguistics Lexical access / lexical structure Syntactic structure Referent selection The role of working memory Disfluencies

Page 37: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Disfluencies and new information Disfluencies: pause, repetition, restart Often just seen as production /

comprehension difficulties Arnold, Fagnano, and Tanenhaus (2003)

How are disfluent references interpreted? Componenets to referent selection

lexical meaning discourse constraints

Page 38: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Candle, camel, grapes, salt shaker

a. DISCOURSE-OLD CONTEXT: Put the grapes below the candle. DISCOURSE-NEW CONTEXT: Put the grapes below the camel.

b. FLUENT: Now put the candle below the salt shaker. DISFLUENT: Now put theee, uh, candle below the salt shaker.

Page 39: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Predictions on 4 conditions: (Target = candle)

Disfluent/New, Fluent/Given: Target Put the grapes below the camel.

Now put theee, uh, candle below the salt shaker. Put the grapes below the candle.

Now put the candle below the salt shaker.

Disfluent/Given, Fluent/New: Competitor Put the grapes below the candle.

Now put theee, uh, candle below the salt shaker. Put the grapes below the camel.

Now put the candle below the salt shaker.

Page 40: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Disfluencies affect what we look at

Percentage of fixations on all new objects from 200 to 500 ms after the onset of “the”/“theee uh” (i.e. before the onset of the head noun)

Page 41: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Target is preferred in two conditions

Percentage of target fixations minus percentage competitor fixations in each condition. Fixations cover 200–500 ms after the onset of the head noun.

Page 42: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

A lot of information is integrated in sentence processing!

Stroop test [i.e. color words]: form, meaning

Tanenhaus et al (1997) [i.e. “they all rose”]: phonology, meaning, syntactic category

Allopena et al (1998) [i.e. cohorts & rhymes]:phonology, visual context

Arnold et al (2003) [i.e. “theee, uh, candle”]:discourse information, visual context

Page 43: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Producing words from pictures or from other words

A comparison of aphasic lexical access from two different input modalities

Gary Dell with

Myrna Schwartz, Dan Foygel, Nadine Martin, Eleanor Saffran, Deborah Gagnon, Rick Hanley, Janice Kay, Susanne Gahl, Rachel Baron, Stefanie Abel, Walter Huber

Page 44: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

A 2-step Interactive Model of Lexical Access in Production

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Semantic Features

Adapted from Gary Dell, “Producing words from pictures or from other words”

Page 45: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

1. Lemma Access: Activate semantic features of CAT

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Semantic Features

Adapted from Gary Dell, “Producing words from pictures or from other words”

Page 46: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

1. Lemma Access: Activation spreads through network

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

Page 47: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

1. Lemma Access: Most active word from proper category is selected and linked to syntactic frame

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 48: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

2. Phonological Access: Jolt of activation is sent to selected word

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 49: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

2. Phonological Access: Activation spreads through network

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 50: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

2. Phonological Access: Most activated phonemes are selected

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

Syl

On Vo Co

Page 51: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Modeling lexical access errors Semantic error Formal error (i.e. errors related by form) Mixed error (semantic + formal) Phonological access error

Page 52: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Semantic error: Shared features activate semantic neighbors

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 53: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Formal error: Phoneme-word feedback activates formal neighbors

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 54: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Mixed error: neighbors activated by both top-down & bottom-up sources

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

NP

N

Page 55: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

Phonological access error: Selection of incorrect phonemes

FOG DOG CAT RAT MAT

f r d k m ae o t g

Onsets Vowels Codas

Adapted from Gary Dell, “Producing words from pictures or from other words”

Syl

On Vo Co

Page 56: Abstract Neuron w2w2 wnwn w1w1 w0w0 i 0 =1 o u t p u t y i2i2 inin i1i1... i n p u t i 1 if net > 0 0 otherwise {

I’ve shown you... Behavioral experiments, and A connectionist model with the goal of understanding how

language is represented and processed in the brain

Next time: Lisa will talk about imaging experiments