whatdohiddenrepresentationslearn?...

9
What do hidden representations learn? Plaut and Shallice (1993) Mapped orthography to semantics (unrelated similarities) Compared similarities among hidden representations to those among orthographic and semantic representations (over settling) Hidden representations “split the difference” between input and output similarity 16 / 25 Other animals don’t like onions (but primates do) No correlation for “specific” weights because inputs vary No correlation for “more general” weights because output varies Only intermediate “general” weights build up due to correlations 25 / 29

Upload: duongnhi

Post on 09-Apr-2018

231 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

What do hidden representations learn?

Plaut and Shallice (1993)

Mapped orthography tosemantics (unrelatedsimilarities)

Compared similarities amonghidden representations tothose among orthographicand semantic representations(over settling)

Hidden representations “split the

difference” between input andoutput similarity

16 / 25

Other animals don’t like onions (but primates do)

No correlation for “specific” weights because inputs varyNo correlation for “more general” weights because output variesOnly intermediate “general” weights build up due to correlations

25 / 29

Page 2: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Semantic hierarchy

Progressivedifferentiation indevelopment

Progressive deteroriation in semantic dementia

Rumelhart and Todd (1993)

Page 3: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Progressive differentiation of concepts

Internalrepresentations

“Basic” level

Perceptual to conceptual shift

Page 4: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Semantic memory

Unitary, amodal semantic system

• General conceptual knowledge

abstracted from a large number of

individual episodes or experiences

(Tulving, 1972).

• Mediates among multiple input and

output modalities

• Can be selectively impaired by

brain damage, usually to the

anterior temporal lobes (Warrington,

1975).

Vision Touch Hearing

Action WritingSpeaking

Semantics

Challenges

• Modality-specific effects in semantic priming

• Category- and modality-specific semantic deficits

• Modality-specific naming disorders (e.g., optic aphasia)

Optic aphasia

Selective impairment in visual object naming

• Brain damage to left medial occipital lobe (visual cortex and underlying white matter)

• Not visual agnosia—relatively preserved visual gesturing and other tests of visual

comprehension

• Not general anomia—relatively preserved naming from other modalities (e.g., touch,

spoken definitions)

• Relatively preserved naming of actions associated with visually presented objects

(Manning & Campbell, 1996)

% Correct Performance

Visual Visual Tactile ActionStudy Naming Gesturing Naming Naming

Lhermitte & Beauvois (1973) 73 100 91

Teixeira Ferreira et al. (1997) 53 95 81 75

Manning & Campbell (1996) 27 75 90 67

Coslett & Saffran (1989) 0 50 92

Analogous selective naming deficits have been observed for tactile input (Beauvois et al.,

1978) and for auditory input (Denes & Semenza, 1975)

Challenge to unitary semantics account

Post-semantic lesion

• No basis for sensitivity to input modality

Action(Gesturing)

Phonology (Naming)

Visual Input

Verbal Input

Tactile Input

Semantics

Semantic lesion

• Would impair visual gesturing and non-visual

naming

Action(Gesturing)

Phonology (Naming)

Visual Input

Verbal Input

Tactile Input

Semantics

Pre-semantic lesion

• Would impair visual gesturing (and other

measures of comprehension)

– But might be preserved relative to naming due to

priviledged access (Caramazza et al., 1990)

Action(Gesturing)

Phonology (Naming)

Visual Input

Verbal Input

Tactile Input

Semantics

Alternative view of semantic organization

Multiple modality-specific semantic

systems (Beauvois, 1982; Lhermitte &

Beauvois, 1973; Shallice, 1987; Warrington,

1975)

Action(Gesturing)

Phonology (Naming)

VisualSemantics

TactileSemantics

VerbalSemantics

Visual Input

Verbal Input

Tactile Input

Optic aphasia: Disconnection from visual to verbal semantics

Problems

• Unparsimonious and post-hoc

• Poor accounts of acquisition, cross-modal generalization and priming

• No account of relative sparing of visual action naming

Page 5: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Current approach

• The semantic system operates according to connectionist/parallel distributed

processing (PDP) principles:

– Processing: Responses are generated by the interactions of large numbers of simple,

neuron-like processing units.

– Representation: Within each modality, similar objects are represented by overlapping

distributed patterns of activity.

– Learning: Knowledge is encoded as weights on connections between units, adjusted

gradually based on task performance.

• Semantic representations develop a graded degree of modality-specific

specialization in learning to mediate between multiple input and output

modalities.

• Graded specialization derives from two factors:

1. Task systematicity: Whether similar inputs map to similar outputs

⇒ Naming is an unsystematic task

2. Topographic bias: Learning favors “short” connections (Jacobs & Jordan, 1992)

⇒ Mappings rely most on regions of semantics “near” relevant modalities

Vision

Action(Gesturing)

Phonology(Naming)

Touch

Task

• Continuous recurrent

attractor network

• Two input groups

(Vision, Touch)

equidistant from two

output groups (Action,

Phonology)

• Task units: object vs.

action tasks

Tasks

• Naming objects from vision or touch

• Naming and gesturing action associated with objects from vision or touch

Stimulus

(Vision or Touch) Task ⇒ Phonology Action

object ⇒ “bed” –

action ⇒ “sleep”

Training procedure

• Object (N = 100), modality of presentation (Vision vs. Touch), and task (object vs.

action) chosen randomly during training

• Activations clamped on appropriate input modality; network settled for 5.0 units of

time (τ = 0.2); error injected only over last time unit

• Error derivatives for each weight calculated by back-propagation-through-time

adapted for continuous-time networks (Pearlmutter, 1989)

• Weight changes scaled by Gaussian function (SD = 10) of connection length

(cf. Jacobs & Jordan, 1992)

• 110,000 total object presentations (≈275 per condition);

all output activations on correct side of 0.5 for all objects and tasks

Page 6: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Vision

Action(Gesturing)

Phonology(Naming)

size of white square =unit activity level

Touch

Task

1 3 5 7 9 11 13 15X Position

0.0

0.2

0.4

0.6

0.8

1.0

Wei

ght-

Cha

nge

Fact

or Vision InputsTouch Inputs

Acquisition

0 10 20 30 40 50 60 70 80 90 100Object Presentations (x 1000)

0

10

20

30

40

50

60

70

80

90

100

Perc

ent C

orre

ct

Visual GesturingVisual Action NamingVisual Object Naming

Semantic similarity

Mean correlations among pairs of semantic representations generated by each object

in each input modality (note: relatedness is relative to visual categories).

Identical Same Category UnrelatedItem Pairs

0.0

0.1

0.2

0.3

0.4

0.5

0.6

0.7

Mea

n Se

man

tic C

orre

latio

n

Same ModalityDifferent Modalities

.63 (.08)

.36 (.19)

.22 (.16)

.03 (.12).02 (.11)

• An object is most similar to itself regardless of modality of presentation

(but cross-modal representations are not identical)

• Accounts for reduction in semantic priming with cross-modal presentation

Magnitude of incoming weights to semanticsVision Input

Vis

ion

Touch Input

Touch

1 3 5 7 9 11 13 15X Position

0.0

0.2

0.4

0.6

0.8

1.0

1.2

1.4

Inco

min

g W

eigh

t Mag

nitu

deVisual InputTactile Input

Page 7: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Topographic lesions

Lesions applied to Vision ⇒ Semantics connections

• Probability of removing each connection is a Gaussian function of the distance of unit from

lesion location; SD of Gaussian controls severity (1.5 below).

Vision

Action(Gesturing)

Phonology(Naming)

Touch

Task

Effects of lesion location

Performance on visual naming and visual gesturing after lesions to Vision ⇒

Semantics connections centered at each location (10 repetitions each; SD = 1.5).

Visual Naming

Action

Vis

ion

Phonology

Visual Gesturing

Action

Touch

Phonology

Effects of lesion location

% correct visual gesturing − % correct visual naming

(white: gesturing > naming; black: naming > gesturing)

Action

Vis

ion

Touch

Phonology

Effects of lesion severity

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0Vision−to−Semantics Lesion Severity (SD)

0

10

20

30

40

50

60

70

80

90

100

Per

cent

Cor

rect

Tactile NamingVisual GesturingVisual Action NamingVisual Object Naming

1

2

3

4

• Highly selective impairment

of visual object naming

relative to visual gesturing

and tactile naming

• Relative preservation of

visual action naming

% Correct Performance

Visual Visual Tactile ActionStudy Naming Gesturing Naming Naming

1. Lhermitte & Beauvois (1973) 73 100 91

2. Teixeira Ferreira et al. (1997) 53 95 81 75

3. Manning & Campbell (1996) 27 75 90 67

4. Coslett & Saffran (1989) 0 50 92

Page 8: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Control simulation: No topographic bias

• Exact replication of simulation except without topographic bias

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0Vision−to−Semantics Lesion Severity (SD)

0

10

20

30

40

50

60

70

80

90

100

Per

cent

Cor

rect

Tactile NamingVisual GesturingVisual Action NamingVisual Object Naming

1

2

3

4

• Relative degrees of impairment due to differences in task systematicity

• Effect is not as large as in patients

Generating object names vs. action names

Performance after lesions to semantics (SD = 2.0).

Visual Object Naming

Action

Vis

ion

Visual Action Naming

Action

Touch

• Generating names of actions associated with objects involves interactions with

Action representations.

Modality-specific impairment of grammatical categories

Selective impairment on nouns vs. verbs restricted to a particular input or output

modality (Caramazza & Hillis, 1991; Hillis & Caramazza, 1995; Rapp & Caramazza, 1997)

• Patient HW (Caramazza & Hillis, 1991)

– nouns > verbs in spoken output but not in written output

• Patient EBA (Hillis & Caramazza, 1995)

– two lesions: left frontal and left temporal

Modality-specific impairment of grammatical categories

Account 1: Modality-specific lexical representations divided by grammatical category

(Caramazza & Hillis, 1991; Hillis & Caramazza, 1995; Rapp & Caramazza, 1997)

orthographic lexical form

nouns verbs

A P P L E

G I R L E A T

G I V E

phonological lexical form

/ it /

/ gIv /

nouns verbs

/ girl /

/ apl /

[eat]V

[give]V

[girl]N

[apple]N

semantic/syntacticlexical specification

nouns +verbs −

nouns −verbs +

Page 9: Whatdohiddenrepresentationslearn? …cnbc.cmu.edu/~plaut/IntroPDP/slides/semantics.pdfWhatdohiddenrepresentationslearn? Plaut and Shallice ... 0.6 0.8 Weight-Change Factor 1.0 Vision

Modality-specific impairment of grammatical categories

Account 2: Modality-specific access pathways divided by grammatical category (Rapp

& Caramazza, 1997)

orthographic lexical form

phonological lexical form

/ it /

/ gIv /

nouns

verbs nouns

verbs

[eat]V

[give]V

[girl]N

[apple]N

semantic/syntacticlexical specification

A P P L E

G I R L

nouns +verbs −

nouns −verbs +

Modality-specific impairment of grammatical categories

Object naming − Action naming (white: nouns > verbs; black: verbs > nouns)

Incoming Lesions

Vision ⇒ Semantics (SD = 2.0)

Vis

ion

Outgoing Lesions

Semantics ⇒ Phonology (SD = 1.5)

Touch

• Nouns depend more on semantics near Vision/Phonology; Verbs depend more on

semantics near Action/Touch

• Lesions to incoming/outgoing connections produce modality-specific grammatical category

impairments

Modality-specific impairment of grammatical categories

[4,14] [10,5] [5,9] [14,6]0

20

40

60

80

100

Per

cent

Cor

rect

Incoming Lesions Outgoing Lesions

Nouns (visual object naming)Verbs (visual action naming)

Summary

• Differences in task systematicity and a topographic bias on learning produce a

graded degree of modality-specific specialization within semantics.

⇒ The network does not develop separate modality-specific semantic

systems—both visual and tactile input engage all of semantics.

• Damage to connections from vision to regions of semantics near phonology

produces a highly selective visual naming deficit as observed in optic aphasia.

• The relative sparing of naming visual actions in optic aphasia results from the

preserved support of action representations.

• Modality-specific impairments to nouns vs. verbs can arise from lesions to/from

regions of semantics partially specialized for these grammatical classes.