language, rules, and serial processing prof.dr. jaap murre university of amsterdam university of...

Post on 31-Dec-2015

215 Views

Category:

Documents

0 Downloads

Preview:

Click to see full reader

TRANSCRIPT

Language, Rules, and Serial Processing

Prof.dr. Jaap Murre

University of Amsterdam

University of Maastricht

jaap@murre.com

http://www.neuromod.org

Overview

• Speech

• Symbols

• Language

• Language acquisition

• Semantics

Speech

What is speech?

• Speech are modulated wave forms that are produced by a source (lungs and glottis) and filtered by the vocal tract and lips and cheeks

Source-filter model of speech

Speech perception is very difficult. It is also categorical.

Categories, Symbols, Subsymbols

Categories

• Items within a category are considered equal with respect to that category

• Categories thus induce their own similarity structure on whatever they represent

• A dolphin seems to be a big fish, but as a mammal is in a different category

• Network always impose a priori constraints on categories

Induced categories

• Categories are formed on the basis of regularities in the input (a posteriori constraints) and network architecture (a priori constraints)

• Neural network categories are subsymbolic

Symbol

• What we now call symbol is really what de Saussure calls a sign (his symbol is what we would now call icon or logo)

• Ferdinand de Sausssure (1857-1913), Swiss Founder of Modern Linguistics

From Course in General Linguistics (1916)

Nature of the Linguistic Sign

• The linguistic sign unites, not a thing and a name, but a concept and a sound-image. The latter is not a material sound, a purely physical thing, but the psychological imprint of the sound, the impression it makes on our senses. (p.66)

Generalization of the sign

• I propose to retain the word sign [signe] to designate the whole and to replace concept and sound-image respectively by signified [signifié] and signifier [signifiant]; … (p.67)

Principle I: The Arbitrary Nature of the Sign

• The bond between the signifier and the signified is arbitrary (p.67)

• … the individual does not have the power to change a sign in any way once it has become established in the linguistic community… (p.69)

Principle II: The Linear Nature of the Signifier

• The signifier, being auditory, is unfolded in time from which it gets the following characteristics:– (a) it represents a span, and– (b) the span is measurable in a single

dimension; it is a line. (p.70)

Problems for connectionism

• Neural network categories are co-determined by input regularities and network architecture: – They are not arbitrary (symbols/signs are)

• Neural network representations are parallel– Symbols/signs are linear (serial)

Language

Language

• What is language?

• Is it innate or learned?

• Where located in the brain?

• Can neural networks represent language processes?

What is language?

• De Saussure distinguished ‘langue’ from ‘parole’

• Chomsky distinguished ‘competence’ from ‘performance’

• Chomsky strongly defended the idea of the innateness of language

Language is hierarchical and can be extremely ambiguous

Grammar may be innate

The essence of grammar is recursion

Simple grammar

G = {N,V,S,P}

S aSaS bSbS c

E.g., c, aca, bcb, aacaa, aabacabaa

S aSa aaSaa aabSbaa aabaSabaa aabacabaa

The man lit his awful cigar

The man that you thought was old lit his awful cigar

The man that you thought that your mother had seen lit his awful cigar

et cetera

It allows an infinite number of sentences to be generated by just a few rules

Where does language come from? • Certain aspects of the development of language

and thought appear to be universal in that they – (i) preceed any learning by the individual

– (ii) are found in all individuals in the same way

• These universalia are often of a deep and abstract nature

• It is not known at present how they are respresented in the brain, or how they emerge from brain organization

Universal constraints in thought development

• Spelke shows that from a very early age, infants know about the continuity and solidity of objects

• These constraints lie at the core of the developmental learning system

• It is not clear how these are represented in the brain or how they emerge

Selection versus instruction

• Chomsky/Pinker: The child must select a grammar

• Bickerton: The child is provided with a specific grammar, which it than modifies in the direction of the caretaker’s language

Bickerton: Not all languages may be equally hard to learn

• Children’s errors when learning English often resemble Creole, for example, the so called double negative

• Perhaps, Creole is the ‘original mother language’

Willem Levelt’s model of speech production and perception

From concept to speech signal

Very complicated transformation take place during speaking

• A conceptual representation is a network of neurons that fire with a complex associative correlational pattern

• This conceptual-semantic pattern is transformed into a hierarchical syntactic pattern

• This pattern is transformed into a serial speech pattern

Language acquisition

Simple recurrent networks

Inducing syntactic structures

Simple Recurrent Network

Buffer (copy of hidden layer)…, D, C, B, A

Simple Recurrent Network

• Introduced by Jeff Elman as a simplification of the general recurrent backpropagation algorithm (Rumelhart, Hinton, & Williams, 1986).

• Feedforward plus 1 buffer

• Allows learning of sequences

• Can learn simple grammars if embedding is not too deep (N.B. grammar induction is NP-Complete!)

Limitations (Sharkey et al., 2000)

• Hard and unreliable to train– Extremely sensitive to initial weight

configurations– Only 2 out of 90 networks learned to be Finite

State Grammar Recognizers

• Cannot handle remote dependencies

• Useless as a psychological model for grammar acquisition

Finite State Grammar L1 (Sharkey, Sharkey, and Jackson, 2000)

D

D

E

E

B

C

H(S)

A

B

A

C

2

1

3 5

4

6

Connectionism offers a battle ground for debate

• McClelland and Rumelhart’s model of past tense learning has ignited a furious and fertile debate

• Rather than rhetoric and assertions, models are used to support arguments

• These models typically offer existence proofs at this point

Existence proofs

• You say: Your idea cannot work because you could never do X

• I make a model that implements my idea and that can do X

• Now have an existence proof that my idea covers X

• This does not in any way prove that my idea is correct or plausible from a psychological (or biological etc.) perspective

Much more of language can be induced than expected

• Past tense, pluralization, case systems can all be learned from examples

• Also text-to-speech

• Also exceptions can be acquired in this way without disturbing the behavior of the network

• Phonological, segmental, and prosodic regularities can successfully be detected (induced), e.g., word stress

Semantics

Semantic networks may be used to help think about the associative networks in the brain

Better is it to view concepts as vectors of abstract ‘features’

Acquisition of semantics

• Can semantics be induced?

• How much a priori structure needs to be present and what form should this take?

• Are rules superfluous?

• What is the relationship between episodic and semantic knowledge?

How can semantic organization be organized according to category?

• Self-organizing maps in the brain can explain the emergence of topological mappings

• Examples are: – the somatosensory homunculus (discussed in

lecture 7)– retinotopic maps in V1 (area 17, discussed in

lecture 3)

Semantic organization can emerge on the basis of word context (Ritter and Kohonen, 1990)

Interesting is that words organize into both semanticand grammatical categories

Example of a semantotopic map

Recent extension to Latent Semantic Analysis by Landauer and colleagues

• Meaning determined by context

• Reduce the semantic space by singular value decomposition

• This improves generalization

• Applications to automatic dictionaries and even essay grading!

Conclusions

• Neural networks are able to do language and serial processing

• They are not great at it

• The attempts to have neural networks process and above all learn language have ignite an important debate with the proponents of rule-based methods

top related