synthesis of neural networks: the case of cascaded hebbiansszepesva/papers/tr96-102.pdf ·...

22
Synthesis of neural networks: the case of cascaded Hebbians Csaba Szcpcsvari August 1996 TR-96-102 Research Group on Artificial Intelligence Hungarian Academy of Sciences, "Attila 6sef" U niversity Seged, 6720, Aradi vrt tere 1. HUNGARY http://www.inf.u-szeged.hu/-rgai This researh \vas sllpported by OTKA rant No. F201:)2.

Upload: dinhnga

Post on 27-Apr-2018

224 views

Category:

Documents


6 download

TRANSCRIPT

Page 1: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural networks: the case of

cascaded Hebbians

Csaba Szcpcsvari

August 1996

TR-96-102

Research Group on Artificial Intelligence

Hungarian Academy of Sciences,

"A t tila .J 67:sef" U ni versi ty

S7:eged, 6720, Aradi vrt tere 1. HUNGARY

http://www.inf.u-szeged.hu/-rgai

This resear('.h \vas sllpported by OTKA (i-rant No. F201:)2.

Page 2: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

Synthesis of neural networks: the case of

cascaded Hebbians

Csaba Szepesviri

Bolyai Institute of Mathematics, "Attila .J6zsef" University,

Research Group on Artificial Intelligence

Hungarian Academy of Sciences,

"Attila .J6zsef" University

Szeged, ()720, Aradi vrt tere l.

RCNGARY

Abstract

\iVe show that cascading Hebbian learning with any other convergent

{l.lgoriLhrn (called the rorwiLrd algoriLhrn) res liltS ill 1.lle cOllvergerrce or Lhe

Hebbian weights to a stationary point where the Hebbian algorithm ·would

converge if the weight& of the forward algorithm had already converged.

Further, it is shown that the convergence rate of the composite algorithm

does not deteriorate because of the cascading. This result is a conse­

quence of a more general theorem which is also stated and proved here.

nlr. proors hr.ing hasr.d on a global Lipsdlit:;o:ian a..<;sllmpLion. Tllr. tllr.ory

is illllsLriLtcd by <I. composiLe PC/\-Hehhiiul ar<:hit.ccLllre illirodll(:ed by

Micheals (Michaels, 199G).

Page 3: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

:! Synthesis of neural netl-rorks

1 Introduction

Hybrid learning rllles for training llllllti -hyer nellr<ll networks have been <::onsid-

ered by scores ofrese<ll'<::lwrs in the TIeld . These in<::lllde Hecht-Nielsen (Hecht-'Jielsen, 1987),

Ritter et a!. (RitLer et a!', 19R8), :Vlooay and Darken (Moody ana Darken, 1989) ,

Pogglo and Giror:;1 (Pogglo aud Girotil, 1990), Kowlau (�ovvla.u, 1990), I3eualm

aud TOlIlaslni (I3eualm aud Toma�lul, 1991,I3enaIlll aud TOlIlatilnl, 1992): :Mar-

LIUelZ and Schulten (:rVlarLlueLz aud Schulten, 1991) li'o llliu el al. (FoIlllu eL al., 1994 L

Szepesva,ri et a1. (Szepesvari et aL, 1994), J outsensalo et a1. (Joutsensalo et al. , 1995),

and Michaels O.Jichaels, 1995) to name but a fe,,"\'.

r-

N1 �

llil N2 �

� L.-

Figure 1: Single hidden layer cascaded network The weight u of E-ubnetwork Nt is trained independently of the weight v of subnet\,",ork N2. V{e investigate in this article if the convergence of the weight vector of N']. follO\ys from the convergence of the weight vector of Nl. We impose <::onditiollS on the tr<lining rllle of iV'l when training N1 and iV'l silllllltaneollsly, or traIuIng N2 after Nl has couverged, whIch make the 1,\,"0 1110deti equivaleut . Specifically, we prove that if N']. is trained by the Hebbian law then the conver­gence speed ofv is simply O(max( l /VI, Illlt -1111)).

Consider, for the sake of simplicit�y, a single hidden-layer network (see Fig-

me 1). Formally, the output of the network is given by y = N,(x,N,(x,u),v). where w = (u, v) E Rml x Rm2 are the weights of the network, x E Rn is the

input, y E Rtn is the output, Nl : Rn x Rm'l � Rk and N2 : RrJ x Rk X Rm2 -----+

R,m are the tr<lnsfer fllll<::tions of the h idden and the Olltpllt layer, respectively.

Page 4: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

The integer k is the number of hidden units. Here .iVI and �V2 describe the

computation performed by the hidden layer neurons and output layer neurons,

respectivel:y.

Such a network can be viewed as the cascade of two subnetworks provided

that the weights v do not influence the training of weights u. Usually, the

weights u are trained by an unsupervised learning rule (e.g. some form of

competitive learning or a maximum likelihood algorithm) and the weights v

are trained according to a t:uperv itied algorithm (Ilebbian learning, compeLi­

Live IIebbian learning, eLc . ) . The Lraining of \veighLt: itl a tlpecial cat:e of the

tltochat:tic approximation (SA) procedure introduced by Robbint: and :Monro

(Robbins and Monro, 1951). Here we are interested in the convergence and rate

of convergence of the weights of the cascaded part. Since, b:y assumption, the

weights v do not influence the weights u we expect that the training of weights

Ut (t denoting the training time) is such that the weights Ut converge to some

limiting value 11*. AftN that no furthN asslllnption is put on the training of

N1. As for the other weights, we provide suffi<::ient conditions llllder ,vhich the

weights of the se<::ond layer, VI, converge to the value they should <::onverge to

w·hen the ,'veighttl Ut are bounded Lo their lilniting value u'" during Lhe training

of lV2. An et:timate of Lhe convergence rate iti abo provided in a tiimple but

i111porLanL cat:e when N2 is trained by t:upervitied IIebbiall learning.

The convergence of cascaded algoriUUllti have been studied in some ddall

by Benaim (Benaim, 1993,Benaim, 1996). First, he considered continuous-tilue

learning under the adiabatic assumption (i.e., that the cascaded part works

on a substantially slower time-scale). Recently, he extended his method to

discrete time SA algorithms and ,vas ahle to drop the adiahati<:: asslllllption

as \vell (Renaim, 1996), hut did not prove any <::onvergence rates. Renaim's

recent convergence results are, in a way, hoth more and less general than ours.

He considers hyhrid algorithms with the <::ascaded part trained hy an arhitrary

gradient rule (in thitl respect hit: retiulL itl very general), bUL he altlo pretluppotles

Page 5: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

4 Synthesis of neural nehvorks

the input layer is trained by another gradient rule which we do not require here.

Actually, he docs not prove the convergence of the cascaded "Tights (except for

the linear case elaborated upon later by us): but just says that if those weights

converge they would converge to the "right'; limit point. The advantage of

our approach is not only its simplicity but that it allows one to estimate the

convergence rate of the compound process directly.

The article is organized as follows: in the next section the key results are

cILeJ and Jiticw::itied. Then in Sec LIon ;) the resulLtl obtained are w:::eu La derive

Lhe convergence anu tlpeed of convergence of a hybrid alg01'ilhll1 introduced in

a dilfel'enL conLexL by Michaeb, all algorithm which is an aHernati ve 1,0 the

classical LMS algorithm. Finally, all the proofs needed are presented in the

Appendix section of the paper.

2 Results

The follmving theorem forms the basis of our results:

'i'HF,ORR'v! 2. 1 I.et X and \' he Hanarh. 8pare8, Un : X X Y ---, X (11. = 0, 1,2, . . .)

he a 8equence of mapping8" let Bn E Y he an arhitrary 8eries, let Boo E Y and

Xoo E X. ret

(1)

and

(2) 1.uith ;1.:0 = Yo. 1 Then

(3)

IThis condition, i.e., that Xo = 'Yo is not necessary at all. However, it is a useful assumption for simplifying the proof.

Page 6: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

5 Synthesis of neural nehvorks

whcrc2 If is the uniform Lipschitz index of Uk(x, e) w.r.t. f) at Boo and, simi­

larly, L% is the uniform Lipschitz index of Uk(x, eG�) w.r.t. x.3

Now suppose thai ;"Cn and en converge to XC<J and Binfty, rcSpcctiDcly, in the

norm of the appropriate spaces. If the Lipschitz constants L� and L� satisfy the

relations

L� :S C(1 - L�), (4)

and 00

II L;;., = 0, (0) m=n

'where n = 0, 1, 2, . .. linn also Limn_c.", llYn - XC-'O II = o.

If, L�, = c/n fur 'Olflt c > 0, L;' :S 1-1/'11 and 110, -000 II = 0(1/0"), '/lihere

0< (t < 1 lhen

llYn -x�11 = O(max(l/n", Ilx" - xooll)),

It is clear that the given convergence ra.te Blust be a strict one hence we cannot

expect Yn converge faster than either ;fn or On. Theorem 2,1 could have many applications, for example, if Un(X, 0) =

fIll', 0) is a contraction with contraction index l' in x (L:� = 'Y) and L� < n for

some B > 0 then the above theorem yields the result that Yn+l = Un (Yn, en)

converges to the fixed point of [Jl, erx,). One simple application of this is III

reinforcement learning, where en corresponds to the estimated parameters of

a Markovian decision process and Un(·, e) is the optimal value operator corre­

sponding to a discounted l\,farkovian decision problem \vith parameter vector e. This is just the synchronous version of the convergence theorem of Gullapalli

and Barto (Gullapalli and Barto, 1094),

In this article the above theorem is applied to Lhe simplest sup ervised train­

ing algorithm, Lo that of IIebbian learning (IIaykin, 199-1). That is, \ve assume

2 Here l,ve a.dmit the convention tha.t n;=m at :::: 1 if m > n. :JThat is, for all ,r E X and 8 E Y IIUk(x,8) - [:k(x,eoo)11 :::; LhllG - eooll and for all

X,Y E X IIUk(x,B,x) - Uk(y,B�)11 <: L�llx- yll,

Page 7: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

6 Synthesis of neural nehvorks

that the output neurons arc trained by the rule

(6)

where (}:t 2: 0, t = 0, 1, '2, . satisfy the 11s11al criteria, 1.e. , they are square

�ulIllllahle uuL nuL linear �ulIllllahle and Ft E rrm2 lIlay depend on Lhe def:)ll'ed

output value Yt, the adual ,�,:e.ighLs of the hidden layer Ut and input of Lhe neL

Xt, buL noL depend 011 Vt:

Ft = F(Yt,Xt,Ut).

(C sually, the dependence of Fi on Xl and Ul goes through the output of the

hidden layer neurons. ) Note that almost all the hybrid algorithms cited above

use Heb bian learning; for the training of the "hidden layer" weights.

The [ollo\vlng Lheorelll is a COllt;equence of Theorem 2.1.

TH;;OKKv' 2.2 Let 11·11 denote an arbdrary norm defined for random vectors.

Assume that for any random 'vector Y, tht norm of v depends oni:1J on the dis­

tributlOn ofv. Let Ut be a stochastic process which converges to so'me vector u*

in the chosen norm: limt_.� IIUt-utll = O. Assume that (Xt,Yt) is a stochastic

process of identically-distributed random variables. Assume further that for all

vector u the followinq inequality holds:

IIP(yo,xo, u) - P(yo, x, u')11 :S [(llu- u'll,

'where f{ > 0 is a Iiud real '/w'/nber. Then lhe fulluwing 'kings huld:

1. The u'u.lpul weighls Vt lrained by rule (fi) cun'(lerge ijJ they cUJ/.'vtrgt when

lmined by lilt 'rule

(7)

Further. in case of converqence the limitinq val/1.es qenerated by the two

rules art equal.

Page 8: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

7 Synthesis of neural nehvorks

2. Let this limit point be v'" and assume that at = cit, where c > O. Then

the convcr.9cncc rate of IIVt - v"'ll is no worse than the maximum of the

conVCT.qcncc rates aivt and Ut.

:'Jote that the theorem is easily extended to the case when the condition that lIt

converges to sonw limit point 11* is relaxed to the assumption that 11t convNges

to one element of a set U* containing isolated vectors.

In order to ohtain a practical convergence rate let llS aSSUllW that llt i s

computed by some SA method and that 11·11, is the [' norm £[11·11'1'/2 where

E[·] stands for the expectation value operator. Kow under usual conditions

(Wasan, IfJ69,ilenveniste et al., 199(1)

IIUt - u'II" IIVt - v'lI, = 0 (�) ,

so hy Theorem 2.2 also

I IVt-v'1I2=O (�) .

It is easy to prove a theorem similar to Theorem 2.2 but for probability-

one convergence cases and when the norm I I ·11 is an ordinary vector norm; all

we have to do is to consider pointwise convergence over the probability space

underlying the sampling process. The precise restriction on F is that F should

be Lipschitzian in u, independently of its first two variables. This normally only

holds for bounded x and y, so we must assume that Xl and Yl are bounded with

probability one (w.p.l): ]f it is further assumed that the process (x"y,) consists

of independent, identically d istributed (i.i.d.) random variables and at = lit

then under sonle mild conditions on the forward algorithm the iterated logarithm

theorem of P. Major (Major, HJ7:1) gives that

'II · 'II (J

IOgIOgt) I IUt-U ,I IVt-V = 0 t

and eventually allows one to prove that

IIVi -v'll = u ( (8)

Page 9: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

holds with probability one.

V{e note that Theorem 2.1 enables us to extend Theorem 2.2 to the case

where F depends on v and just F should be a non-expansion w.r.t. v, indepen­

dently of the other parameters of F (it is sufficient if it holds for vectors in the

tail of Yd. The one-dimensional example F(u, v) = uvL shmvs that if F does

not satisfy this condition then the conclusion of Theorem 2.2 no longer holds.

In this case the limit of the process VI. converges to a fixed point or to infinity

depending on Va, tlO the 1nitial value::; of the procet;t: Ut can deLerm1ne the unal

behaviour.

Though the proo[t; presented belovv are qU1Le tll111ple thit: 1t; not counLed as a

drawback but rather as an advantage. The proofs are based on the classical ap­

proach to SA, i.e., the Robbins-1\,fonro formalism which reduces the convergence

of SA algorithms to the convergence of difference equations. Here we prefer this

formalism to the method of ordinary differential equations (ODEs) introduced

by Ljllng (Ljllng, 1977) and Kushner aud Clark (Kushner and Clark, 19(8) be­

<::allse of its simpli<::ity.

3 Application: LMS minimisation by hybrid

learning

:\Echaels proposed a t,,,,'o-layer: hybrid neural network that performs principal

<::Olnponent analysis (i .e.: de<::orrelation of the inputs) in its first layer and learns

<::orrelations (i .e. , performs Hebbian-learning) in its se<::ond layer (ivfi<::haels, 1095).

He ha.s shown that this network solves the a.sso<::iation of finite samples <::omposed

of input-output pairs with the minil1lUm LMS error. Here \'\'e exlend his analy­

sit: to in/inile random. samples under Lhe ast:umpLion that the training samples

are dra\""n independently frolll a common probability dit:tribution hav1ng a u­

nite second moment, It is demonsLrated LhaL the proposed learning algorilhm

solves the LMS (linear) approximation problcIll and also give an estimate of its

Page 10: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

convergence speed.

3.1 Preliminaries

Let E"t E Rtl., t = 0, 1, 2, . . . be i.i,d. randmll variables with zero mean and finite

second moment, and let f : Rn -----7 R1f! he an arhitrary hOllnded function. The

task is to find the hest linear approximation of I , that is the matrix lV E RIHx'fI

sllch that

f;(W) = �f;[IIW( - f(OII']

i� minimal (here and in Lhe following � denoLef:; a random variable wiLh the

same distribution as �t.) The classical increIllental solution to this problem is to

perform gradient-descent on the error surface Enr) (Widrow and Hoff, 1960).

Gradient-descent is guaranteed to converge to the optimal solution: since Enr) has only one minimum "\yhich satisfies the equation

(9)

where RJ((),( = E[f(Oe] is the cross-correlation matrix of the input and out­

put data and [(,( = E[�e] is the autocorrelation matrix of the input data,

Therefore the Widrow-Iloff LMS algorithm

incrementally complltes the produd of 1i.f(O,(. and inverse of matrix Ii('l'

3.2 LMS algorithm based on decorrelation and correlation

Assume that all the eigenvalues of RE.,€ are posit ive and that they are distinct

as well. LeL U be Lhe maLrix composed of Lhe orthonormaliseJ eigenvectors of

R,,( (LhaL is U UT = E, where E is the unit matrix) and leL A be the diagonal

InaLrlx whose £Lh diagonal elemenL corresponds Lo lhe iLh eigenvecLor in U, lhal

IS

(10)

Page 11: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

1 0 Synthesis of neural nehvorks

Rewriting (9) in terms of (10) yields

that is, if we let

th�ll VV" = QUT. Nmv let us introduce the variahle

Then Lhe optllnal LrvIS approxiula,Lion of J it; given by

y(O = Q,,(O·

llesides this Q can be written in terms of 1} since

Q Rj(O.(UA-1 = E[J(OeUA-'] =

E[J(O(A-1UT�n = E[J(O(A-'"n = Rri(lk'w

(11)

( 1 �)

(1�)

(14)

Equations (12) and (Ll) can be interpreted as the output computation per­

formed by a two-layer (linear) neural network. This idea was developed fur­

ther by xlichaels (Michaels, 1995) , who proposed that the learning of Q and

[/'1' should be performed by neural algorithms (though he considered the finite

sample case "\vhich requires a different treatment). There arc scvcral lcarning al­

gorithms that are known to converge to UT (see e.g. (Oja and Karhunen, 19S5,

Rubner and Tavan, 19S9,Baldi and Hornik, 19S9,Sanger, 19S9,Xu, 1993,Xu and Yuille, 1995,

Palmieri and Zhu, 1995,Chen and Chang, 1995) ) and Hebbian learning can be

used to learn the correlation matrix and the eigenvalues A. One possible algo-

rithm is shown in Table 1. The equation of Step 2 ma.y be repla.ced by any one

of Lhe many learning algoritllllls that learn the principal c.omponenLs of R�l. In

Step G diag(A) is defined by diag(A)ij = 0 if it j and diag(A)" = aii. Thus diag(Af' is easy to compute: diag(A);:;

' = 0 if it .i and diag(A);;' = l/aii. It is possible to avoid the inversion of the diagonal elements of At by learning

Page 12: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

I I Synthesis of neural nehvorks

Algorithm 1: LMS algorithm based on dccorrclation and correlation.

repeat forever{ 1. Generate the training data: (�t, f.), ·where f( = f(�t). 2. Learn the principal components:

3 Compute output of the first layer:

-1. Learn the eigenvalue matrix:

5, Learn the cross-correlation matrix:

6. t :� t + l. }

th� inverse of A. directly. This can he done by replacing the learning eqlli,dioll

of Step 4 by a special LMS rule:

,-1 _ ., -1 + (E ,-1, T) 1l..t+1 - �l.t O:t ' - ht 1]'1} .

3.3 Convergence properties

By Theorem 2.2 and the arg1lment that followed it , we have that the proposed

algorithm converges in the mean-sq1lare error norm \yith prohability one since

it uses only Hebhian learning in the cascaded part (Step 4 and ;5). Assuming

Lhe L'2 nonn of ranJOlll vecLors, Lhe rate of convergence of Lhe cascaded part is

O(l/-Vl} In general, the convergence raLe of Lhe whole algorithm \'"ill depend

on Lhe raLe of convergence of Lhe fon"ard algorithm. GOlnputer simulaLions

illdicate LhaL Lhis convergence raLe is also O(l/Vl) (see e.g. the experimenLs of

Xu and Yuille (Xu and Yuille, 1995)), so by TheoreIll 2.2 the convergence rate of

the ''''Thole algorithul is 0(1/-0), which is just the same as the convergence rate

of the LMS algorithm. Kote that the actual convergence rate of both algorithuls

Page 13: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

I � Synthesis of neural nehvorks

depend on the cigcnstructurc of the correlation matrix RE,f,o

Appendix

First we prove Theorem 2.1 through a series of simple lemmas and then consider

the proof of Theorem 2.2. VVe start '\vith a basic lemma concerning iterated linear

inequalities.

LEMMA 3 . 1 Let {kn},{an},{bn} be sequ ences alreal numbers and assume that

lilt lwrnberti k�!l (til sa li41J lhe sarne intquali ly

Then " " kn <:: ko II ai + 2:), II a).

;:::0 i:::D j:::i+l (15)

Proof- By assumption kl � aoko + bo and k2 :S al kl + b1. Substituting the esti­

LuaLe of ki into the laLLer inequality yielJtl k2 :::; alaok�o + alba + h. Contilluing

ill Lhio way we get (15). o

LErv:[MA 3 . 2 Let X and 1" be Banach spaces,. Lin: X X Y -----+){ 6e a sequence of

a rbit rary mappings (n = 0, 1 , 2, . . .j, and ret {}n E Y be a convergent series 1cith

lim.,t Bw (that !SIIB" - Bwll ron1Jprgps to urn). f'nrth pr, 1ft

and lei

with Xo = Yo. Then n n

(10)

( 1 7)

( 1 8)

where L� is the uniform Lipschitz index ofUk(x, 8) w.r.t. e at (}'XJ and, si mi la rly,

L% is thE uniform Li ps ch itz index of Uk(x, ()=) w.r.t. x.

Page 14: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

Proof. The proof is the application of Lemma 3.1 to the sequence On = I I XI]. -Yn II. Indeed, by the triangle inequality we have that

< IIUn(,r", eoc) - Un(Vn, eoo)11 + IIUn(Vn, eoo) - Un(v", en)11

< 1,�f,n+I,�llen-eooll

Thus On satisfies the conditions of Lemma :Ll with an = L;� and bn = L�], liOn -o

Part 1 of Theorem 2 . 1 follows immediately from this lemma since IIVn -"00 II s I lv" - ,r" II + II,E" - ,roc II· Now let liS consider the second part.

r.ir�i.l'y, �ome ex1.ra cond.i1.ion� are required 1.0 make IIY'll - :tn II converge Lo

ljero. The r.h.�. of (18) can be Laken a� 1.he "1.riangle" Lranf:;[orma1.ion of the

�eq uenee liOn - O.X,) II. If .6..n if:; a seq uenee of llumberf:;, and all,m: 0 < m < TI if:; a

Lriangular ma1.rix 1.hen Lhe Lranf:;[ormeJ sequence

n bn = L an,m.6..m

tn;:::Q

is called the triangle transform of .6..n with transformation sequence all.m. Trian­

gle transforms of infinite series for which L�,=o an,m is a constant independent

of n have been analysed in (P6lya and Szego: 1972). Here we consider a broader

class. A transformation sequence an,m. is called regular if the following two con-

ditions hold:

1 . lilIln.--+= an,m, = 0 for all m and

2. lil11n.--+= (L�,;:::o an,m) = c with SOIne c > O.

The constant c is called the asymptotic value of the sequence an,m. Regular

transformations actually keep the limit of the transformed seq1lence, as can be

show11 by the following lemma:

Page 15: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

14 Synthesis of neural nehvorks

LErvt:MA 3.3 Let a = (an,m.), an,m 2: 0 (n, m E N, O:S m S n) be an arbitrary

rcqular sequence with asymptotic value c. Let �n be an arbitrary convcr,qcnt sc-

rics with limit �. Let bn be the "trian,qlc transform)' of �fl with transformation

sequence an,m: "

b1/, = 2:= an,m�tJ1,. m=O

Then b" converges to c�" and the following estimate holds as well:

K Ibn - rill:S Ifnill + 2cHK + Hu L "n,m, (HI)

tn.=!)

where 0 ::; K S n is arbitrary (but it may depend on n)! en = c - L�=o an;m I

and

HK = sup liln -ill, n>J{

Proof- In th� following sequence of estimates

n

rw:::O K

< 2:= an,tnl�m - �I + rn=O

K < He L o,n,m + 2rHK + Ifnill,

rn=()

(20)

where 0 :S }( :; n is arbitrary, we have exploited the fact that Ull,m 2: O. Hence

(19) follows,

Now, let us consider the convergence of b"o Pick an arbitrary E > O. Then

leL us choose the number K so LhaL 2cIIK < fJ2, and choose .i.Y � K such lhal

kn.6.1 + JIo L�=o an,m < (/2 is sa.lisfied \"henever 11. � N. Such numbers exisL

because JIK a.nd (n converge Lo zero and also an,m converges Lo zero as 11. Lends

to infinity for any fixed m, Then by (21) if n ::> N then Ibn - eLI. I < f, Since f was arbitrary bn converges to c�. o

Page 16: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

1 5 Synthesis of neural nehvorks

The inequality relation (19) enables us to estimate the convergence speed of

bn. However, since the above inequality holds for any regular triangle transfor­

mation the resulting estimate will in general be rather crudc. vVc note too that

inequality (19) is strict in the worst case sense. This becomes evident when

considering; .6..'{I = � = 0 for all n. If the convergence of U'fI,m towards zero is

independent of m we can do much better in the derivation of the convergence

speed as shown by the next lemma:

LE:r..IMA 3 .4 Assume that rln.m decrwses uniformly in rn,! i.e.! that there exists

a functIOn h( n) surh that

(� 1 )

holds for all n. Let g : It+ ---, It be a monotonically decreaSing, posltzve-valued

integrable functIOn which satisfies

D.n S g(n). (22)

Then lilt triangle "ransfurm uf �n , bn! salit;fies lhe inequalit y

b" S h(n)G(n) + CJ(h(n)),

where G(x) = f: g(s)ds.

Proof. From LIte mean-value theorem of illtegl'al�

n-1

OS [q(S)dS - L g(i) S g(O) - q(n) -c: q(O), i=O

while from equa tions (21) and (22) we have that

bn S h(n) L g(112). In=O

Combining the above bro inequalities �yiclds the desired result. D

Note that if G(., ) converges to zero then we get the result that bn = CJ(h(n)),

while if G(n) does not converge to zero then bn = CJ(h(n)G(n)). For example.

Page 17: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

1 6 Synthesis of neural nehvorks

if an,m = l/(n + 1) (which corresponds to averaging) and fl.n <:: l/(n + 1)'

with 0 < s < 1 then we get that bn = O(l/(n + 1)'), i.e . . averaging docs not

change the speed of convergence. Note that although the primitive function of

g(x) = y'loglogx/x cannot be written in a closed form, clearly

/' 16 Jlog log x/x dx <:: JT log log T

which can be casily verified by differentiation. From these follow our extension

of the iterated logarithm theorem of P. �Jajor to cascaded Hcbbian learning.

touched upon in Section 2 . In order to derive this and the convergence rate

estimate of Theorem 2 .1 , simply observe that if L� :S c/(n + 1) and L� :S

1 - 1/(n + 1), then after some computation equation (1S) will reduce to

n Ilxn - Yn l l <::

n: 1 L liB, - Bnll 3=0

The following; lemma completes the proof of Theorem 2. 1 :

LE1IMA 3 .5 Let xn,Yn,L�"L�,:Bn be the series as defined in Lemma 3.2. As­

�nl,rne Uw l lIltre e:C'is l s ([ cons lant C > 0 s'Uch lha t

L�. <:: e(l -L�) (24)

hold,4 jor all n 2" 0 and also for all k 2" 0

.::xJ

II J.� = O. (�5) i=k

Then lin,,_oo Ilx" - Yn II = n.

Proof. LeL �n = liOn - 020 II and Cln,m = L�. TIj=m+l Lj ""here we have w�eJ the

convention again that an empty product (i.e., when m � n) is equal to 1 . Ob­

serve that by Lemma 3.2 II;cn -Yn II can be estimated from above by the triangle

transform of �n with coefficients an,m. Since L� � U(1 -L��J then (In,m �

i Note that this condition requires that L;' :::; 1 holds for all n since L;, 2: u.

Page 18: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

1 7 Synthesis of neural nehvorks

bn,tJ]., where bn,m = C( 1 -L::n) TIj=m+l L]. Therefore since L:�=IJ an;m�n ::;

L��=o bn,mD..n, it is sufficient to estimate the sum L��=o bn)m.�n from the above

to obtain an upper bound on Ilx n - Unll. Now, observe that (bn,m) satisfies the

assumptions of Lemma 3.3: bn,rJL 2: 0,

and

00

nl�� bn,m = C(l -L�) II LX = 0 J

n n n � bn.m = C 2:)1 - L�,) II m=1J m=O j=m+l

LJ! J

which is a telescopic sum that reduces to C(I-[1j�o Lj) after expansion. There­

fore n lim " bn.m = C·(1 - II Lj) = C > o.

n-co L-m=O j=O Tl lUtl LellllIla :1.2 can be applied and \ve tlee that Ilxn - Ynll converges Lo :L,ero.

o

Now let llS consider a c(l.<;;caded Hehhian algori thm of the form

where the involved quantities satisfy the conditions of Theorem 2 .2 . First, ob­

serve that this corresponds to a diagonal process Yn of form (2): with Yn = Vn, Bn = Un, and Un(Un,fJn,w) = F(Yn:xn,un). (In order to make the differ-

entiation easier between the symbols of the two contexts, symbols related to

the cascaded Hebbian algorithm have a boldface typeset, while the symbols

of the Banach-space context have a normal typeset.) UTi does not depend on

fJn so the Lipschitz constant of Un w.r.t. fJT! is zero. Therefore the possible

L.ipschitz constanLs are L�, = 1 - an and L� = f{ Ctn, where f{ is the Lips­

chitz constanL of F ,\\T.r.t. U defined in Theorem 2.2 . Since we assumed lhal

L�o Ct't = 00 the LipschiLZ constanLs satisfy the conditions of Lemma :1 .. ) (in­

equality bound (24) and equation (25)) thus IIVt - Utll converges to zero, and

hence by Lemma 3.211Vt -v*11 also converges to zero (here v* denotes the limit

Page 19: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

I S Synthesis of neural nehvorks

of process Vt+1 = (l-at)Vt+atF(Yt,Xt, u' ) where limt_GC Ilu'-Utll = 0). The

convergence rate estimate follmvs from the properties of the Lipschitz constants.

References

(Haldi and Hornik, I !!S!!) Hald i , 1'. and Hornik , 1<' (1 !IS!I). �e1lral net.works and

principal component analysis: Learning from examples withOllt local minima.

/\'eural /\'dworks, :!:,)�-58.

(Henaim, 1 DD:)) Henaim, .\'1. (199:)). The (:off l ine learning approximation;' ill

conLilluou� Lime neural neLViOrkf:): an adiabatic Lheorem. N fund .:Vei'U)orA:s:

5( ,j) :6f.ifi-oGf).

(Benallll, 1996) I3 enal111 , M . (1996). A dynamical �yslelIl approach La f:;Locha.<:;Lic

apPfOxlula,Llom;. 8IAJf 1. Contrul and Oplim,i.?aLiun, ;14(2):437-472.

(I3enailll and TOlnaslni, 1991) I3 enalm, M. anu Tomasini , 1. (1991). Compet­

itive and self-organizing algorithms based on the minimisation of an infor­

mation criterion. In Artificial �Veural �Vetworks f.. volume 1, pages 301-306.

Korth-Holland, Amsterdam.

(Benaim and Tomasini, 1992) Bcnaim, M. and Tomasini, L. (1992). Approxi,

mating function and prediciting time series with multisigmoidal basis func­

tions. In Artificial Neural },letworks If., volume 1, pages 407-411. Elsevier,

Amsterdam.

(Uenveniste et a!., liJ!JO) Uenveniste, A. , I\Wivier, M., and Priouret, P. (li)!)O).

Adaptive Algorithms and Stochastic ApproximatlOns. Springer Verlag, New

York.

(Chen and Chang, 1995) Chen, L. and Chang, S. (1995). An adaptive learning

algorithm for principal component analysis. IEEE Tran. on .l'lev,ral Netlvorks.

6(5):1255 126;1.

Page 20: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Hi Synthesis of neural nehvorks

(Fomin et a!. , 1994) Fomin, T., Szepcsvari, C. , and Lorincz, A. (1994). Sclf­

organizing neurocontral. In Proc. of IEEE Y-l"CCI IC]"{S'94, volume 5 , pages

2777-2780, Orlando, Florida. IEEE Inc.

(Gullapalli and Barto, 1994) Gullapalli, V. and Barto, A. (1904). Convergence

of indired adaptive asynchronous value iteration algorithms. In Advances

in N eu.ral Information Processing Systems 6, pages 695-702. San Francisco:

lVlorgan Kauffmann.

(Haykin, 1904) Haykin, S. (1994). Neural networks: a comprehensive founda­

tion. MadJillen College Publ. Co., New York.

(Hecht-Nielsen, 19S7) Hecht-'1ielsen, R. (19S7). Nearest matched filter classifi­

cation of spatiotemporal patterns. Applied Optics, 26(10).

(Joutsensalo et a!., 1005) J outsensalo, J., lvI iettinen, A., and Zeindl, Iv!. (1995).

Konlinear dimension reduction by combining competitive and distributed

learning. In Prot. oj ICAA']'{'.95, volume 2, pages ;19,]-·'100, Paris.

(Kushner and Clark, 1978) Kushner, II. and Clark, D . (1978). Stuchastic ap­

pro;C'inwfion Tnellwds Jo!' cun!:iLraintd and u'fIcU'fu:drai'fl td systemt;. Springer­

Verlag, I3erlin, Heidelberg, Ne\v York.

(Ljung, 1977) Ljung, L . ( 1977). Allaly�i� of recur�ive f:)Lochaslic algoriLllIn�.

IEEE Trim. A'Ulo'ffwl. Co nL-,.ol, 22:.].]1-.]7.] .

(Major, 19n) Major, P. (1973 ) . A law o[ the iterated logariLhm [or the Il.obbins­

rvlonro method. SIHdia 8cienliaTum lvlalll fmalicarum IIungarica, 8 :9.]-102.

(Martinetz and Schulten, 1991) Martinetz, T. and Schulten, K. (1991). A

'�neural-gas" network learns Lopologies. In Proc. oj ICA�Vi\;-, volume 1 , pages

397-402. Elsevier, Amsterdam.

(Michaels, 1995) xlichaels, R. (1005). Associative memory with un correlated

inputs. Neural Computation, 8(2):256-259.

Page 21: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

Synthesis of neural nehvorks

(Moody and Darken, 1989) xloody, J . and Darken, C. (1989). Fast learning

in neural networks of locally tuned processing units. N cura l Com pl1.tati on ,

1 :281-582.

(Kowlan, 1990) '1owlan, S. (1990) . M,,"'Cimum likelihood competitive learning.

In 1'ro c. of Ne ura l Inf ormation 1'ro ce ssing Sy stem s, pages 074-582.

(Oja ond Karhllllen, HIS.;) Ojo , �:. and I{orhllnen, J . (HIS.»). On stoch8Bt.ic

approximation of the eigenver.tors and eigenvalues of the expectation of a

random matrix. J. o f .Math . Ana L and Appl £(:(}.t io o8, 1 06 :6D-S4.

( Palmieri ano �hll, 1 995) Palmieri, �'. ond �hll, J . ( 1 9%). Self-associat.ion ano

IIeLbian learning in linear neural network�. IEEE Tum. on A'e u ml .:Ve ftvorA:s :

6(5) 1165-1184.

(Poggio and Girosi, 1990) Poggio, T. and Girosi, F. (1990) . Regularization al­

goriLluIls for learning that are equivalent Lo multilayer neL�;orks. Scienct,

247:979-982.

(Polya and Szegii, 19(2) Polya, G. and Szegii, G. (1972). Problem s and Th eo­

rem s in Analys is 1. Springer-Verlag Berlin.

(Rittcr ct a!., 1988) Rittcr, R . , Martinetz, T. , and Schulten, K. (1988). Topol­

ogy conserving maps for learning visuo-motor coordination. �Vcl1.ra l _Nct wor ks .

2:159-168.

(H.obbins and Monro, 1il01) H.obbins, ll. and Monro, S. ( lD" l ) . A stochastic

approximation met hod . Ann . Mat. Stat . , 22:400-4117.

( l-tuhner and Tavan , HlSD) I-tuhner, .J . and Tavan , P. ( I DSD). A self-organir,ing

network for principal r.omponent analysis. !".:umphy8. I-,rtt., 1 O :6::Hl-6SD.

(Sanger, 1 989) Sanger, T. ( 1 989). Optimal unsupervised learning in a single­

layer linear feedfonvard neul'al ndw·ork. JVe ural j\le l'lro r ks, 12:459-473.

Page 22: Synthesis of neural networks: the case of cascaded Hebbiansszepesva/papers/TR96-102.pdf · Synthesis of neural networks: the case of cascaded Hebbians ... Synthesis of neural networks:

�1 Synthesis of neural nehvorks

(Szepesvari et a!., 1994) Szepesvari, C., Balazs, L., and Lorincz, A. (1994).

Topology learning solved by extended objects: a neural network model. N cu­

ml Computation, 6(3):441-458.

(Wasan, 1969) Wasan, T. ( 1969). Stochastic Approximation. Cambridge Uni­

versity Press, London.

(Widrow and Hoff, 1 960) Widrow, H. and Hoff, Ivl . (I\ J60). Adapt.ive swit.ching

cir('.llits. In 1960 Jl-U",' Wf�'SO.t\r Convention Hecord, pag�s D6-104, New '{ork.

I-teprinted in : iVellrorompllting, .LA . AndNson and �:. I-tosenfeld (eeL) , The

MIT Press, Cambridge, Massachusetts, 1988, pp. 12(1 LH.

(Xu, 199;1) Xu: L. (1993). Lea�L lIlean r:;qual'e error reconr:;LrucLioIl princ.iple for

f:)elI-organiz,ing Ileural-neL�. lYeul'al �Vt hL'OTks, 6{,:.I ) :627-648.

(Xu and Yuille, 199,») Xu, L. and Yuille, A. ( 19%). Robust principal compo­

nent allaly�i� by �el[-orgalli1jing rule; based on statistical physics approach.

IEEE Tran. on Neural Networks, 6(1) : 131-143.