Storage capacity of a Potts-perceptron - Journal de Physique I

4 downloads 0 Views 588KB Size Report
«Potts-perceptron» with. N. Q-states input neurons and one. Q' states output neuron, we compute the maximal storage capacity for unbiased pattems. In the.
Phys.

J.

France

I

(1991)

1

l109-l121

AOOT

1991,

PAGE

l109

Classification

Physics

Abstracts

05.20

87.30

Storage

capacity

Jean-Pierre

(')

Lab.

Cedex

Nadal

f) Dept.

(')

Physique

de

05,

of

Potts-perceptron

a

and

AJbrecht

Statistique (*),

f)

Rau

Ecole

Supdrieure,

Normale

24

rue

Lhomond,

F-75231

Paris

France

Theoretical

of

Physics,

University

of

Oxford,

I

Keble

Rd.,

GB-OxfordoXl

3NP,

G-B-

(Received

22

January

1991,

accepted

in

final form

30

April 1991)

networks where each be in consider the properties of « Potts » neural Abstract. We neuron can with different For «Potts-perceptron» N Q-states input and Q states. a neurons one Q' states output pattems. In the storage capacity for unbiased compute the maximal neuron, we be stored is found to be proportional to of pattems that large N limit the maximal number can f(Q'), where f(Q') is of order 1. N(Q I

Pom-perceptrons,

1.

The

Hopfield

and

was

linear

machines

and

winner

take

all

systems.

[I] Of a formal neural network used having two possible states, neurons extension of this model of analogy with spin glass systems. A natural associative taking more than instead of having is to consider two states : memory neurons neuronal described by an Ising like variable, one has a Potts [2] like variable. One then state obtains a neural network Potts [3], which is to the Potts Glass [4] what the Hopfield model is to the Spin Glass model. The physics of Potts neural networks has been studied statistical for networks with Hebbian leaming rules ([3, 5-7]). In these studies, one is neural attractor considering a Potts-Attractor Neural Network ~PANN) : each whose activity can neuron, take Q different values, is connected to every other and the stimulus (a to a neuron, response given initial configuration of activities) is the which the dynamics of the natural attractor to leads. Clearly one Hebbian learning rules and in particular also consider net can non non symmetric couplingsfor the of binary One also consider Pottsas case neurons. can feedforward networks. Perceptrons (PP), that is ~possibly multilayer) In that the case, of possible layer. the following number In will consider only states may dirtier from layer to we the simplest that is one input layer with Q-states and output layer with case, neurons one Q' states (and no hidden layer). Such systems are well known in data analysis neurons literature under the of «linear machines» [8]. The binary [9] is one perceptron name particular case. In fact the perceptron algorithm, as well as its variants, is easily adapted to multistates ([8, 10]). neurons (*)

based

Laboratoire

model on

an

associd

au

CNRS

(URA1306)

et

aux

Ulfiversitds

Paris

VI

et

PaHs

VII.

II10

JOURNAL

PHYSIQUE

DE

I

lNf 8

considering such systems are manifolds. Consider the of a I) case just mentioned~ such allows deal with multiclass system to classification tasks, and with strings of data made of N items~ each one taking Q possible different values. Here few examples. In image processing, Q Q' would be the number are of grey each input node would be in one of the levels. In the analj,sis of DNA sequences~ the might be binary for codon Q 4 letters~ A, T~ G or C and output exxon see versus interested in predicting the secondary structure e.g. [I I]). In the analysis of proteins, if one is number from the of amino acids, Q would be 20, the of different acids, and sequence amino (a helix, p sheet or randomj iii Q' would be 3, the number of different possible structures Consider feedforward with N inputs and Q'-state output One network one neuron. a can binary winner take updating consider Q' with all rule, the output as neurons, a neurons » receiving the highest input field will fire. Such systems commonly and only the neuron are both for biological modeling and engineering applications. There are found in the literature, the competition models of self-organisation based between in particular (see e-gon neurons mentioned the fact that of [l2]). iii) We have already such natural extension system are a addition, In for a from the physicist point of view. multidass binary neural networks classification task (including the two class, binary casej, the use of a Potts perceptron neuron theory (see e-g [8] (as precisely defined in the next section) can be justified in Bayes decision indications that p.16 and [11]). iv) Lastly we note that, on the biological side, there are coherently, having a small of possible cortical columns behave number coherent [13]. states neural network insight on the behavior of a network of Hence a Potts-attractor may give some The

for

motivations

feedforward

As

network.

=

=

cortical

columns.

Here

will be

we

classification the neural the

pattems~ present the

v'

network.

define

first

us

specifying

encoding

compute

the

chosen

be

for

neurons)

associative

algorithm. the

maximal

fractional

limit

biased

to

properties. the

volume

will

We

patterns. extended

patterns.

will

We

show

capacity

storage

jPANNj or particular we will storage capacity of a of the weights which study to unbiased our memory

In-

its

results.

In

section

relevance The

?

we

to

the

details

of

Appendix.

the

dynamics.

and

model

the

inputs (Q-state is given bv

[14] for

will

in

given

for

Gardner we

the

invariance gauge section 3 we present

In

be

networks

leaming computing

randomly computation can

discuss

will

neural

Potts

of

set

and

invariances,

Gauge

Let N

by Namely,

learning of a although the model

with without

initiated

network.

dynamics of the computation the

2.

(PP),

approach

follow Potts

realize

concerned

tasks

simplest

the

and

one

h~

=

feedforward

Q'-state

output

jj (js',s)

3~~

the

case,

The

neuron.

Potts

local

perceptron field

at

with state

(lj '

J.~

The state

synaptic s~

on

ii..Q)

.v~

jn~,j

I... =

lip,

T

is

((s', s)

matrix the

and

N),

s'~

with

defined

the weight of a signal coming from node j, which is in processing unit. otherwise indicated ii.. N ), Unless j ~ A input activities denoted (I.. Q'). of will be pattern probabilist il. The decision rule Q ). at «temperature» ~ a

indicates

the

of

s'

state

n~

by

=

s[~~

s' =

with

probability

exp

jj ~

p (h~,

exp p

(fi,.

H~,

0, j

(2j

M

STORAGE

8

the

where

by

thresholds.

the

are

OF

CAPACITY

s[~~

At

(s(

=

POTTS-PERCEPTRON

A

rule

is

s()

0~, Vs' #

@~,) h~,

h~~

decision

the

temperature,

zero

II II

(3)

o

winner-takes-all consider This is a « rule : the Q' state output neuron » can as one s' receives the input field h~, defined by (I), and only the Q' binary neuron neurons neuron with the highest local field will become active. The generalization to N' output neurons or to a is straightforward. Indeed, fully connected net (PANIQ~ for which N' N and Q Q' for the basic properties considered in this section as well for of binary as for the case neurons, the analysis in the next section, one can always focus on one particular (output) storage which is equivalent to consider with only one perceptron output a neuron. neuron, invariant under the following translations Updating rules (2) and (3) are =

( (S', S )

( (S', S )

-

@~,

and

=

0~,

-

(S) j

(4)

uo

+

under

((S', S) @y

where

the

uo,

local

each

to

output

uj(s) field

thresholds.

In

of

set

addition

translations

updating

the

positive

gauge

invariances

possible

One

s(,

strictly

and

authors

to

set

to

([8, iii). another

make

choice zero

It is

choice

real

number

allows

namely

under

(4) add

transformation

first

absorbed

be

can

independent

is

the

in the

(6)

redefinition

global rescaling

a

of the

fields by

local

a

of

:

j3-p/A

(7)

parameters

by fixing the gauge ». particular output state

@~,-A@~,,

A.

reduce

to

that

however

input

state

so

( (s', so).

and

choice

gauge

any

of

number

the

is to pick one particular all the couplings ( (s(, s) clear

The

input pattem, couplings (5) modifies invariant

is

rule

(6)

numbers.

then

j(s',s)-AJj(s',s), any These

(5)

of the

the

on

which

pattern,

vj (S')

+

£ uj(s')

@y

-

arbitrary real are although a function

input

of the

((S', S)

-

vj(s')

which,

term

second

independent

the

the

and

a

The

states.

term

for

+ U

and

one

This

is

will

do.

In

a

choice the

made

by

several

section

next

we

will

:

Z((s',

s

o ;

vs'

(8)

=

o ;

vs

=

(9)

)

s

Z((s',

s)

~.

by considering the dynamics, is more natural. we argue as now dynamics of the network, whether it is an neural attractor net or a quantities. This can be easily shown should only depend on gauge invariant dynamics. For simplicity, consider a perceptron with NQ-states inputs and (Q'= 2). In that case one can write the updating rule (3) as

which,

feedforward

The

"

with

w~(s)

=

((2, s)

((I, s)

"

(h) h

sgn

and @o

=

@~

=

£

j,

w~(n~) and

«

for one

the

binary

±

I is the

step

output

(10)

@o

=

net,

one

binary

output.

Suppose

JOURNAL

II12

that

now

for

As

the

binary

the

variable~

presented

perceptron here

that

so

input

the

is

pattern

n~

=

n~

=

[15] one probability

version

I

of

M

teamed

a

pattern,

8

pattern

say

probability f

f'

getting

for

(l j

Q

that~ for large N,

show

can

~

probability

with

# n~'

s

noisy

a

with

n~'

PHYSIQUE

DE

the

~'"lfij

input

the

field is

Gaussian

a

is

output

correct

~~~~

where m

H(x)

i13)

Dy,

=

~

Gaussian

the

being

measure

by

indicated

Dy The

first

two

conditions

with

the

(h)

of h,

cumulants

(h~)~ (the

and

fi,

bias

same

(- y~/2 if,.'2

exp

=

j14)

ar

being

averages

I li(nj')l~

tn)

(i

+

=

j

tn)

the

renormalized

couplings

initial

and

I li(.I)l~ ,

,

and

possible

jj ((n~')-0

=m

J

tn(i

all

given by

are

(h)

lh~l~

over

threshold

j(s)ww~(s)

~

by

given

are

list

zw~(i)

Q

,

0

The

parameter

noisy

the

which

Hi

appears

w

characterize

above

the

=

~

is

the

natural

in

the

and

pattern

the

(17)

~

value

~~

pattern

between

II, ,,

m

It

correlation

m(Q3~i

l ;

takes

the

(16)

version

(Q3~ and

jjw.,(t). Q~,

Ho

case

generalization of binary

of neurons

the

usual

IQ

2 =

~

(18)

Q

=

«

).

overlap It

takes

between the

value

the

if

input there

is

string no

with the pattern. It is important to notice input, and 0 if the input is uncorrelated concentrated dependance in the initial conditions is in this The above parameter. easily be generalized finite also Note that, the of temperature. to can in case

and

noise

in

that

all

formula an

the

the

the

j12)

attractor

M 8

CAPACITY

STORAGE

characterizes network, (12) fully [15]. conclude, it is thus natural more

neural

dynamics

the

II13

POTTS-PERCEPTRON

A

OF

the

in

of

case

highly

a

diluted

network To

hy

and

explicitly

to

fixe is

=

threshold»

particular,

point

this data

way

=

a

it is

useful

to

make

in

the

the

as

for

field

neural

a

comments

some

input layer.

string of Q nodes. The direct interpretation Q-bit string with only the nj-th bit on. In site j.

at

This is the

consider

most

of

choice

common

representations. Q nodes

other

configuration

Q

the

the

((s') j, Q

the

(xf,n I. representation,

vectors

thermometric

«

(I)

of

»

whose

vector

different

for

values

3.

forming

where

a

given

are

of the binary representing the

case

capacity

choice

of

and

of

a

the

letters

first

is

is

field

coupling

the

to

=

RQ. A

all

What

on

the

can

be

from But

as a

node

one

by

can

given

a

have

couplings layer of any

the

choice,

that

definition

0

from

and

shown

is

(19)

there

neural

adopt for

s

that

in the

the

local

the note

we

that

thresholds.

of the

feedforward we

If

is

node

is

matters

just

example

otherwise.

coming

what

we

typical

and

nj,

coupling

(3) with the

rule

of I

s

The

».

that

each j, the activity is encoded the s-th state is represented by

((s',s)

base

a

for

a

for

encoding problem, (I) of the local

Q),

the

given j. reinterpretation of s'

that

case

stated,

Otherwise

considerations

Storage

I

is

component

s

((s').x".

equivalent, via a apply as well which the input data are strings of corresponds to take the updating choices (8) and (9). gauge

choices

the

in

encoding found in the literature. defined general case, is state n

xi

=

for

is

that

the

In

the

entries

of

Q )

=

((s',n)=

then

these

choice, standard

definition

xn=(g,s=i. with

(19)

l)

on

The

with NQ

net

a

also

as

=

represented

are

understood

s

fields

s) (Q3~,~~

by (8) and (9). This 2, corresponds to the

gauges

for

activity by

At

£ ((s',

=

local

the

Q' Q With these choices, the case « without spin like variable, ) ± I. the between the Q is the only situation which states : in symmetry preserves the optimal thresholds will unbiased be when leaming random, patterns, zero. that

perceptron, neuron

the

define

to

site

at

are

possible

all

that

Note

network

for

following, fieldi, and

thus the

Pom-perceptron.

consider with N inputs, each We now the storage capacity of a Potts having perceptron one will number Q possible states, and one Q' state output We the maximal of compute neuron. input-output pairs (n~P, j which random for there exists I N ), n'~ ), p I set p, a of couplings such that, for every p (p n'~ if the input is I,... p ), (3) is true with s[w =

=

=

=

number

pattern which

is

p.

equivalent

All

Potts

states

in

occur

the

the

thresholds The

pattern

average to

translational

with

pattems

are

the

probability,

same

is

indicated

by (... )

=0. f

J

Here

pattems

to

lQ8~~»-1)

the

of

set

the

Since ~.

unbiased,

we

can

set

all

zero.

freedom

of the

£ ((s',

field

local

s)

0 =

is

fixed

Vs' =

I

by Q'

(20)

JOURNAL

l14

Z((s',

PHYSIQUE

DE

s)

vs

o

=

M

I

Q'

i

8

(21)

=

~.

should

It

that

noted

be

the

constraint

the

scalar

of (20) for s' Q' degree of freedom

is

included

(21). We

in

=

only the .v' by

when

case

((s', s))~

£

is

consider

will

independently

fixed

for

(22)

Ny

=

each

j,s

The

constant

of

is

y

arbitrary,

course

y

the

From

each

For

invariance

gauge

there

pattem

expect

that

N(Q

i

the

we

maximal

) (Q'

see

i

),

(Q

i

that

there

that

will

this

that

see

is

the

=

a

which

(Q

be

to

can

I

) free

satisfied. will

stored

be

parameters.

Thus be

one

of

can

order

(24)

N

slowly

a

IQ'

I

need

that

I

a

(23J

only N(Q

are

take

we

)/Q'

i

patterns

with

case~

convenience

(Q'-

of

p~~~ iS

p~~~ We

later

inequalities

I

number

)/(Q'

i

for

=

(Q'-

are

but

varying,

bounded,

function

of

Q' only. the

s,;rem

entropy

method,

&[h~(n'" )

h

maximum

and

we

define

the

partition

function

as

dp (J)

Z =

h"(s')

Here

[14]

Gardner's

follow

We of

is

local

the

fl

fl

»

~j~~.~j

field

is

jj ((s',

h"(s') =

dp (J) is the

input

the

when

"

,~)

is')

(25)

x

pattern

p~

s) (Q3~_~~

),

l

in the space of interactions which is compatible with (?0)measure asking for minimal stability the of the (In Q Q' we are a 2~ case x. / defined differs in (25) by from stability factor the standard parameter as a x introduced by Gardner [14]). In order to evaluate the quenched average the distribution over of (ln Z) the replica trick. of the Assuming the validity of a entropy S patterns we use and shrinking the replica symmetric solution volume of interactions find for the to zero we capacity a (as defined by (24)) maximum storage and

(22). As stability

in

=

~~

_, "

H~ 2

Dy is the

Here

natural

[14],

~

(l

~° ~~~

~~~~~~ ~

Ho)Q'-

+

(Ho)~

Q'Ho

(14) and

measure

H, (_i')

«

~

Dt(t

m x

we

+ j'

have

+

x

introduced

I'

;

I

=

j,'2 ~~~~

(Ho)~

Q'(Q'-

Gaussian

Q'(Q'-

=

functions

the

(27)

0, 1, 2

i

derivation of (26) is given in the Appendix. The justification for restricting to symmetric calculation is the for the usual, Q Indeed, any Q' 2, case. same as continuously transformed solution other solution, be from be into the can any as can seen geometrical picture (one particular solution is a set of Q' vectors with well defined angles them, and one go through any between solution by global rotations). Details

the

of

the

replica

=

=

M 8

STORAGE

For

(x

a

=

Q'= 2 0) 2.

K

=

finds

one

one

=

=

limit

~

,

~

~

the

particular with in capacity analytically, and

Gardner,

storage

-1

$

~

II15

~'~~~'

~

~~~~

gets

one

«

~~

by

found

evaluate

can

/

large Q'

the

3

:

~

In

Q'

0 and

POTTS-PERCEPTRON

A

capacity

critical

the

recovers

one

For

OF

CAPACITY

(Q,

»

1)

DyiH~

=

(H~)2/Hoi

»

(29)

3.850.

a 3

Fig.

Maximal

I.

The

results

table I. In Table

capacity

for

4

2

-0

x

0 =

Q' as

function

a

Maximal

storage

capacity

at

K

that

critical of

about

capacity 3.85.

as

measured

However,

the

by

a

2

2.000

3

2.320

4

2.546

5

2.714

6

2.844

is

for a

K

0 =

function

shown

are

of

K

in

figure

I

and

Q'= 2, 3 and 4.

for

a(Q')

j.850

co

The

Q"

0.

=

o'

value

of

for solving the integral in (26) numerically figure 2 is shown the maximal capacity as

I.

lo

8

6 ,

thus

information

an

increasing function I(Q'), in nat

content

of per

saturating at (free) parameters,

Q',

a

is

I(Q') =

«

(Q')

in

Q'/(Q'-1),

(30)

JOURNAL

II16

PHYSIQUE

DE

lNf 8

I

25

j ~

>

~~

io

0

~

io

os

is

lo

is K

Fig.

Maximal

2.

capacity

as

a

function

stability

the

of

parameter

K

Q'

for

2.

3

and

4

infinity. Hence we find that~ by Kanter in the case of to one mentioned in the introduction~ there as are find of whenever which allow couplings there exists at algorithms type set perceptron to a solution ([8, 10]). such algorithms allow to respect any particular Moreover, least gauge one choice. To see this~ let us give the algorithm for the choice of the gauges (20) and perceptron with couplings. Then, the following is repeated until (21). One starts convergence zero Pick a pattern p at random. For any s'# ii'" such that hf ~ hf~, make a learning step for every j and every s by

decreasing function of Q', going to zero a qualitatively, the optimal behavior with Q' is Hebb rule. It is important that, to note a

/(n'", s)

((s', It is

clear

that

at

each

s

j

time

step

the

storage

-

-

the

similar

i(n'", s) j(s',

Q'

when

is

to

goes

found

the

(Q3

+

i

nj

(Q3s,n/

s

couplings

current

1) will

(3'

satisfy (20)

and

(?1).

Conclusion.

In the

this

article

we

maximal

analysed

number

of

pattern

that

capacity can

be

of

a

Potts

We

perceptron.

proportional function of Q'. Our

stored

is

to

N

Q

have

obtained

I ), the

being a slowly increasing and bounded calculation can generally, generalized to other cases~ e-g- the case of biased More it is clear patterns. of the analysis done for the binary Q' could be generalized to 2 perceptron (Q most it interesting consider function than Potts In particular would be other to cost perceptron.

a

=

number

of

errors,

as

considered

that

prefactor easily be that the

=

the

consider the above saturation. here, and to properties (We method, simpler than the replica techniques, has been derived

recently ([19]) a function. in order to study binary perceptrons at and above criticality, for any choice of cost checked that it can be easily applied to Potts-perceptrons.) In practical applications We have of neural networks, take all systems currently used. This is presently the main winnner are of for pursuing the analytical study such systems. reason note

that

very

M 8

STORAGE

OF

CAPACITY

POTTS-PERCEPTRON

A

II17

Acknowledgements. Normale Supkrieure for their kind of us (A.R.) thank the group at the Ecole wants to further acknowledges financial by the hospitality during his stay there. He support Studienstftung ties dbutschen Volkes, the SERC and Corpus Christi College, Oxford. This and BRAIN twinning (ST2J.0312.C(EDB) work has been supported by EEC contracts discussions. ST2J.0422.C(EDB)). We would like to thank Marc M6zard for several

One

Appendix. order

In

(In Z)

evaluate

to

we

~

identity

the

use

(

till z)

=

I

n

=

and

measure

the

on

fl

m

d

dA(a mm

I

exP

x

js~

fi

la

Ala

(it)

N

«

introduced

we

A(~

dP (J~)

(I)

0 ~

,~

~,j,n,~~

p.a

a

d4a

fl

fl

dP (J~)

z"

The

I

consider

thus

where

/

=

~

n

We

)

Z"

lim

m

£ ( ((n'~, is

Q~fl

pass'

(f(s',

)

)) (Q3~, ~

s

l

)

(iii)

~

couplings

(f(s', s )

s

by

defined

'

£ ((s',

~jS'

s

)

S

fl

£ /f(s',

~jS

S'

)

s

x

jj /f(S',

§

~

~

S)~

~~'

~~

Ny

~~~~

~

where will n

V is

a

compute

limit

normalization

(Z"),,

the

constant

and

average

of

the

the

over

nN

the

patterns

are

unbiased

one

(Q has

f

I

normalization

distribution,

pattem

in

(22). the

We

small

=

l

~~

) (Q8

performing

the

average

exp

nN

'

the

over

(Q

I

) f =

#

~

~

Q3 f

fl fl ldq()~'exp (G~ a.b

v.

(vii)

l ~

gets

one

n~P

))

J

(vi)

0 f

l

~~

f

J

(vj ~

l

~~ J

((Q8~

jZ"j

:

(Q3~

Then

fixing the

constant

:

exp

Since

y is

Z"



+

Gj)

(viii)

JOURNAL

1118

PHYSIQUE

DE

M

I

8

with

G~

exp

=

fl ~

x

fl ~~~

~

«

dA

d,~(~

j~

exp

(- I,;[~ A j~)

I

-~(a

x

$fi

~

£

exP

x

ah,~,

x)b lqli~

qli"'~

qli~

~'~

+

qlfi"I

fix)

~~~,

and

exp

G

j

ill

=

dp (J~)

x

~

ZJ~(S' a

~~

~

','

~~

Ny

s=1+> fl ~-~

~

k

with

critical

~

Q'

'

+

gives

which

~~~

x

t

taken

are

fl

~~~

~'

(X,

s'=i~-K

jj i=i

~~~

(3' X~.j

we

k

x

~'~'~' ~~

~~

(x~,»j

(Z' X~ +

following

express>on

for

the

capacity

i1

i

Q'-1

~

Pmax

Qj'

Q'~ k

~_j

Q '~ ~~~ ->

~~~ x

~

j

i

~



=i

Qj'

~~~,

~-~

-i~i

,

~

~~~

~

-~

2

~'

jj ~.~

jj z) k+

,'Q' ~

~~

,

~~

(~X'lll)

bt 8

STORAGE

Introducing

Gaussian

a

CAPACITY

variable

in

OF

order

~

~~

(

to

decouple

~~

(~~~~)

one

sum

can

k

over

kA~

using the

and

~~,,

identities

'

(I

KA

A)~~'

+

(xxix)

k

i

and

the

K =

k

l12I

POTTS-PERCEPTRON

A

finally

arrives

expression (26)

the

at

for

the

Lent.

55

critical

capacity.

storage

References

[Ii [2]

WU

[3] [4] [5]

[6]

J. J.,

HOPFIELD PoTTs

B,,

R.

F. Y.,

SHIM

D.,

DUPONT

G. M.,

KIM

R. O.

DUDA

MINSKY

1990

S.,

RUMELHART

PERETTO

GARDNER KRAUTH

KEPLER

and

P.

VAN

CHoi

and

D.

P. E.,

HART

and

S. I.,

LAPEDES

DAYHOFF

[16]

(1982)

54

(USA) 79 (1982) (1952) 106j

2554.

235.

H., Phys. the

Sitges

J.

Phys.

Rev.

Conference

(1985)

304.

Neural

on

networks

(Sitges,

J.,

MOURIK

M.

Y.,

1990

Seoul

Classification

Pattem

AM

(1991)

National

Univ.

and

Scene

1065.

preprint

(SNUTP 90/33). York Wiley, 1973)

Analysis (New

5.

1986) [13] [14] [15]

and

M.

GALLANT

BRUNAK

[12]

48

1990)

June

Chap. [10] [I Ii

Sci. Soc.

I., Phys. Rev. A 37 (1988) 2739. GRoss SOMPOLINSKY D. J., KANTER I. and CooK J., J. Phys. All (1989) 2057. BOLLt D. and DUPONT P., Proceedings of BOLLt

[9]

Phys.

Mod.

Rev.

Acad.

Phil.

Camb.

KANTER

Spain, [7] [8]

Nail.

Proc.

Proc.

Neural

preprint

and

J.

ZIPSER

T. B.

GUTFREUND

and

1988).

Mass.,

1

S.,

KNUDSEN

D., in

Parallel

Nucl.

Acids

Distributed

Res.

18

Processing

(1990) (MIT

4787,

Press,

Cambridge,

;

Network J., Neural P., The Modeling of

E., J. Phys. W., MtzARD

Net.

;

and

lsl-193

Cambridge

(M.I.T. Press, (1990) 179.

Perceptrons

Trans.

ENGELBRECHT D. E.

pp.

S.,

PAPERT IEEE

(1988)

All M.

Anon

and L.

H., 1990, private

(van

Architectures Neural

Nostrand

(Cambridge

networks

Reinhold, Univ.

New

Press,

257.

NADAL

J.-P.,

F., J. Phys.

Complex Systems 1 (1988) 387 49 (1988) 1657.

France

communication

and

in

preparation.

1990) Chap.

York,

1991)

;

BURNOD

Y.

6.