38
What to make of: • distributed representations • summation of inputs • Hebbian plasticity ? Competitive nets Pattern associators Autoassoci

What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Embed Size (px)

Citation preview

Page 1: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

What to make of:• distributed representations• summation of inputs• Hebbian plasticity ?

Competitive nets Pattern associators Autoassociators

Page 2: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

"Hebb-Marr" associative nets

McNaughton & Morris, 1987

Page 3: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Between 1968 and 1971, David Marr published the first series of theoretical papers that attempted to interpret the anatomical connections and known physiology of specific brain structures (cerebellum, hippocampus, neocortex) in terms of associative memory systems using Hebb's rule. He made numerous specific predictions about the physiological properties of these nets that have since been verified.

Page 4: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 5: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 6: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 7: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 8: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Competitive Networks

• “discover” structures in input space• may remove redundancy• may orthogonalize• may categorize• may sparsify representations• can separate out even linear combinations• can easily be generalized to self- organizing topographic maps

Page 9: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 10: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Competitive Networks

• “discover” structures in input space• may remove redundancy• may orthogonalize• may categorize• may sparsify representations• can separate out even linear combinations• can easily be generalized to self- organizing topographic maps

Page 11: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Pattern associators

• generalize to nearby input vectors• gracefully degrade/fault tolerant• may extract prototypes• may reduce fluctuations• are fast (feed forward)• p (the # of associations) is proportional to C (# inputs/unit) and also grows with decreasing a (sparser patterns)

Y

X

r (X) r (Y) after learning

Page 12: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

1 1 0 1 0 0

Page 13: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

1 1 0 1 0 0

Page 14: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

0 0 1 0 1 1

Page 15: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

1 0 0 1 1 0

Page 16: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

3 2 2 3 3 2 3

1 0 0 1 1 0

3

Associative retrieval: The inhibitory interneuron divides down the excitation in proportion to the total input to the net, so that only the most strongly excited cells reach threshold (i.e., integer division).

Page 17: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

3 2 2 3 3 2 3

1 0 0 1 1 0

3

Page 18: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

1 1 2 1 2 2 2

0 0 1 0 1 1

2Pattern completion: The inhibitory interneuron divides down the excitation in proportion to the total input to the net, so that only the most strongly excited cells reach threshold (i.e., integer division).

Page 19: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

0 0 1 0 1 1

Page 20: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Pattern associators

• generalize to nearby input vectors• gracefully degrade/fault tolerant• may extract prototypes• may reduce fluctuations• are fast (feed forward)• p (the # of associations) is proportional to C (# inputs/unit) and also grows with decreasing a (sparser patterns)

Y

X

r (X) r (Y) after learning

Page 21: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Storage is formally equivalent to forming the so-called outer product of two binary vectors. Multiple memories are stored by combining outer product matrices using the logical OR operation. Retrieval is formally equivalent to taking the inner (or dot) product of the input vector and the storage matrix (i.e., multiply through the rows and sum the columns) and subjecting it to a non-linear normalization (i.e., integer division).

Page 22: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

•Inhibitory cells must be driven by the same excitatory afferents that activate the principal cells .

•The inhibitory mechanism must implement a division operation on the excitation that reaches the cell body from the dendrites.

•Inhibitory cells can be much fewer in number than principal cells, but they must have extensive connectivity.

•Inhibitory cells must respond to a synchronous input at lower threshold and at shorter latency than principal cells. Shorter latency is necessary to ensure that the appropriate division operation is already set up at the somata of the principal cells by the time the excitation from the same input arrives there via the principal cell dendrites.

•Whereas principal cells will be quite selective in their response characteristics, inhibitory neurons will not be particular about which afferents are active at a given time, only about how many are active. Thus they will convey little information in the principal cells' response domain.

•Excitatory synapses onto interneurons should not be plastic

•In unfamiliar situations (i.e. when current input elicits reduced output) extrinsic modulation of inhibitory neurons might lower output threshold, successively probing for a complete pattern. This might also serve as a gate enabling the activation of the synaptic modification process.

Properties of interneurons predicted by the simple Hebb-Marr net model

Page 23: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Freund & Buzsaki, 1996

Axons of basket inhbitory interneurons project widely.1-2 mm in transverse plane, 2-4 mm in long axis

Page 24: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

CA3 is dominated by recurrent collaterals

Page 25: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Autoassociation in a recurrent net

Page 26: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Autoassociation in a recurrent net

The input pattern is kept on for 2 time steps so that the output at time t-1 is associated with the input at time t

Page 27: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Recurrent net after storage of 3 input vectors

Page 28: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Error correction (pattern completion)

Page 29: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

"Reverberation": persistence of pattern after removal of input

Persistencerequires synaptic weights to be symmetrical: Wij = Wji ,although for large nets symmetry does not have to be perfect.

Page 30: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

If inputs change each cycle, the output at time t-1 is associated with the

input at time t.

Sequence Learning in Recurrent Nets

Page 31: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

This leads to sequence storage.

Page 32: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

The sequence may be of arbitrary length.

Page 33: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Presentation of any vector in the sequence leads to sequence completion if and only if all elements of the sequence are unique.

Page 34: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 35: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators
Page 36: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Autoassociators

• complete a partial cue• can continue a sequence• generalize• gracefully degrade/fault tolerant• extract prot. / reduce fluct.• are also fast, but their feedback may sustain short term memory• pc is again proportional to C and grows with decreasing a pc = k C / [a log (1/a)]

r (Y’) r (Y) after learning

Y

Page 37: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Constraints on Storage Capacity

1) Connectivity. In real networks connectivity is much less than all-to-all. Capacity is proportional to the number of synapses/unit, C.

2) Sparsity of coding, a. Sparsity roughly speaking refers to the fraction of the units active at any one time. The continuum ranges from extremely sparse (a=1/N) to fully distributed (a=1).Sparse activity allows more patterns, but not more information.

3) "Orthogonality" of coding. By definition, orthogonal vectors have an overlap of zero. In practice, the maximum number of stored patterns is achieved when the patterns are uncorrelated, not mutually orthogonal. Correlated patterns cause local saturation.

Page 38: What to make of: distributed representations summation of inputs Hebbian plasticity ? Competitive nets Pattern associators Autoassociators

Sparser inputs to the associative memory system can be generated by more efficient coding of the input (redundancy reduction). Much of the processing of visual and other sensory information in 'primary' and 'secondary' sensory cortex is devoted to this process of "feature extraction", with various simple feature detectors developmentally programmed. Association cortex learns to develop high order feature detectors based on past experience.