Search for notes by fellow students, in your own course and all over the country.

Browse our notes for titles which look like what you need, you can preview any of the notes via a sample of the contents. After you're happy these are the notes you're after simply pop them into your shopping cart.

My Basket

You have nothing in your shopping cart yet.

Title: Computational Neuroscience 2011 paper answers (UCL)
Description: Computational Science module for Neuroscience BSc at UCL I got 69 in the module and a first class degree in Neuroscience

Document Preview

Extracts from the notes are below, to see the PDF you'll receive please use the links above


2011
1
...

There is the same winner-takes-all structure but learning rule is
modified so that weights to outputs neighbouring the winner (Ok)
are also modified using a ‘neighbourhood function’ F
...
So F(i,k) decreases
(maximum is 1, minimum is 0) from 1 to 0 as the distance increases
between two neurons
...

Biologically speaking, this could occur by physical spread of chemical
neurotransmitters or messengers
...

So the perceptron learns to classify two groups of input patterns by
finding weights to establish a ‘decision boundary’ between them
...
There is a set of training set of inputs with target outputs
values
...

Present the input pattern and find the corresponding output values
...

(tin –oin) is known as delta, it is the error made by output i for the
input pattern n
...

The delta value could be positive or negative, if it’s positive,
connection weight increases, if it’s negative, connection weight
decreases
...

After training, the network can be used to sort data
...
There are anomalies because the
training data set will have different distribution to the data set being
1

2011
presented
...

Another important about perceptrons is that the delta rule only
learns connection weights
...
So in order to find the threshold value using
the delta rule, the threshold value can be changed by making it into
another connection weight
...
e
...

Perceptrons can have many output units forming a single layer
neural network
...

What restrictions are there on the power of the perceptron as a
classification device, and how can such restrictions be overcome (5
marks)?
Perceptrons can perform linear discrimination but cannot solve
‘non-linearly’ separable problems, e
...
the ‘exclusive OR’ problem
...
In this case, the simple
delta rule has to be changed
...

The neurons can use a continuous transfer function instead of the
threshold logic function
...
If the output value is closer to the target
value then the change is good
...

Discuss the biological plausibility of these learning algorithms (5
marks)
...

Competitive learning- place cells may use this learning algorithm
...
Training sets
may have distributions of data to actual sets of data, the linear
decision boundary generated by perceptrons would be too rigid
...

Error-back propagation- very unlikely to be used in real life
...
Unlike
2

2011
Hebbian learning which is based on one synapse and its presynaptic
and postsynaptic firing rates
...

2
...

The problem is that action signal is needed to be generated to
achieve a desired state from a given current state
...
Then the brain needs to learn how to get to a
desired state from the current state, e
...
like playing a stroke in
tennis, or trying to shoot hoops in basketball
...
All the hidden values are unknown, so an
algorithm (e
...
error-back propagation) is needed to find the values
in order to make accurate movements
...

Forward models is used to train for predictions about the outcome
of an arbitrary signal
...

The very first prediction would be inaccurate, but there is sensory
input (visual and proprioceptive) telling the brain where the arm
actually ended up
...
This means the differences between the observed
and predicted effect is the delta used in error-back propagation to
train the model
...

The inverse model is trained after the forward model becomes fully
trained
...

When there is a desired state to be achieved, the difference
between the desired state and the predicted state is the error used
in error back-propagation for this part of the network
...

Describe how convergent force fields generated from the spinal
cord could aid the control of limb movement (8 marks)
...

The idea of convergence force fields originated with Bizzi et al
...
He stimulated the spinal cord, either electrically or using
glutamate at different levels and different areas of the spinal cord
...
This was expected
...

They were called convergence force fields (CFF)
...
In other words, if the leg was already at the
equilibrium position to start with, stimulation of lateral neuropil
region would produce no movement
...

Micro-stimulation at different levels of spine produces CFFs with
different equilibrium positions
...

This wiring is advantageous to the organism
...
This was what Bizzi et al
...
So if an animal wants their leg to go anywhere, you can move
it there from any position just by giving the right of relative
activation to different CFF generators, which all have equilibrium
positions around the edge of the work space
...
This is the
equilibrium point control hypothesis
...
of CFF generators
...
While the equilibrium point
control hypothesis may be relevant for frogs, it might not be as
applicable to primates
...
In
particular, fine movements of fingers and thumbs don’t seem to be
focused in this way
...


4

2011
Give an example of how the firing rates of many neurons, each
with broadly tuned responses to specific values of a behavioural
variable, can be combined to give an accurate estimate of that
variable (5 marks)
...
A population vector is the
firing rate weighted vector sum of preferred firing directions of
neurons
...

Basically each neuron has a preferred firing direction, and all the
neurons of different firing direction has different firing rates
...
This is a
good way of averaging across all the noisy neurons to get a good
overall estimate of which direction the neurons are encoding
...
g
...

Reaching movement could be coded by population vector
...
These
different neurons with different preferred reaching directions are all
fire at different rates, so just take the firing rate weighted vector
sum of preferred directions to get an estimate of which reaching
direction the neurons are encoding (Georgopoulos et al
...
The
estimate is not exact but accuracy increase with number of neurons
to filter out noise
...
Each place cell has a preferred location of firing,
which could be a vector, so a rat’s location is well predicted by firing
rate weighted vector average of preferred locations
...
Also,
population vector is only actually the best optimal estimator when
the tuning is a cosine function
...

Population vectors would work more efficiently for head direction,
as there is a possible coverage of all directions
...
Make eye movement to retinal
position A, there is a bump of activity in all the neurons around A
centred on A
...

3
...


5

2011
Associative memory is being able to learn and remember
relationships between unrelated items
...
E
...
remember an event by linking the people, objects,
time and space together
...
When a partial cue from the memory is presented, pattern
completion occurs and the whole episode is retrieved
...

Describe how the hippocampus might act as an associative
memory system (6 marks), including how its anatomical structure
relates to this function (5 marks),
CA3 neurons in the hippocampus have recurrent connections to
other CA3 neurons
...
This is a classic network of the hippocampus
...






CA3 neurons receive input from perforant pathway from the
entorhinal cortex, these are weak distal synapses
...

o The inputs are small in number but very powerful- a
single DG synapse can make a cell fire; they are
detonator synapses
The other set of inputs from the perforant path of the
entorhinal cortex could be used for retrieval
...
Then EC present partial cue and
causes retrieval of the pattern
...
1995
A way to separate learning and retrieval by using neuromodulator
...
Ach coming from medial septum
and it changes properties of neurons in CA3 to go between learning
and recall
...
So during learning, there is a large flood of Ach to
suppress recurrent excitatory connections, enhance synaptic
plasticity and boost detonator synapses
...


6

2011

There are problems with this model: slow- neuromodulators will
take a few seconds to diffuse
...
It is known that there is a lot of acetylcholine release from
the medial septum in rats when they are in a novel environment;
Ach modulates synaptic plasticity in the hippocampus; it is know
that it suppresses recurrent connections
...
Perforant path synapses in the dentate gyrus undergo selforganisation to form new representations of input form
entorhinal cortex
2
...
Excitatory recurrent connections in stratum radiatum (in
hippocampus) mediate auto-associative storage and recall
of these patterns
4
...

5
...
The
comparison of recall activity in region CA3 with direct input
to region CA1 regulates cholinergic modulation, allowing a
mismatch between recall and input to increase ACH, and a
match between recall and input to decrease Ach
...

There is a network that an only store a limited number of patterns,
too many patterns means interference occur, spurious memories
7

2011
form
...
In order of the hippocampus to do this, you need
very plastic synapses at a very fast learning rate- rapid ‘one-shot’
associative learning
...
This is the price to pay for fast learning rate
...
The slow system has no
problem with interference, can look at events over time and slowly
extract underlie structures, look at commonalities between events
and make them stronger
...

This idea is based on Marr (1971)
...
People argue
that during sleep, the hippocampus replays the event and
trains the neocortex to remember and associate the
different inputs (smell, sight etc
...

o Is this transfer or training? Do memories truly leave
the hippocampus?
Timescale? One night or twenty years → graded retrograde
amnesia?
Generalisation / abstraction of meaning from multiple
events leads to semantic memory from episodic memory?

Hippocampus is an index (this index represents cross-modal biding)
or convergence zone for disparate neocortical sites which store the
memory content
...
This is anatomical evidence why memory
transfer might be true
...

The standard consolidation model (Squire and Alvarez 1995)


States that all declarative memories are treated in the same
way by the medial temporal lobe, this means semantic

8

2011



memory of facts and episodic memory of events are the
same
...
It binds together neocortical
memory traces
...


Memory consolidation occurs when the hippocampal system
repeatedly reactivates representations in neocortex, leading
eventually to strong interconnections among cortical sites, which
can support memory independently of the hippocampal system
...
So older memories are less susceptible to be
wiped away by hippocampal damage – as they are distributed more
widely so larger lesions needed to affect remote memories
...
e
...

However, true and detailed episodic memories always depend on
the hippocampus
...
Eventually a memory can become
independent of the hippocampus but to remember it vividly, the
hippocampus is needed
...
Discuss how the biophysical properties of a single neuron can
determine what type of function it performs (4 marks)
...

A simple artificial neurons have a threshold logic function, the
network uses a simple integrate and fire model
...
The ‘net input’ to the
neuron is calculated as the ‘weighted sum’ of input activations
...
This is like calculating a dot product
of vectors
...

Time of firing of neurons are especially useful in scale-invariant
recognition
...
g
...
Doesn’t matter if the person speaks softly or
loudly, the identity of the speaker can still be recognised
E
...
scent, either weak or smell, still recognisable
Pattern of activity in input neurons is a vector, x
...
To recognise input pattern of activity x as
similar to stored pattern x*, i
...
x ≈ λx*
...

Standard neural networks have to normalise inputs to do
scale-invariant recognition
o Normalisation could be achieved by feedback
inhibition
o But this still poses a problem as standard neural
networks are strength dependent rather than
pattern dependent
...
This weak input is not
fundamental in deciding what the outcome is
...

Linear pattern separation (perceptrons: weighted sum of
firing rates and threshold) are not suitable for scale invariant
recognition
Tuning to respond at a given absolute value (e
...
head
direction and place cells, radial based functions) are not
suitable for scale-invariant recognition either
Tuning to respond to given relative values are suitable for
scale invariant recognition
...
This model of
computation explains how one scheme of neuroarchitecture can be
used for very different sensory modalities and seemingly different
computations
...
g
...


10

2011
When there is no input current, the sub-threshold cell potential
oscillation (solid line) never exceeds the firing threshold (horizontal
broken line)
...

The pattern is coded by the timing of spikes relative to a subthreshold membrane potential oscillation
...
e
...
If there is a big input
current, then the shift of cell potential oscillation upwards will occur
relatively early in the oscillation wave
...
So small inputs must fire
near the peak of the sub-threshold membrane potential oscillation
for it to be able to bring it to firing threshold
...

Olfactory bulb, hippocampus have this oscillation so this might be
how information is coded in them
...
The input
neurons are coded by the strength of a smell they’ve detected and
their inputs are one spike each, but the timing of that input is crucial
instead
...
So neuron one fires early
in the oscillation wave because it has a large current and brings the
cell’s oscillation above firing threshold
...

So all the neurons fire once per cycle at slightly differently points in
the cycle
...
So all the input neurons’ spikes will arrive simultaneously
11

2011
at the output neuron, and that will make a big impact on the output
cell, especially if the cell does coincidence detection
...
If there is a
different pattern, the conduction delays will not fit and the firing will
not summate
...

Provides alternative to standard models of input values = activation
(firing rates) and processing using variable connection strengths
(more plausible due to existence of LTP and LTD)
...

Adding scale invariant:
If the time advance (Ti) is proportional to the log of the input
current, log(xi), then the network will do scale-invariant recognition,
if the input current strength is doubled: log(2xi)= log(2) + log(xi)
...
By doubling, the output neuron will respond
earlier by log(2)
...

Consider the shape of the oscillatory curve, it is roughly sinusoidal so
it could produce something like a log coding of stimulus intensity
...
So the shape of this membrane oscillation
curve over time means small increases in inputs are actually
magnified in giving a big time advance than big inputs
...

Evaluation of model
The oscillation is 10 Hz, so that is 100 msec between each peak,
therefore time advance of different neurons have to be 10, 20 msec
...
e
...
However in myelinated axons,
transmission delays are much shorter than this, usually 1 or 2 msec
...
In the hippocampus there are weird axons
from dentate gyrus granule cells going to CA3 pyramidal cells which
are not myelinated, so transmission may be slow
...
However even if it is possible to
have different delays, there is no learning rule for setting delays or
means of changing them
...
So a developmental setting might exist so
that important odours are hardwired and automatically detected,
but it is not particularly flexible like the standard model is, where
there is plenty of evidence that you can rapidly change the strength
of connections in synapses
...
Discuss the neural representations that allow a mammal to
know where it is within its environment (8 marks)
...

Place cells fire whenever a rat is in a given location (O’Keefe and
Dostrovsky 1971)
...
They are mainly found in CA1 and CA3 of the hippocampus
proper, but have also been found in the subiculum and entorhinal
cortex
...
However,
in narrow arm mazes, where the animal is running back and forth on
linear tracts, the place cells fire differently depending on the
direction of movement of the rat
...
If all cues are removed, the place cells
might rotate slowly because the mouse doesn’t known the definitive
location anymore
...
Hippocampal pyramidal cells are the final,
output layer of the device
...

Sensory inputs  competitive learning  entorhinal cortex with
functions similar to place cells, but not as accurately tuned 
completive learning  place cells in the hippocampus (CA1 and
CA3)
O’Keefe and Burgess (1996) - Place cells also have cues from
boundary vector cells (BVCs)
...
Sharper tuning of BVCs for shorter distances,
more broad tuning for longer distances
...
Place
fields are modelled as the threshold sum of 2 or more BVC firing
fields
...
Place cell firing in new environmental
13

2011
layouts can be predicted using BVCs
...
2009)
...
The
BVCs model is good at capturing sensory aspects, but it doesn’t take
learning into account
...
Place field stability in a novel environment depends
on NMDA receptors in CA3 (Kentros et al
...
2002)
...
2002) and place cells distinguish locations with
experiences (Barry et al, 2006)
...

Head direction cells: Ranck (1984) first discovered them in the dorsal
presubiculum
...
The primary correlate is the
azimuthal orientation of the head in the horizontal plane
...
(2005)
...
Nearby gird cells have grids of similar
orientation and scale, but they are shifted to tile the environment
...

What is the problem of “temporal credit assignment” in learning (3
marks)?
These problems occur in reinforced learning, a form of slightly
supervised learning
...

The solution is to have an internal critic to assess performance and
guides actions
...

The temporal difference learning is a slow ‘trial and error’ based
method, but it is biologically plausible
...

Describe the neural mechanisms by which a mammal might learn
to find its way to a previously visited reward location, including

14

2011
how the temporal credit assignment problem would be solved (9
marks)
...
g
...

The first learning problem is how to act, i
...
what to do to approach
the goal
...
g
...
So this means
place cells are the input and movement is the output
...
The critic learns that location
at goal is good, which is easy because presence of immediate
reward
...
The critic value (V) of
each state/input (u) is the expected future reward from that state
...

δ(t) = v(t+1)-v(t)+r(t)
if δ(t)>0, action o(t) was good
But how to learn v? A simple learning rule creates a set of
connection weights (w) so that v(t) = w x u(t) and w → w + εS(t)δ(t),
i
...
you can also use δ to learn weights for the critic!
Spatial example:
First encounters goal at location x: r=1, so increase weight to critic
from place cell at x
...

Next time go to x from x’, v(t+1)-v(t)=1, so increase weight to critic
from place cell at x’
...

Then nest time go to x’ from x’’, v(t+1) –v(t)=1, so increase weight to
critic from place cell at x’’, and so on
...

There are sensory cells that project to the hippocampal place cells,
causing a particular output pattern in the place cells
...
There are two sets of nucleus accumbens cells, left or
right, so one set of place cells may project to the left turn groups
15

2011
while the other set of place cells project to the right turn groups
...

Also, head direction cells have modifiable connections to all of the
nucleus accumbens cells- this helps to decide whether the left or
right pool of neurons get more activation to make the animal move
...
Recently active synapses which led to the ‘correct’
locomotor response are strengthened, so that any move which took
place in a particular locational and directional context, and resulted
in reinforcement, is ‘stamped in’ in a Thorndikian (1898, 1911)
manner
...
Each of those connections had a decaying
memory of how often there had been Hebbian pre- and postsynaptic activity
...

So under this scheme synapses which have been chronically or very
recently active are strengthened the most, while synapses with the
least cumulative activity are strengthened least
...
e
...

0term
...


This type of model still requires trial error and it only learns when it
is at the goal
...

Performance of simulations of the model
It learns stereotype trajectories rather than optimal trajectories
...

Unfortunately the model doesn’t take into straight
...

Place constant means the goal is always in the same place, place
random means the goal gets moved
...

Evaluation

16

2011
Captures the idea of ‘response’ (body turn) learning in basal ganglia,
versus ‘place’ presentation in hippocampus
...

The Brown and Sharp model performs complex ‘stimulus-response’
learning- takes into account place and direction and then make a
body turn of either left or right in order to get to the goal
...

Solves the temporal credit assignment problem by using the
recency-weighted cumulative Hebbian value associated with each
connection
...
There may be
something in the synapses that governs LTP which takes into
account decay
...
But then it is impossible for one
synapse to know all the other synaptic activity of the other
synapses
...


17


Title: Computational Neuroscience 2011 paper answers (UCL)
Description: Computational Science module for Neuroscience BSc at UCL I got 69 in the module and a first class degree in Neuroscience