VS265: Class project Fall2010: Difference between revisions

From RedwoodCenter
Jump to navigationJump to search
No edit summary
 
(17 intermediate revisions by one other user not shown)
Line 1: Line 1:
'''Project presentations''' will take place on project presentation day, TBD (typically just before or during finals week).  You may present your project either as an oral (15 min.) or poster presentation.   
'''Project presentations''' will take place on project presentation day, '''Monday, December 13th -- 3-6pm'''.  You may present your project either as an oral (15 min.) or poster presentation.   


Think of the class project as an extended lab assignment.  In most of the labs we just scratched the surface of various network models.  This is a chance to explore these ideas in more depth.  In addition to giving an oral or poster presentation on your project at the end of the term, you should turn in a short writeup that describes the problem you investigated, why it is interesting, your approach, results, and conclusions.  There is no length requirement but about 5-10 pages would be in the right ballpark.   
Think of the class project as an extended lab assignment.  In most of the labs we just scratched the surface of various network models.  This is a chance to explore these ideas in more depth.  In addition to giving an oral or poster presentation on your project at the end of the term, you should turn in a short writeup that describes the problem you investigated, why it is interesting, your approach, results, and conclusions.  There is no length requirement but about 5-10 pages would be in the right ballpark.   
Line 7: Line 7:
This is not an exhaustive listing, just some suggestions to get you started thinking.
This is not an exhaustive listing, just some suggestions to get you started thinking.


* '''NETtalk.'''  Train a multi-layer perceptron to convert text to speech.  You can get Sejnowski & Rosenberg's original paper and the data they used [http://www.cnl.salk.edu/ParallelNetsPronounce/index.php here].  (You will need a DECtalk speech synthesizer to play the phonemes - you can probably pick up a used one online.)
* '''NETtalk.'''  Train a multi-layer perceptron to convert text to speech.  You can get Sejnowski & Rosenberg's original paper and the data they used [http://cnl.salk.edu/Research/ParallelNetsPronounce/ here].  (You will need a DECtalk speech synthesizer to play the phonemes - you may be able to pick up a used one online.)


* '''Recognition of handwritten digits.'''  Train a MLP to classify handwritten digits 0-9.  You can get some training data [http://yann.lecun.com/exdb/mnist/ here].  You may wish to follow the convolutional network methodology of [http://yann.lecun.com/exdb/lenet/index.html Yann LeCun] (try the simpler, [http://yann.lecun.com/exdb/publis/pdf/lecun-89e.pdf earlier model]), or invent your own method.   
* '''Recognition of handwritten digits.'''  Train a MLP to classify handwritten digits 0-9.  You can get some training data [http://yann.lecun.com/exdb/mnist/ here].  You may wish to follow the convolutional network methodology of [http://yann.lecun.com/exdb/lenet/index.html Yann LeCun] (try the simpler, [http://yann.lecun.com/exdb/publis/pdf/lecun-89e.pdf earlier model]), or invent your own method.   


* '''Sparse coding and decorrelation.'''  Implement [http://redwood.berkeley.edu/vs265/foldiak90.pdf Peter Foldiak's network] and train it on the handwritten digits above to learn the features of this data.  You may wish to then try supervised learning on the learned features to see if it has simplified the classification problem.
* '''Sparse coding and decorrelation.'''  Implement [http://redwood.berkeley.edu/vs265/foldiak90.pdf Peter Foldiak's network] and train it on the handwritten digits above to learn the features of this data. How do the learned features change as a function of the number of units in the network? You may wish to then try supervised learning on the learned features to see if it has simplified the classification problem.


* '''Cortical maps.'''  The elastic net model of [http://redwood.berkeley.edu/vs265/durbin-mitchison.pdf Durbin and Mitchison] is typical of many cortical map models in that they learn directly on a parameterized feature space.  But the cortex simply gets a bunch of inputs from the LGN, and so it needs to learn features such as orientation at the same time as it organizes them into a feature map.  How would you go about learning a feature map for orientation and position directly from simulated LGN inputs?  (You may wish to consult the book of [http://nn.cs.utexas.edu/computationalmaps/ Risto Miikkulainen] for recent efforts in this area.)
* '''Cortical maps.'''  The elastic net model of [http://redwood.berkeley.edu/vs265/durbin-mitchison.pdf Durbin and Mitchison] is typical of many cortical map models in that they learn directly on a parameterized feature space.  But the cortex simply gets a bunch of inputs from the LGN, and so it needs to learn features such as orientation at the same time as it organizes them into a feature map.  How would you go about learning a feature map for orientation and position directly from simulated LGN inputs?  (You may wish to consult the book of [http://nn.cs.utexas.edu/computationalmaps/ Risto Miikkulainen] for recent efforts in this area.)
Line 17: Line 17:
* '''The 'magic TV''''  Let's say you wake up one day to find someone rewired your optic nerve (or you have been implanted with a prosthetic retina).  The signals from retina to brain are still intact, but the wires are all mixed up in the wrong place.  Since neighboring pixels in natural images are correlated, it should be possible to learn a remapping that "descrambles" the image by exploiting these correlations.  See if you can train a Kohonen-style network to learn the proper topographic mapping of an image based on the statistics of natural images.  (Kohonen dubbed this problem 'the Magic TV'.)
* '''The 'magic TV''''  Let's say you wake up one day to find someone rewired your optic nerve (or you have been implanted with a prosthetic retina).  The signals from retina to brain are still intact, but the wires are all mixed up in the wrong place.  Since neighboring pixels in natural images are correlated, it should be possible to learn a remapping that "descrambles" the image by exploiting these correlations.  See if you can train a Kohonen-style network to learn the proper topographic mapping of an image based on the statistics of natural images.  (Kohonen dubbed this problem 'the Magic TV'.)


* '''Feedforward vs. recurrent weights.'''  As we discussed in class, one can implement a given input-output mapping in a neural network using just feedforward weights: <math> y = W x</math>, or using just recurrent weights: <math> \tau dy/dt + y = x + M y</math>, or both: <math>\tau dy/dt + y = W x + M y</math>.  Probably there is a trade-off here in terms of minimizing overall wiring length and settling time - i.e., feedforward networks are fast but require lots of synapses, while recurrent networks are slower but can implement more complex functions with local connections.  Explore these tradeoffs for a particular problem - e.g., implementing an array of Gabor filters in model of V1.
* '''Feedforward vs. recurrent weights.'''  As we discussed in class, one can implement a given input-output mapping in a neural network using just feedforward weights: <math> y = W x</math>, or using just recurrent weights: <math> \tau dy/dt + y = x + M y</math>, or both: <math>\tau dy/dt + y = W x + M y</math>.  Probably there is a trade-off here in terms of minimizing overall wiring length and settling time - i.e., feedforward networks are fast but require lots of synapses, while recurrent networks are slower but can implement more complex functions with local connections.  Explore these tradeoffs for a particular problem - e.g., implementing an array of Gabor filters in a model of V1.


* '''Sparse codes and associative memory.'''  The advantage of storing and recalling patterns using an associative memory as opposed to a conventional computer memory is 1) parallel search, 2) distributed storage, and 3)  denoising (recall of an uncorrupted pattern from partial or degraded input).  However, associative memory models do not work well with natural data such as images or sound directly.  Rather, they are best suited (have highest capacity) for sparse patterns (i.e., patterns with many zeros).  Recent work (discussed in class) has shown how it is possible to convert natural images and sounds into a sparse format, and there is some evidence for this happening in the brain.  See if you can link these ideas in order to store natural images or sounds in an associative memory.  
* '''Sparse codes and associative memory.'''  The advantages of storing and recalling patterns using an associative memory as opposed to conventional template matching are 1) parallel search, 2) distributed storage, and 3)  denoising (recall of an uncorrupted pattern from partial or degraded input).  However, associative memory models do not work well with natural data such as images or sound directly.  Rather, they are best suited (have highest capacity) for sparse patterns (i.e., patterns with many zeros).  Recent work (discussed in class) has shown how it is possible to convert natural images and sounds into a sparse format, and there is some evidence for this happening in the brain.  See if you can link these ideas in order to store natural images or sounds in an associative memory.  


* '''Hardware implementation of associative memory.'''  The analog Hopfield model has a direct physical implementation as an electrical circuit of resistors, capacitors, and op-amps.  Try building a scaled-down version of this model in hardware.  What issues arise in the implementation of this model?  How long does it take to converge to a local minimum?
* '''Hardware implementation of associative memory.'''  The analog Hopfield model has a direct physical implementation as an electrical circuit of resistors, capacitors, and op-amps.  Try building a scaled-down version of this model in hardware.  What issues arise in the implementation of this model?  How long does it take to converge to a local minimum?


* '''Analysis of Hopfield dynamics.'''  The Hopfield dynamics may be seen as performing a form of gradient descent on the energy function - i.e., the state of each unit, <math>V_i</math>, follows a monotonically increasing function of the gradient rather than the gradient itself.  Is the resulting trajectory more efficient for reaching the energy minimum than what you would get from doing steepest descent?
* '''Analysis of Hopfield dynamics.'''  The [http://redwood.berkeley.edu/vs265/hopfield84.pdf Hopfield dynamics] may be seen as performing a form of gradient descent on the energy function - i.e., the state of each unit, <math>V_i</math>, follows a monotonically increasing function of the gradient rather than the gradient itself.  The same is true of the [http://redwood.berkeley.edu/vs265/rozell-sparse-coding-nc08.pdf LCA] (locally competitive algorithm) used for sparse coding.  Is the resulting trajectory more efficient for reaching the energy minimum than what you would get from doing steepest descent?


* '''Bump circuits.'''  Implement Kechen Zhang's bump circuit model discussed in class.  How robust is the model to perturbations of the weights?  How might such a circuit to self-correct for any imperfections in the weights?
* '''Bump circuits.'''  Implement Kechen Zhang's [http://redwood.berkeley.edu/vs265/zhang96.pdf bump circuit model] discussed in class.  How robust is the model to perturbations of the weights?  How might such a circuit to self-correct for any imperfections in the weights?


* '''Map-seeking circuits.'''  David Arathorn has described a neural circuit for doing invariant object recognition which utilizes three-way interactions among units - see [http://www.amazon.com/Map-Seeking-Circuits-Visual-Cognition-Computational/dp/0804742774/ref=sr_1_3?ie=UTF8&s=books&qid=1225087289&sr=1-3 "Map-seeking circuits in Visual Cognition,"] Stanford University Press, 2002.  However, his implementation of the associative memory uses grandmother cellsTry using instead a distributed representation for the memory using a Hopfield network.
* '''Map-seeking circuits.'''  David Arathorn has described a neural circuit for doing invariant object recognition which utilizes three-way interactions among units - see [http://www.amazon.com/Map-Seeking-Circuits-Visual-Cognition-Computational/dp/0804742774/ref=sr_1_3?ie=UTF8&s=books&qid=1225087289&sr=1-3 "Map-seeking circuits in Visual Cognition,"] Stanford University Press, 2002.  Try implementing this model.  One issue to explore is the memory matching scheme, which basically uses template matchingCan this be improved using a distributed representation such as in a Hopfield network?


* '''Entropy of natural images.'''  We live in a highly structured environment, and thus the images that fall upon our retinae exhibit  strong statistical dependencies among local pixel values.  To get a feel for this, consider how long would it take to find an image resembling a natural scene if you viewed a series of images composed from random pixel values.  This tells us that the true entropy of an image patch is far less than N x (entropy/pixel), where N is the number of pixels.  But calculating the true entropy is impossible in practice because it requires collecting the full joint pdf over an image patch, and for any reasonably sized patch this is intractable.  Thus it is currently unknown.  But there are other ways to estimate the entropy of data beyond direct caculation from the pdf.  See for example the recent work of [mailto:djf3@cornell.edu David Field] and colleagues, or methods based on the Hausdorff dimension.  Implement one of these methods, or see if you can come up with your own method, for estimating the true entropy of images.
* '''Entropy of natural images.'''  We live in a highly structured environment, and thus the images that fall upon our retinae exhibit  strong statistical dependencies among local pixel values.  To get a feel for this, consider how long would it take to find an image resembling a natural scene if you viewed a series of images composed from random pixel values.  This tells us that the true entropy of an image patch is far less than N x (entropy/pixel), where N is the number of pixels.  But calculating the true entropy is impossible in practice because it requires collecting the full joint pdf over an image patch, and for any reasonably sized patch this is intractable.  Thus it is currently unknown.  But there are other ways to estimate the entropy of data beyond direct caculation from the pdf.  See for example the recent work of [http://redwood.berkeley.edu/vs265/chandler_field_2007.pdf David Field] and colleagues, or methods based on the Hausdorff dimension.  Implement one of these methods, or see if you can come up with your own method, for estimating the true entropy of images.


* '''ICA applied to EEG.'''  ICA has been successfully applied to EEG signals, to separate noise artifacts from brain signals, and also to find separate signal sources within the brain.  See if you can get some EEG data from one of the labs on campus (e.g., Klein or Knight labs) and use ICA to reveal the hidden structure in this data.  To get started, look at Makeig S, Bell AJ, Jung T-P, Sejnowski TJ (1996)  "Independent Component Analysis of Electroencephalographic Data."  In: Advances in Neural Information Processing Systems 8, 145-151.
* '''ICA applied to EEG.'''  ICA has been successfully applied to EEG signals, to separate noise artifacts from brain signals, and also to find separate signal sources within the brain.  See if you can get some EEG data from one of the labs on campus (e.g., Klein or Knight labs) and use ICA to reveal the hidden structure in this data.  To get started, look at [http://sccn.ucsd.edu/~jung/artifact.html this page].


* '''Hierarchical restricted Boltzmann machines.'''  [http://redwood.berkeley.edu/vs265/hinton06.pdf Geoff Hinton] has recently developed a hierarchical auto-encoder network, based on the restricted Boltzmann machine, for modelling structure in data.  Implement this network and train on the handwritten digits data or some other dataset of your choosing.
* '''Hierarchical restricted Boltzmann machines.'''  [http://redwood.berkeley.edu/vs265/hinton06.pdf Geoff Hinton] has developed a hierarchical auto-encoder network, based on the restricted Boltzmann machine, for modelling structure in data.  Implement this network and train on the handwritten digits data or some other dataset of your choosing.


* '''Modeling dynamical systems with recurrent neural networks.'''  As David Zipser showed in class, it is possible to train recurrent neural networks to learn limit cycles, chaos, and other forms of dynamics.  Try using these learning rules to train a recurrent network to learn to recognize or emulate the structure of time-series data.  (David will make his software available for this purpose to save you some time.)  One form of dynamics you might try to model is [http://psych-s1.psy.vanderbilt.edu/faculty/blaker/BM/BioMot.html "biological motion."]
* '''Modeling dynamical systems with recurrent neural networks.'''  As David Zipser showed in class, it is possible to train recurrent neural networks to learn limit cycles, chaos, and other forms of dynamics.  Try using these learning rules to train a recurrent network to learn to recognize or emulate the structure of time-series data.  (David will make his software available for this purpose to save you some time.)  One form of dynamics you might try to model is [http://psych-s1.psy.vanderbilt.edu/faculty/blaker/BM/BioMot.html "biological motion."]
Line 39: Line 39:
* '''Echo-state networks.'''    [http://redwood.berkeley.edu/vs265/echo_state.pdf Herbert Jaeger] and others have shown that it is possible to utilize echo-state networks to learn rich forms of dynamics without having to train specific connections within the network itself, but rather by learning to "read out" the network in the right way.  Try experimenting with this approach to see what sorts of dynamics it is capable of learning.   
* '''Echo-state networks.'''    [http://redwood.berkeley.edu/vs265/echo_state.pdf Herbert Jaeger] and others have shown that it is possible to utilize echo-state networks to learn rich forms of dynamics without having to train specific connections within the network itself, but rather by learning to "read out" the network in the right way.  Try experimenting with this approach to see what sorts of dynamics it is capable of learning.   


* '''Integrate-and-fire model neurons.'''  The [http://redwood.berkeley.edu/vs265/int-fire/int-fire.html integrate and fire model] is a simple model for capturing the temporal integration and spiking aspects of real neurons.  Using such a simplified model it is possible to begin exploring some interesting questions about sensory coding in neurons.  For example, how is it possible to encode a continous, time-varying signal using a population of spiking neurons?  (See "Spikes" by Rieke et al., or "Principles of Neural Engineering" by Eliasmith and Anderson for an extended discussion of this issue.)  You may also wish to explore the effect of adding more realistic biophysical properties, such as the dependence of threshold on membrane potential (see work by [http://connes.berkeley.edu/vs265/azouz-gray00.pdf Gray and Azouz]).
* '''Integrate-and-fire model neurons.'''  The [http://redwood.berkeley.edu/vs265/int-fire/int-fire.html integrate and fire model] is a simple model for capturing the temporal integration and spiking aspects of real neurons.  Using such a simplified model it is possible to begin exploring some interesting questions about sensory coding in neurons.  For example, how is it possible to encode a continous, time-varying signal using a population of spiking neurons?  (See "Spikes" by Rieke et al., or "Principles of Neural Engineering" by Eliasmith and Anderson for an extended discussion of this issue.)  You may also wish to explore the effect of adding more realistic biophysical properties, such as the dependence of threshold on membrane potential (see work by [http://redwood.berkeley.edu/vs265/azouz-gray00.pdf Gray and Azouz]).


* '''Oscillations.'''  Oscillations in neural activity are pervasive throughout the brain.  What kinds of neural circuits are capable of eliciting oscillating behavior in spiking neurons?  How could it be coordinated across large regions of cortex?  (ask Fritz Sommer)  What role might it play in the processing of information?  [http://redwood.berkeley.edu/vs265/hopfield95.pdf John Hopfield] has suggested that spike timing relative to the phase of an ongoing oscillation could code information.  What factors would need to be considered in order to make this idea viable?
* '''Oscillations.'''  Oscillations in neural activity are pervasive throughout the brain.  What kinds of neural circuits are capable of eliciting oscillating behavior in spiking neurons?  How could it be coordinated across large regions of cortex?  (ask Fritz Sommer)  What role might it play in the processing of information?  [http://redwood.berkeley.edu/vs265/hopfield95.pdf John Hopfield] has suggested that spike timing relative to the phase of an ongoing oscillation could code information.  What factors would need to be considered in order to make this idea viable?
* '''Memristors.'''  The memristor was originally proposed by Leon Chua at UC Berkeley back in the 1970's as a hypothetical '4th circuit element.'  Recently, a group at [http://redwood.berkeley.edu/vs265/memristor-nature08.pdf HP labs] has discovered a memristive like element (see also [http://spectrum.ieee.org/semiconductors/processors/how-we-found-the-missing-memristor this article]).  What is intriguing about the memristor is its synapse like properties ([http://redwood.berkeley.edu/vs265/memristor-stdp.pdf ref]), and according to Chua it even leads to a more straightforward model of action potential dynamics!  Might memristive elements exist in neural circuits in the brain?  Try simulating a memristive device and explore its relevance to plasticity or dynamics in neural systems.
* '''Learning in sensorimotor loops.'''  A goal of research in both robotics and neuroscience is to understand the principles of adaptive behavior in embodied systems.  However, currently there are few theories for guiding work in this area.    One approach advanced by O'regan and colleagues is based on learning "sensorimotor contingencies" ([http://redwood.berkeley.edu/vs265/noe-oregan.pdf ref1], [http://redwood.berkeley.edu/vs265/philipona-oregan.pdf ref2]).  Another by Ralf Der and colleagues is based on minimizing both predictive and 'postdictive' error ([http://redwood.berkeley.edu/vs265/taming_the_beast_martius2010.pdf ref]).  Both propose simple algorithmic examples that you can implement in computer simulation, or you may wish to implement in a real physical system using the robot lab (currently in progress) in the Redwood Center.




Line 50: Line 54:
* [http://www.cs.toronto.edu/~roweis/data.html '''Sam Roweis'''] - Many datasets in Matlab format: MNIST and USPS handwritten digits, faces, text,  speech.
* [http://www.cs.toronto.edu/~roweis/data.html '''Sam Roweis'''] - Many datasets in Matlab format: MNIST and USPS handwritten digits, faces, text,  speech.


* [http://hlab.phys.rug.nl/archive.html '''Hans van Hateren'''] - Natural stimuli collection: still images, intensity time series, video
* [http://hlab.phys.rug.nl/archive.html '''Hans van Hateren'''] - Natural stimuli collection: still images, intensity time series, video.
''Note:'' Hans van Hateren's website is down, there are two mirrors for the still images collection: [http://www.kyb.mpg.de/bethge/vanhateren/index.php mirror 2](Germany, detailed description), [http://pirsquared.org/research/#van-hateren-database mirror 2] (US,CA - faster)


* [http://www.robjhyndman.com/TSDL/ '''Time Series Data Library''']
* [http://www.robjhyndman.com/TSDL/ '''Time Series Data Library''']


* [http://www.face-rec.org/databases/ '''Face Recognition Databases''']
* [http://www.face-rec.org/databases/ '''Face Recognition Databases''']

Latest revision as of 02:45, 28 August 2012

Project presentations will take place on project presentation day, Monday, December 13th -- 3-6pm. You may present your project either as an oral (15 min.) or poster presentation.

Think of the class project as an extended lab assignment. In most of the labs we just scratched the surface of various network models. This is a chance to explore these ideas in more depth. In addition to giving an oral or poster presentation on your project at the end of the term, you should turn in a short writeup that describes the problem you investigated, why it is interesting, your approach, results, and conclusions. There is no length requirement but about 5-10 pages would be in the right ballpark.

Project Suggestions

This is not an exhaustive listing, just some suggestions to get you started thinking.

  • NETtalk. Train a multi-layer perceptron to convert text to speech. You can get Sejnowski & Rosenberg's original paper and the data they used here. (You will need a DECtalk speech synthesizer to play the phonemes - you may be able to pick up a used one online.)
  • Recognition of handwritten digits. Train a MLP to classify handwritten digits 0-9. You can get some training data here. You may wish to follow the convolutional network methodology of Yann LeCun (try the simpler, earlier model), or invent your own method.
  • Sparse coding and decorrelation. Implement Peter Foldiak's network and train it on the handwritten digits above to learn the features of this data. How do the learned features change as a function of the number of units in the network? You may wish to then try supervised learning on the learned features to see if it has simplified the classification problem.
  • Cortical maps. The elastic net model of Durbin and Mitchison is typical of many cortical map models in that they learn directly on a parameterized feature space. But the cortex simply gets a bunch of inputs from the LGN, and so it needs to learn features such as orientation at the same time as it organizes them into a feature map. How would you go about learning a feature map for orientation and position directly from simulated LGN inputs? (You may wish to consult the book of Risto Miikkulainen for recent efforts in this area.)
  • The 'magic TV' Let's say you wake up one day to find someone rewired your optic nerve (or you have been implanted with a prosthetic retina). The signals from retina to brain are still intact, but the wires are all mixed up in the wrong place. Since neighboring pixels in natural images are correlated, it should be possible to learn a remapping that "descrambles" the image by exploiting these correlations. See if you can train a Kohonen-style network to learn the proper topographic mapping of an image based on the statistics of natural images. (Kohonen dubbed this problem 'the Magic TV'.)
  • Feedforward vs. recurrent weights. As we discussed in class, one can implement a given input-output mapping in a neural network using just feedforward weights: <math> y = W x</math>, or using just recurrent weights: <math> \tau dy/dt + y = x + M y</math>, or both: <math>\tau dy/dt + y = W x + M y</math>. Probably there is a trade-off here in terms of minimizing overall wiring length and settling time - i.e., feedforward networks are fast but require lots of synapses, while recurrent networks are slower but can implement more complex functions with local connections. Explore these tradeoffs for a particular problem - e.g., implementing an array of Gabor filters in a model of V1.
  • Sparse codes and associative memory. The advantages of storing and recalling patterns using an associative memory as opposed to conventional template matching are 1) parallel search, 2) distributed storage, and 3) denoising (recall of an uncorrupted pattern from partial or degraded input). However, associative memory models do not work well with natural data such as images or sound directly. Rather, they are best suited (have highest capacity) for sparse patterns (i.e., patterns with many zeros). Recent work (discussed in class) has shown how it is possible to convert natural images and sounds into a sparse format, and there is some evidence for this happening in the brain. See if you can link these ideas in order to store natural images or sounds in an associative memory.
  • Hardware implementation of associative memory. The analog Hopfield model has a direct physical implementation as an electrical circuit of resistors, capacitors, and op-amps. Try building a scaled-down version of this model in hardware. What issues arise in the implementation of this model? How long does it take to converge to a local minimum?
  • Analysis of Hopfield dynamics. The Hopfield dynamics may be seen as performing a form of gradient descent on the energy function - i.e., the state of each unit, <math>V_i</math>, follows a monotonically increasing function of the gradient rather than the gradient itself. The same is true of the LCA (locally competitive algorithm) used for sparse coding. Is the resulting trajectory more efficient for reaching the energy minimum than what you would get from doing steepest descent?
  • Bump circuits. Implement Kechen Zhang's bump circuit model discussed in class. How robust is the model to perturbations of the weights? How might such a circuit to self-correct for any imperfections in the weights?
  • Map-seeking circuits. David Arathorn has described a neural circuit for doing invariant object recognition which utilizes three-way interactions among units - see "Map-seeking circuits in Visual Cognition," Stanford University Press, 2002. Try implementing this model. One issue to explore is the memory matching scheme, which basically uses template matching. Can this be improved using a distributed representation such as in a Hopfield network?
  • Entropy of natural images. We live in a highly structured environment, and thus the images that fall upon our retinae exhibit strong statistical dependencies among local pixel values. To get a feel for this, consider how long would it take to find an image resembling a natural scene if you viewed a series of images composed from random pixel values. This tells us that the true entropy of an image patch is far less than N x (entropy/pixel), where N is the number of pixels. But calculating the true entropy is impossible in practice because it requires collecting the full joint pdf over an image patch, and for any reasonably sized patch this is intractable. Thus it is currently unknown. But there are other ways to estimate the entropy of data beyond direct caculation from the pdf. See for example the recent work of David Field and colleagues, or methods based on the Hausdorff dimension. Implement one of these methods, or see if you can come up with your own method, for estimating the true entropy of images.
  • ICA applied to EEG. ICA has been successfully applied to EEG signals, to separate noise artifacts from brain signals, and also to find separate signal sources within the brain. See if you can get some EEG data from one of the labs on campus (e.g., Klein or Knight labs) and use ICA to reveal the hidden structure in this data. To get started, look at this page.
  • Hierarchical restricted Boltzmann machines. Geoff Hinton has developed a hierarchical auto-encoder network, based on the restricted Boltzmann machine, for modelling structure in data. Implement this network and train on the handwritten digits data or some other dataset of your choosing.
  • Modeling dynamical systems with recurrent neural networks. As David Zipser showed in class, it is possible to train recurrent neural networks to learn limit cycles, chaos, and other forms of dynamics. Try using these learning rules to train a recurrent network to learn to recognize or emulate the structure of time-series data. (David will make his software available for this purpose to save you some time.) One form of dynamics you might try to model is "biological motion."
  • Echo-state networks. Herbert Jaeger and others have shown that it is possible to utilize echo-state networks to learn rich forms of dynamics without having to train specific connections within the network itself, but rather by learning to "read out" the network in the right way. Try experimenting with this approach to see what sorts of dynamics it is capable of learning.
  • Integrate-and-fire model neurons. The integrate and fire model is a simple model for capturing the temporal integration and spiking aspects of real neurons. Using such a simplified model it is possible to begin exploring some interesting questions about sensory coding in neurons. For example, how is it possible to encode a continous, time-varying signal using a population of spiking neurons? (See "Spikes" by Rieke et al., or "Principles of Neural Engineering" by Eliasmith and Anderson for an extended discussion of this issue.) You may also wish to explore the effect of adding more realistic biophysical properties, such as the dependence of threshold on membrane potential (see work by Gray and Azouz).
  • Oscillations. Oscillations in neural activity are pervasive throughout the brain. What kinds of neural circuits are capable of eliciting oscillating behavior in spiking neurons? How could it be coordinated across large regions of cortex? (ask Fritz Sommer) What role might it play in the processing of information? John Hopfield has suggested that spike timing relative to the phase of an ongoing oscillation could code information. What factors would need to be considered in order to make this idea viable?
  • Memristors. The memristor was originally proposed by Leon Chua at UC Berkeley back in the 1970's as a hypothetical '4th circuit element.' Recently, a group at HP labs has discovered a memristive like element (see also this article). What is intriguing about the memristor is its synapse like properties (ref), and according to Chua it even leads to a more straightforward model of action potential dynamics! Might memristive elements exist in neural circuits in the brain? Try simulating a memristive device and explore its relevance to plasticity or dynamics in neural systems.
  • Learning in sensorimotor loops. A goal of research in both robotics and neuroscience is to understand the principles of adaptive behavior in embodied systems. However, currently there are few theories for guiding work in this area. One approach advanced by O'regan and colleagues is based on learning "sensorimotor contingencies" (ref1, ref2). Another by Ralf Der and colleagues is based on minimizing both predictive and 'postdictive' error (ref). Both propose simple algorithmic examples that you can implement in computer simulation, or you may wish to implement in a real physical system using the robot lab (currently in progress) in the Redwood Center.


Data

There are many sources of data on the web that you can use for these projects. Here are a few to get you started:

  • Sam Roweis - Many datasets in Matlab format: MNIST and USPS handwritten digits, faces, text, speech.
  • Hans van Hateren - Natural stimuli collection: still images, intensity time series, video.

Note: Hans van Hateren's website is down, there are two mirrors for the still images collection: mirror 2(Germany, detailed description), mirror 2 (US,CA - faster)