RW001: Difference between revisions
(Created page with "Welcome to the Matrix Analysis class wiki. Instructor: Christopher Hillar Scribe: Sarah Marzen Web co-ordinator: Mayur Mudigonda") |
|||
(11 intermediate revisions by 2 users not shown) | |||
Line 1: | Line 1: | ||
This is the spring '13 Matrix analysis class wiki | |||
== Course description == | |||
Web | This is the first ever course of its kind on Matrix Analysis. Matrices are generally awesome and since most of us use Matlab, it would help to know more about them. | ||
The intention is to provide people some matrix theory and tools that are relevant for machine learning, artificial intelligence, and neuroscience. | |||
Topics will attempt to cover (in no particular order and with no guarantees of them happening): | |||
*basic notions of matrix theory - vector spaces, rank, dimension, determinants, etc. | |||
*eigenvectors and eigenvalues, norms and inequalities, matrix functions | |||
*special matrices - stochastic, orthogonal, unitary, symmetric, positive definite, etc. | |||
*matrix decompositions - SVD, polar, diagonalizability, Jordan form, etc. | |||
*applications to probability & data analysis - Max/Min Entropy, Markov chains, PCA, ICA, etc. | |||
*applications to optimization - Newton and Hessian methods, conjugate gradient, natural gradient etc. | |||
*applications to machine learning - spectral relaxations for clustering, gradients of objective functions in matrices, etc. | |||
*applications to dynamical systems - linear systems theory, associative networks, etc. | |||
*applications to neuroscience - ACS, MPF, spike-timing max entropy neural codes, etc. | |||
*applications to noneq stat mech - Schnakenberg theory, Crooks & Jarzynski theory, etc. | |||
=== Instructors === | |||
[Chris Hillar] | |||
* [http://www.msri.org/people/members/chillar/ Website] | |||
* Office: 573 Evans | |||
* Office hours: immediately following lecture or through e-mail | |||
[Sarah Marzen], Scribe and GSI | |||
* Email: | |||
* Office: 567 Evans | |||
[http://redwood.berkeley.edu/mayur Mayur Mudigonda] Web stuff | |||
=== Lectures === | |||
* NO CLASS April 4, 2013 !!! | |||
* '''Location''': 560 Evans (Redwood Center Conference Hall) | |||
* '''Times''': First and Third Thursdays - 3:30 PM to 5 PM | |||
* Lecture 1 (intro lecture): [http://www.msri.org/people/members/chillar/files/Day1.pdf] | |||
* Lecture 2: [http://www.msri.org/people/members/chillar/files/Day2.pdf] | |||
* Lectures 3,4: [http://www.msri.org/people/members/chillar/files/Day4.pdf] | |||
=== Enrollment information === | |||
Open to all interested members of the Berkeley community. The lectures will not be recorded, so attending the fortnightly talks and e-mails are the only interactions. | |||
=== Email list and forum === | |||
=== Grading === | |||
Self-grading with solutions in class. Since this is the first time this class is being offered, the details are still under work. | |||
=== Required background=== | |||
Prerequisites are calculus, ordinary differential equations, basic probability and statistics, and linear algebra. Familiarity with programming in a high level language such as Matlab is also required. | |||
=== Textbooks === | |||
Matrix Analysis, Horn and Johnson. | |||
[http://www.maths.gla.ac.uk/~ajb/dvi-ps/2w-notes.pdf Basic Linear Algebra Notes] |
Latest revision as of 01:28, 4 April 2013
This is the spring '13 Matrix analysis class wiki
Course description
This is the first ever course of its kind on Matrix Analysis. Matrices are generally awesome and since most of us use Matlab, it would help to know more about them.
The intention is to provide people some matrix theory and tools that are relevant for machine learning, artificial intelligence, and neuroscience.
Topics will attempt to cover (in no particular order and with no guarantees of them happening):
- basic notions of matrix theory - vector spaces, rank, dimension, determinants, etc.
- eigenvectors and eigenvalues, norms and inequalities, matrix functions
- special matrices - stochastic, orthogonal, unitary, symmetric, positive definite, etc.
- matrix decompositions - SVD, polar, diagonalizability, Jordan form, etc.
- applications to probability & data analysis - Max/Min Entropy, Markov chains, PCA, ICA, etc.
- applications to optimization - Newton and Hessian methods, conjugate gradient, natural gradient etc.
- applications to machine learning - spectral relaxations for clustering, gradients of objective functions in matrices, etc.
- applications to dynamical systems - linear systems theory, associative networks, etc.
- applications to neuroscience - ACS, MPF, spike-timing max entropy neural codes, etc.
- applications to noneq stat mech - Schnakenberg theory, Crooks & Jarzynski theory, etc.
Instructors
[Chris Hillar]
- Website
- Office: 573 Evans
- Office hours: immediately following lecture or through e-mail
[Sarah Marzen], Scribe and GSI
- Email:
- Office: 567 Evans
Mayur Mudigonda Web stuff
Lectures
- NO CLASS April 4, 2013 !!!
- Location: 560 Evans (Redwood Center Conference Hall)
- Times: First and Third Thursdays - 3:30 PM to 5 PM
- Lecture 1 (intro lecture): [1]
- Lecture 2: [2]
- Lectures 3,4: [3]
Enrollment information
Open to all interested members of the Berkeley community. The lectures will not be recorded, so attending the fortnightly talks and e-mails are the only interactions.
Email list and forum
Grading
Self-grading with solutions in class. Since this is the first time this class is being offered, the details are still under work.
Required background
Prerequisites are calculus, ordinary differential equations, basic probability and statistics, and linear algebra. Familiarity with programming in a high level language such as Matlab is also required.
Textbooks
Matrix Analysis, Horn and Johnson. Basic Linear Algebra Notes