Machine learning and neural networks for neuroscience

              

               Machine learning and neural networks for neuroscience

Course number: 27-504-01/02

 Instructor: Dr. Ossnat Bar-Shira

TA: TBD

Academic year: 2023/4

Semester: A       Scope of hours: 4h lecture + 2h practice

Online course website: TBD

Course Goals (general and specific):

Machine learning (ML) is a branch of artificial intelligence that enables computers to learn a model given a large amount of data in a way that resembles human learning. Neural networks (NNs) are a type of machine learning model inspired by the structure and function of the neural networks inside the human brain.  In recent years, ML and in particular NNs have become very popular in many domains, and have been proved extremely effective in a range of applications, from machine vision and speech recognition to decision making and robotics.

The course covers four main topics: (1) Machine learning basics such as regression, classification, regularization and model evaluation (2) Supervised learning methods and algorithms (3) Unsupervised learning (4) Neural networks and deep learning.

The course is accompanied by set of hands-on exercises, including both theory and applied (programming) excercises.

 

Course content:

1.            Generalization and the bias-variance tradeoff in supervised learning.

2.            Linear classification, large margin classifiers

3.            Logistic regression, optimization and regulation methods

4.            Nonlinear classificatoin, SVM, Multi-layer perceptron, convolutional networks.

5.            Neural networks and backpropagation

6.            Recurrent neural networks: RNN, LSTM

7.            Unsupervised learning: Autoencoders, PCA, Infomax.

8.            Hopfield networks as associative memory networks

The course of the lessons (teaching methods, use of technology, guest lecturers):

Mostly board and chalk lectures. Sporadic pre-recorded video lectures.

A detailed teaching plan for all classes:

Introduction to models, from neuron to brain, simplified neurons.

Introduction to learning regression and generalization with polynomial regression.

Classification, logistic regression, tree based algorithms

Regularization and model evaluation

Projection to half spaces, GD, SGD, Newton method

Bias and variance, the perceptron Algorithm

Perceptron as GD. online vs batch learning, log reg (3)

Non linear classification, Multi-layer perceptrons

MLP architectures. ConvNets, ResNets

Backpropagation, multi-class and softmax

Max-Margin classification, SVM

Duality, dual SVM, kernels (advanced class)

Kernel-SVM as a network, kernel examples

Generalization theorems, word embedding

Recurrent neural networks, RNNs and LSTMs

Unsupervised learning: clustering, k-means, Multi-variate Gaussians, PCA

PCA as reconstruction, Oja

Attention, entropy fundamentals of Info theory

Measures of uncertainty, Entropy and source coding

Dkl, The mutual information. Chains. Markov MaxEnt,

Info theory - cont., Compression

Continuous Entropy. Entropy of a Gaussian

Info Max as a learning principle

Denoising auto encoders

Rehearse information theory and learning

Advanced classes: Generative models VAE

Advanced classes: Generative models VAE

Advanced classes: Generative models GAN

linear dynamical systems, fixed points & stability, high dim

Stochastic linear system, Markov chains

Markov chains, non linear systems, bifurcations

non linear systems

Integrate and fire, associative memory, hopfield model

nonlinear models FN

Hopfield model

Hopfield learning

Summary

 

Prerequisites:

Mathematics: Probability and statistics, Linear Algebra, Calculus.

Fluency in a programming language preferably Python.

 

Duties/requirements/assignments:

Students should submit 90% of home assignments, and a final project

 

Grading distribution:

Final Project 45%

Home assignments 30%

Midterm exam 25%

              

Bibliography (required and recommended reading):

Machine Learning and Pattern recognition, C. Bishop (2006)

Deep Learning, I. Goodfellow and Y. Bengio (2015)

The Elements of Statistical Learning, T. Hastie et al. (2001)

 

See full list online here:

https://sites.google.com/site/gondaneuralnetworks/home/notes-and-links