======CS551: Introduction to Deep Learning====== This course will provide a basic understanding of deep learning and how to solve problems from varied domains. Open source tools will be used to demonstrate different applications. =====Class timings===== * Monday - 0800-0900 * Tuesday - 0800-0900 * Wednesday - 1700-1800 =====Syllabus===== Brief introduction of big data problem. Overview of linear algebra, probability, numerical computation. Basics of Machine learning/Feature engineering. Neural network. Tutorial for Tools. Deep learning network - Shallow vs Deep network, Deep feedforward network, Gradient based learning - Cost function, soft max, sigmoid function, Hidden unit - ReLU, Logistic sigmoid, hyperbolic tangent Architecture design, SGD, Unsupervised learning - Deep Belief Network, Deep Boltzmann Machine, Factor analysis, Autoencoders. Regularization. Optimization for training deep model. Advanced topics - Convolutional Neural Network, Recurrent Neural Network/ Sequence modeling, LSTM, Reinforcement learning. Practical applications – Vision, speech, NLP, etc. =====Books===== * Ian Goodfellow, Yoshua Bengio and Aaron Courville, “Deep Learning”, Book in preparation for MIT Press, 2016. (available online) * Jerome H. Friedman, Robert Tibshirani, and Trevor Hastie, “The elements of statistical learning”, Springer Series in Statistics, 2009. * Charu C Aggarwal, "Neural Networks and Deep Learning", Springer. =====Slides===== * {{ :courses:2019:cs551:01_intro_to_dl.pdf |Introduction}} * {{ :courses:2019:cs551:02_linear_algebra.pdf | Linear Algebra and Probability}} * {{ :courses:2019:cs551:03_feature_engg.pdf | Feature engineering}} * {{ :courses:2019:cs551:04_neural_network.pdf | Neural network}} * {{ :courses:2019:cs551:keras.pdf | Introdution to Keras}} - Presented by Niraj * {{ :courses:2019:cs551:05_dl_network.pdf | DNN}} * {{ :courses:2019:cs551:06_regularization.pdf |Regularization}} * {{ :courses:2019:cs551:07_optimization.pdf |Optimization}} * {{ :courses:2019:cs551:08_cnn.pdf | CNN}} * {{ :courses:2019:cs551:09_rnn.pdf | RNN}} * {{ :courses:2019:cs551:10_practical_methods.pdf | Practical methods}} * {{ :courses:2019:cs551:11_drl.pdf |Reinforcement learning}} =====Additional Reading Materials===== * {{:courses:2017:matrixcalculus.pdf|Matrix differentiation}} by Randal J. Barnes, Department of Civil Engineering, University of Minnesota, Minneapolis, Minnesota, USA * [[http://www.visiondummy.com/2014/04/geometric-interpretation-covariance-matrix|A geometric interpretation of the covariance matrix]] * [[http://www.nature.com/nature/journal/v521/n7553/full/nature14539.html|Deep Learning, Nature]] * [[http://ieeexplore.ieee.org/abstract/document/485891/ | Artificial Neural Networks: A tutorial]] * [[http://www.asimovinstitute.org/neural-network-zoo/|The neural network zoo]] * [[https://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example/ | A Step by Step Backpropagation Example]] * [[http://www.jefkine.com/general/2016/09/05/backpropagation-in-convolutional-neural-networks/|Backpropagation in convolutional neural network]] * [[http://karpathy.github.io/2015/05/21/rnn-effectiveness/ | Andrej Karpathy blog]] * [[https://iamtrask.github.io/2015/11/15/anyone-can-code-lstm/ | Blog on LSTM-RNN]] * [[ftp://ftp.sas.com/pub/neural/illcond/illcond.html| Ill-Conditioning in Neural Networks]] =====Information related to projects===== * [[ :courses:2019:cs551:project_information |General information]] =====Other information===== * {{ :courses:2019:cs551:2018_cs551_midsem.pdf |Previous year midsem paper}}