Deep Learning / Jan-May 2025
Updates
- Jan 09: New Lecture is up: (dl-03) Representational Power of an MLP [ReLU-activation-code] [sigmoid-activation-codes] [slides]
- Jan 08: New Lecture is up: (dl-02) Network of perceptrons [slides]
- Jan 06: New Lecture is up: (dl-01) Artificial Neuron Models (MP neuron and Perceptron) [perceptron-learning-code] [slides]
- Jan 02: New Lecture is up: (dl-00) Introduction [slides]
-
Jan 1, 2025: The course website is up!
Course Description
Deep Learning has lately become the driving force behind numerous high-performing AI/ML products deployed in real-world across diverse disciplines. Tech giants such as Google, Microsoft, Facebook, Amazon, etc. have strongly been employing the Deep Learning workforce in the past few years for developing various applications in Computer Vision, Natural Processing, etc. Also, various organizations, such as OpenAI, DeepMind, etc., that intend to work towards developing safe and responsible Artificial General Intelligence (AGI) conduct research in Deep Learning. Hence, it has recently become one of the most sought-after courses. In this course, we will discuss (with a focus on the implementation) the various building blocks required to realize the Deep Learning solutions from a simple neuron model to Generative AI.
Deep Learning (AI2100, AI5100 and CS5480) Course Contents
Starting from an artificial neuron model, the aim of this course is to understand feed-forward, recurrent architectures of Artificial Neural Networks, all the way to the latest Generative AI models driven by Deep Neural Networks. Specifically, we will discuss the basic Neuron models (McCulloch Pitts, Perceptron), Multi-Layer Perceptron (MLP), Convolutional Neural Networks (CNN), and Recurrent Neural Networks (RNN, LSTM and GRU). We will understand these models’ representational ability and how to train them using the Gradient Descent technique using the Backpropagation algorithm. We will then discuss the encoder-decoder architecture, attention mechanism and its variants. That will be followed by self-attention and Transformers. The next part of the course will be on Generative AI, wherein we will discuss Variational Autoencoders, GANs, Diffusion Models, GPT, BERT, etc. We will briefly discuss multi-modal representation learning (e.g., CLIP). Towards the end, students will be briefly exposed to some of the advanced topics and/or recent trends in deep learning.
Prerequisites
A course on Machine Learning (e.g., AI2000, CS3390, EE2802, EE5913, EE5610) and Programming experience in Python
Logistics
Class Room: LHC-7
Timings: Slot-B (Monday-10:00-10:55, Wednesday-09:00-09:55, Thursday-11:00-11:55)
Visit this page regularly for updates and information regarding the course.
Previous Offerings
Instructors
Teaching Assistants
Sunayna Padhye
Madhumitha V
Naveen George
Harsh Udai
Anish Pawar
Puneet Rajan