This course is an elementary introduction to a machine learning technique called deep learning, as well as its applications to a variety of domains. Along the way, the course also provides an intuitive introduction to machine learning such as simple models, learning paradigms, optimization, overfitting, importance of data, training caveats, etc. The assignments explore key concepts and simple applications, and the final project allows an in-depth exploration of a particular application area. By the end of the course, you will have an overview on the deep learning landscape and its applications. You will also have a working knowledge of several types of neural networks, be able to implement and train them, and have a basic understanding of their inner workings.
IRB 1116
Tuesday, Thursday 2:00pm - 3:15pm
Abhinav Shrivastava
abhinav@cs.umd.edu
Office hours: TBD (request by email for now).
Matt Gwilliam
mgwillia@umd.edu
Office hours: Wednesdays 2 PM, IRB-4119
Nirat Saini
nirat@umd.edu
Office hours: TBA, once per assignment/exam
Saksham Suri
sakshams@umd.edu
Office hours: TBA, once per assignment/exam
Hanyu Wang
hywang66@umd.edu
Office hours: TBA, once per assignment/exam
Date | Topic | Slides | Notes & Assignments | |
---|---|---|---|---|
January 25 |
Course Introduction
Motivation Goals, Syllabus, Assignments Policies |
slides | ||
Machine Learning Basics - I
|
||||
January 30 library_books February 1 |
Introduction to Statistical Learning
Simple models
Paradigms of learning
|
slides | Assignment 1 out (Gradescope) |
|
Neural Networks Basics - I
|
||||
February 1 Februray 6 |
Introduction to Neural Networks
Terminology Simple Neural Networks Non-linearities |
slides | ||
Februray 6library_books
Februray 8 |
Problem setup: labels and losses
Types of problems and labels Loss functions |
slides | Assignment 1 due | |
Machine Learning Basics - II
|
||||
February 8
February 13 February 15 |
Optimization basics
Loss function derivatives, minimasGradient descent, Stochastic gradient descent |
slides | ||
Neural Network Basics - II
|
||||
February 15
Febuary 20 library_books Febuary 27 |
Training Neural Networks
InitializationBackpropagation Optimization & hyperparameters |
slides
(backprop) slides (opt) |
Assignment 2 out (Feb 20) |
|
February 27 February 29 library_books |
Training Caveats (Neural Networks and ML models)
OverfittingBias/Variance trade-off Optimization & hyperparameters
Improving Performance of Neural Networks
Optimization tips and tricksBest principles |
(see above) | Assignment 2 due (Mar 1) |
|
Computation Resources Overview
|
||||
February 22 |
Computational Infrastructure (TA Lecture)
Quick walk-through: Colaboratory, Google Cloud Platform Implementation: Neural Network Basics and Run-through (TA Lecture) Introduction to PyTorch Handling data (images, text) Structuring a neural network and machine learning codebase |
slides | ||
Exam:
|
||||
March 5library_books |
Midterm Exam
In class
|
|||
March 19 March 21 |
Spring Break | |||
Convolutional Neural Networks (ConvNets)
|
||||
March 12 March 14 library_books |
Introduction to ConvNets
ConvolutionsPooling |
slides | Assignment 3 out (challenge) (Mar 12) |
|
March 14 March 26 library_books |
ConvNet Architectures
Popular architectures (primarily images, brief overview of videos)Intuitions and key-insights Design principles |
(see above) | Assignment 3 due (Mar 26) Project Proposal due (Mar 15) |
|
Applications of ConvNets
|
||||
March 28 library_books
April 2 April 9 |
Application: Dense Prediction | slides | Assignment 4 out (challenge) (Mar 28) |
|
April 2 April 4 April 9 library_books |
Application: Object Detection | slides | Assignment 4 due (Apr 9) |
|
Recurrent Neural Networks (RNNs)
|
||||
April 9 April 11 library_books |
Introduction to Recurrent Networks
RNNs, GRUs, LSTMsText/Language Applications Language Modeling |
slides | ||
April 16 April 18 |
Introduction to Self-attention or Transformers
Self-attenton or Transformers |
slides | ||
Advanced Topics
|
||||
April 16 April 18 |
Vision + Language (models, tasks, training) | (see above) | ||
April 25 April 30 May 2 |
Image Generative Models
Auto regressive modelsGANs, pix2pix, CycleGAN, etc. VAEs Teasers: Text Generation, Self-supervised Learning |
Minimum grade of C- in CMSC330 and CMSC351; and 1 course with a minimum grade of C- from (MATH240, MATH461); and permission of CMNS-Computer Science department.
We will work extensively with probability, statistics, mathematical functions such as logarithms and differentiation, and linear algebra concepts such as vectors and matrices. You should be comfortable manipulating these concepts.
We will use of the Python programming language. It is assumed that you know or will quickly learn how to program in Python. The programming assignments will be oriented toward Unix-like operating systems. While it is possible to complete the course using other operating systems, you will be solely responsible for troubleshooting any issues you encounter.
If you are unsure that you have the required prerequisites, consult with the instructor.
Here's how you will be graded:
Collaboration: Students are expected to finish the homeworks by himself/herself, but discussion on the assignments is allowed (and encouraged). The people you discussed with on assignments should be clearly detailed: before the solution to each question, list all people that you discussed with on that particular question. In addition, each student should submit his/her own code and mention anyone he/she collaborated with.
Details will be announced in class.
Syllabus subject to change.
The course should be self contained, but if you need additional reading material, you can consult the following:
Helpful review and reference material:
If you need reference/additional readings for statistical learning, you can consult the following:
Note that academic dishonesty includes not only cheating, fabrication, and plagiarism, but also includes helping other students commit acts of academic dishonesty by allowing them to obtain copies of your work. In short, all submitted work must be your own. Cases of academic dishonesty will be pursued to the fullest extent possible as stipulated by the Office of Student Conduct. It is very important for you to be aware of the consequences of cheating, fabrication, facilitation, and plagiarism. For more information on the Code of Academic Integrity or the Student Honor Council, please visit http://www.shc.umd.edu.
Any student who needs to be excused for an absence from a single lecture, recitation, or lab due to a medically necessitated absence shall:
Any student who needs to be excused for a Major Scheduled Grading Event, must provide written documentation of the illness from the Health Center or from an outside health care provider. This documentation must verify dates of treatment and indicate the time frame that the student was unable to meet academic responsibilities. No diagnostic information shall be given. The Major Scheduled Grading Events for this course include midterm and final exam. For class presentations, the instructor will help the student swap their presentation slot with other students.
It is also the student's responsibility to inform the instructor of any intended absences from exams and class presentations for religious observances in advance. Notice should be provided as soon as possible, but no later than the Monday prior to the the midterm exam, the class presentation date, and the final exam.
Any student eligible for and requesting reasonable academic accommodations due to a disability is requested to provide a letter of accommodation from the Office of Disability Support Services within the first three weeks of the semester.
You can find the university’s course policies here.