CSCE 5218 – Deep Learning
Spring 2025
Basic information:
-
Instructor: Heng Fan (heng.fan@unt.edu)
-
Office: Discovery Park F284
-
Office hours: Th 11:30 am-1:00 pm or by appointment
-
Lecture time: Tu/Th 1:00 - 2:20 pm
-
Classroom: NTDP K110
-
Syllabus: PDF
-
TA: Mingchen Li (mingchenli@my.unt.edu)
Office Hours: 3:00-5:00 pm on Wed, F221 or by appointment
-
IA: Piyush Deepak Hemnani (PiyushHemnani@my.unt.edu)
Office Hours: 12:00-2:00 pm on Friday (via appointment)
Course description
This course aims at covering the basics of modern deep neural networks. In specific, the first part will introduce the fundamental concepts in neural networks including network architecture, activation function, loss, optimization, etc. Then, the second part will describe specific types of different deep neural networks such as convolutional neural networks (CNNs), recurrent neural networks (RNNs) and attention-based Transformer, as well as their applications in computer vision and natural language processing. In the final part we will briefly discuss some recent advanced topics in deep learning including graph neural networks, unsupervised representation learning, deep reinforcement learning, generative adversarial networks (GANs), etc. In this course, the hands-on practice of implementing deep learning algorithms (in Python) will be provided via homeworks and course project.
Textbooks
We will have required readings from the following textbook:
-
Deep Learning, by Ian Goodfellow, Yoshua Bengio, and Aaron Courville, 2016. online version
Besides, the following textbooks are useful as additional references:
-
Dive into Deep Learning, by Aston Zhang, Zack C. Lipton, Mu Li, and Alex J. Smola, 2019. online version
(A lot of examples are provided to practice deep learning.)
-
Neural Networks and Deep Learning, by Michael Nielsen, 2019. online version
-
Introduction to Deep Learning, by Eugene Charniak, 2019. link
In addition to the textbooks, extra reading materials will be provided as we cover topics. Check out the
course website regularly for updated reading materials.
Announcements
-
01/13/2025: Course website has been launched.
Links
Paper review list
Important: Read the requirements (click here) for paper review. Here is a review example for your reference.
(
Paper review lists will be gradually added.)
Paper review list 1 (due on 2/8):
-
A Krizhevsky, I Sutskever, and G Hinton, ImageNet Classification with Deep Convolutional Neural Networks, NeurIPS, 2012.
-
A Paszke, et al., PyTorch: An Imperative Style, High-Performance Deep Learning Library, NeurIPS, 2019.
Paper review list 2 (due on 2/20):
-
K Simonyan and A Zisserman, Very Deep Convolutional Networks for Large-Scale Image Recognition, ICLR, 2015.
-
K. He, X. Zhang, S. Ren, and J. Sun, Deep Residual Learning for Image Recognition, CVPR, 2016.
Paper review list 3 (due on 2/29):
-
R. Girshick, J. Donahue, T. Darrell, and J. Malik, Rich feature hierarchies for accurate object detection and semantic segmentation, CVPR, 2014.
-
R. Girshick, Fast R-CNN, ICCV, 2015.
Paper review list 4 (due on 3/10):
-
S. Ren, K. He, R. Girshick, and J. Sun, Faster R-CNN: Towards real-time object detection with region proposal networks, NIPS, 2015.
-
J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, You only look once: Unified, real-time object detection, CVPR, 2016.
-
J. Long, E. Shelhamer, and T. Darrell, Fully Convolutional Networks for Semantic Segmentation, CVPR, 2015.
Paper review list 5 (due on 3/28):
-
J. Chung, C. Gulcehre, K. Cho, and Y. Bengio, Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modelling, NIPS Workshop, 2014.
-
J. Donahue, L. Hendricks, S. Guadarrama, M. Rohrbach, S. Venugopalan, K. Saenko, and T. Darrell, Long-term recurrent convolutional networks for visual recognition and description, CVPR, 2015.
Paper review list 6 (due on 4/11):
-
D. Bahdanau, K. Cho, and Y. Bengio, Neural machine translation by jointly learning to align and translate, ICLR, 2015.
-
A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. Gomez, Ł. Kaiser, and I. Polosukhin, Attention is all you need, NIPS, 2017.
-
A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, J. Uszkoreit, and N. Houlsby, An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale, ICLR, 2021.
Schedule and class notes (being updated)
Date |
Lecture |
Reading |
Note |
1/14 |
Introduction
|
- |
- |
1/16 |
Machine learning overview |
Deep Learning Ch 1-5 |
- |
1/21 |
Machine learning overview (cont)
|
Deep Learning Ch 1-5 |
- |
1/23 |
Neural network basics-1
|
Deep Learning Ch 4.2, 4.3, 6 |
- |
Grading policy
Grading will be based on the following components:
-
Attendance: 5%
-
Paper review: 30%
-
Course project: 50% (15% for proposal, 25% for final report, and 10% for presentation)
-
Final exam: 15%
Resource and Acknowledgment
This course is inspired by the following courses:
-
Introduction to Deep Learning by Adriana Kovashka, University of Pittsburgh, Fall 2021
-
Deep Learning by Dhruv Batra, Georgia Tech, Fall 2021
-
Convolutional Neural Networks for Visual Recognition by Fei-Fei Li, Justin Johnson, and Serena Young, some content by Andrej Karpathy, Stanford University, Spring 2021
-
Natural Language Processing with Deep Learning by Chris Manning, Abigail See, based on an earlier course by Richard Socher, Stanford University, Winter 2019
-
Introduction to Deep Learning by Bhiksha Raj, Carnegie Mellon University, Fall 2019
-
Computer Vision by Kristen Grauman, UT Austin, Spring 2011
-
Introduction to Machine Learning by Rebecca Hwa, University of Pittsburgh, Fall 2015
-
Introduction to Machine Learning by Dhruv Batra, Virginia Tech, Spring 2015
-
Machine Learning by Subhransu Maji, UMass Amhrest, Spring 2015
-
Machine Learning by Erik Sudderth, Brown University, Fall 2015
-
Computer Vision by Derek Hoiem, UIUC, Spring 2015
-
Natural Language Processing Ray Mooney, UT Austin, Spring 2018