The "Black" Art & "Elusive" Science of Training Deep Neural Networks

School of Computing, University of Nebraska-Lincoln Spring 2021 (3-week session): CSCE 496/896

Synopsis: Deep Learning, a cornerstone of modern Artificial Intelligence, tackles complex challenges in computer vision, natural language processing, and speech recognition. Deep Neural Networks (DNNs)—vast networks of artificial neurons arranged in hierarchical layers—uncover intricate patterns by creating increasingly meaningful data representations. Unlocking their potential requires a deep understanding of effective architectures, learning algorithms, and optimization strategies. Often dubbed a "black art" due to its complexity, training DNNs can feel like summoning a genie. This course aims to demystify this process, exploring state-of-the-art DNN architectures and delving into the "black" art and "elusive" science of training them.

Philosophy: Inspired by my mentor Richard Feynman's philosophy on learning—“What I cannot create, I do not understand”—this course emphasizes hands-on implementation. True mastery of DNN training comes from building models from scratch, without high-level libraries like Scikit-Learn, Keras, or PyTorch. Begin with a Multi-Layer Perceptron (MLP) using NumPy (pseudocode), then tackle a Convolutional Neural Network (CNN) using NumPy (pseudocode).

Instructor
Dr. M. R. Hasan
Office Hours
See the course Canvas page
Lecture Time
Monday through Friday: 10:00 AM - 11:30 AM via Zoom

Explore my GitHub tutorials on Machine Learning and Deep Learning for additional resources. Key repositories include:

Schedule

Week Date Topic & PDF Slides Video Links
1 Jan 4
1 Jan 5
1 Jan 6

Jupyter Notebook:

1 Jan 7
1 Jan 8

Jupyter Notebooks:

1 Jan 9
2 Jan 11
2 Jan 12
2 Jan 13
2 Jan 14
2 Jan 15
2 Jan 17
3 Jan 18 Martin Luther King Holiday
3 Jan 19

CNN Visualization:

Jupyter Notebooks:

3 Jan 20
3 Jan 21

Jupyter Notebooks:

Text Resources
  • Lecture slides and Jupyter notebooks provide a thorough and extensive account of the topics.
  • Core Deep Learning Texts:
    • Dive into Deep Learning by Aston Zhang, Zachary C. Lipton, Mu Li, and Alexander J. Smola
    • Deep Learning by Ian Goodfellow, Yoshua Bengio, and Aaron Courville
    • Deep Learning with Python by Francois Chollet
    • Hands-On Machine Learning with Scikit-Learn, Keras and TensorFlow (2nd Edition, 2019) by Aurélien Géron (O'Reilly)
  • Foundational Machine Learning Texts:
    • Machine Learning: A Probabilistic Perspective by Kevin P. Murphy
    • Pattern Recognition and Machine Learning by Christopher M. Bishop
    • Introduction to Machine Learning (3rd ed.) by Ethem Alpaydin
  • Python Refresher/Introduction:
    • Data Science from Scratch by Joel Grus (O’Reilly)
    • Python for Data Analysis (2nd Edition) by Wes McKinney (O'Reilly)
    • Python Machine Learning by Sebastian Raschka (Packt Publishing)
  • Statistics, Linear Algebra & Calculus:
    • Advanced Engineering Mathematics (10th Ed.) by Erwin Kreyszig
    • All of Statistics: A Concise Course in Statistical Inference by Larry Wasserman
  • Interesting & Enlightening Texts:
    • The Master Algorithm by Pedro Domingos
    • Artificial Intelligence: A Guide for Thinking Humans by Melanie Mitchell
    • The Deep Learning Revolution by Terrence J. Sejnowski
    • Rebooting AI: Building Artificial Intelligence We Can Trust by Gary Marcus and Ernest Davis
    • Thinking, Fast and Slow by Daniel Kahneman
    • Interpretable Machine Learning by Christoph Molnar
Deep Learning Courses Elsewhere
Collaboration Tool
Google Colab Tutorials
Python
Open Data Repositories
ML Podcasts
Journals
Conference Proceedings