This class provides an example-based introduction to deep learning using the Keras libary. Monday lectures will focus on a general background of different machine learning techniques including convolutional neural networks, recurrent neural networks, long term short-term memory, as well as applications in image recognition, control and natural language processing. Wednesday lectures will provide an overview over relevant tools for data acquisition and processing, followed by student-driven presentations of selected research papers and homeworks.
Homeworks and research paper reviewed in class must be summarized by an interactive Jupyter notebook, which will be hosted online, that summarizes the paper and lets the reader experience its content by example. The final deliverable for the class is a report on an independent research project consisting of a 3-page paper in IEEE double-column format (overleaf) and a Jupyter notebook demonstrating a technique learned in class on a real-world data set.
MW 4.30-5.45 in ECES 114
- Required: Deep Learning with Keras by Antonio Gulli and Sujit Pal
- Recommended: Pattern Recognition and Machine Learning, Bishop
- Srivastava, Nitish, Geoffrey Hinton, Alex Krizhevsky, Ilya Sutskever, and Ruslan Salakhutdinov. “Dropout: a simple way to prevent neural networks from overfitting.” The journal of machine learning research 15, no. 1 (2014): 1929-1958. http://jmlr.org/papers/v15/srivastava14a.html (presented by Annelise Lynch, February 12).
- Kingma, Diederik P., and Jimmy Ba. “Adam: A method for stochastic optimization.” arXiv preprint arXiv:1412.6980 (2014). https://arxiv.org/abs/1412.6980 (presented by Divya Athoopallil, February 19)
- Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X. and Metaxas, D.N., 2017. Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks. In Proceedings of the IEEE international conference on computer vision (pp. 5907-5915) (presented by Trevor Grant, February 26)
- Redmon, J., Divvala, S., Girshick, R. and Farhadi, A., 2016. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 779-788) (presented by Ashwin Vasan, March 4)
- Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wierstra, D. and Riedmiller, M., 2013. Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602 (presented by Arturo Freydig Avila, March 11)
- Devlin, J., Chang, M.W., Lee, K. and Toutanova, K., 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (presented by Soumyajyoti Bhattacharya, March 18)
10% In-class participation
12% Homework 1: Implementing a simple classification/regression problem
12% Homework 2: Classification/regression on time series data
26% Jupyter notebook summary of a selected paper
40% Final project
Extra credit: narrated YouTube video
As homework and project are being submitted to a public repository, late submissions will lead to a reduction by one letter grade (A->B, B->C etc.).
Week 1: Perceptron algorithm
Week 2: MLK day – Multi-layer networks and back-propagation
Week 3: Deep convolutional neural networks
Week 4: Very deep convolutional networks
Week 5: Generative Adversarial Networks (GAN)
Week 6: Other applications for GANs (WaveNet)
Week 7: Word embeddings
Week 8: Other NLP applications
Week 9: Recurrent Neural Networks (RNN)
Week 10: Long short term memory (LSTM)
Week 11: Regression networks
Week 12: Autoencoders
Week 13: Reinforcement learning
Week 14: Project
Week 15: Project
Week 16: Project
Please follow this link for additional policies regarding accomodations, class room behavior, preferred student names and pronouns, honor code, sexual misconduct, and religious holidays.