Before coming to Duke, I received B.S. in Statistics from Peking University in 2019. In summer 2018, I was an intern in Industrial and Systems Engineering (ISyE), Georgia Tech, working with Prof. Tuo Zhao.
My research interests are in optimization and theoretical machine learning. Recently, I am particularly interested in deep learning theory.
A Local Convergence Theory for Mildly Over-Parameterized Two-Layer Neural Network
Mo Zhou, Rong Ge, Chi Jin
Publications* denotes equal contribution.
Towards Understanding the Importance of Shortcut Connections in Residual Networks
Tianyi Liu*, Minshuo Chen*, Mo Zhou, Simon S. Du, Enlu Zhou, Tuo Zhao
Conference on Neural Information Processing Systems (NeurIPS), 2019.
Towards Understanding the Importance of Noise in Training Neural Networks
Mo Zhou*, Tianyi Liu*, Yan Li, Dachao Lin, Enlu Zhou, Tuo Zhao
International Conference on Machine Learning (ICML), 2019. Long Talk.
- CPS590.04 Machine Learning Algorithms, 2021 Spring. TA
- CPS330 Design and Analysis of Algorithms, 2020 Fall. TA
- CPS330 Design and Analysis of Algorithms, 2020 Spring. TA
Duke University, 2019 - present
Ph.D. in Computer Science
Peking University, 2015 - 2019
B.S. in Statistics