Simon Shaolei Du
Simon Shaolei Du
Assistant Professor, School of Computer Science and Engineering, University of Washington
Verified email at cs.washington.edu - Homepage
Title
Cited by
Cited by
Year
Gradient descent provably optimizes over-parameterized neural networks
SS Du, X Zhai, B Poczos, A Singh
International Conference on Learning Representations 2019, 2018
6782018
Gradient descent finds global minima of deep neural networks
SS Du, JD Lee, H Li, L Wang, X Zhai
International Conference on Machine Learning 2019, 2018
6422018
Fine-grained analysis of optimization and generalization for overparameterized two-layer neural networks
S Arora, SS Du, W Hu, Z Li, R Wang
International Conference on Machine Learning 2019, 2019
4912019
On exact computation with an infinitely wide neural net
S Arora, SS Du, W Hu, Z Li, R Salakhutdinov, R Wang
arXiv preprint arXiv:1904.11955, 2019
4172019
On the power of over-parametrization in neural networks with quadratic activation
SS Du, JD Lee
International Conference on Machine Learning 2018, 2018
1872018
Gradient Descent Learns One-hidden-layer CNN: Don't be Afraid of Spurious Local Minima
SS Du, JD Lee, Y Tian, B Poczos, A Singh
International Conference on Machine Learning 2018, 2017
1862017
Gradient descent can take exponential time to escape saddle points
SS Du, C Jin, JD Lee, MI Jordan, B Poczos, A Singh
arXiv preprint arXiv:1705.10412, 2017
1782017
Stochastic variance reduction methods for policy evaluation
SS Du, J Chen, L Li, L Xiao, D Zhou
International Conference on Machine Learning 2017, 2017
1142017
When is a convolutional filter easy to learn?
SS Du, JD Lee, Y Tian
International Conference on Learning Representations 2018, 2017
1122017
Computationally efficient robust estimation of sparse functionals
SS Du, S Balakrishnan, A Singh
Conference on Learning Theory, 2017, 2017
109*2017
What Can Neural Networks Reason About?
K Xu, J Li, M Zhang, SS Du, K Kawarabayashi, S Jegelka
International Conference on Learning Representations 2020, 2019
1052019
Graph Neural Tangent Kernel: Fusing Graph Neural Networks with Graph Kernels
SS Du, K Hou, B Póczos, R Salakhutdinov, R Wang, K Xu
Advances in Neural Information Processing Systems 2019, 2019
1022019
Algorithmic regularization in learning deep homogeneous models: Layers are automatically balanced
SS Du, W Hu, JD Lee
arXiv preprint arXiv:1806.00900, 2018
992018
Is a Good Representation Sufficient for Sample Efficient Reinforcement Learning?
SS Du, SM Kakade, R Wang, LF Yang
International Conference on Learning Representation 2020, 2019
962019
Provably efficient RL with rich observations via latent state decoding
SS Du, A Krishnamurthy, N Jiang, A Agarwal, M Dudík, J Langford
International Conference on Machine Learning 2019, 2019
932019
Understanding the acceleration phenomenon via high-resolution differential equations
B Shi, SS Du, MI Jordan, WJ Su
arXiv preprint arXiv:1810.08907, 2018
932018
Linear convergence of the primal-dual gradient method for convex-concave saddle point problems without strong convexity
SS Du, W Hu
International Conference on Artificial Intelligence and Statistics 2019, 2018
832018
Harnessing the power of infinitely wide deep nets on small-data tasks
S Arora, SS Du, Z Li, R Salakhutdinov, R Wang, D Yu
International Conference on Learning Representations 2020, 2019
802019
Few-shot learning via learning the representation, provably
SS Du, W Hu, SM Kakade, JD Lee, Q Lei
arXiv preprint arXiv:2002.09434, 2020
772020
How neural networks extrapolate: From feedforward to graph neural networks
K Xu, M Zhang, J Li, SS Du, K Kawarabayashi, S Jegelka
arXiv preprint arXiv:2009.11848, 2020
682020
The system can't perform the operation now. Try again later.
Articles 1–20