Yuanzhi Li
Yuanzhi Li
Assistant Professor at CMU
Verified email at andrew.cmu.edu - Homepage
Title
Cited by
Cited by
Year
A convergence theory for deep learning via over-parameterization
Z Allen-Zhu, Y Li, Z Song
International Conference on Machine Learning, 242-252, 2019
3642019
Convergence analysis of two-layer neural networks with relu activation
Y Li, Y Yuan
Advances in neural information processing systems, 597-607, 2017
3182017
A theoretical analysis of NDCG ranking measures
Y Wang, L Wang, Y Li, D He, W Chen, TY Liu
Proceedings of the 26th annual conference on learning theory (COLT 2013) 8, 6, 2013
283*2013
A latent variable model approach to pmi-based word embeddings
S Arora, Y Li, Y Liang, T Ma, A Risteski
Transactions of the Association for Computational Linguistics 4, 385-399, 2016
258*2016
Learning and generalization in overparameterized neural networks, going beyond two layers
Z Allen-Zhu, Y Li, Y Liang
Advances in neural information processing systems, 6158-6169, 2019
2312019
Learning overparameterized neural networks via stochastic gradient descent on structured data
Y Li, Y Liang
Advances in Neural Information Processing Systems, 8157-8166, 2018
2312018
Linear algebraic structure of word senses, with applications to polysemy
S Arora, Y Li, Y Liang, T Ma, A Risteski
Transactions of the Association for Computational Linguistics 6, 483-495, 2018
1172018
Algorithmic regularization in over-parameterized matrix sensing and neural networks with quadratic activations
Y Li, T Ma, H Zhang
Conference On Learning Theory, 2-47, 2018
1142018
An alternative view: When does SGD escape local minima?
R Kleinberg, Y Li, Y Yuan
arXiv preprint arXiv:1802.06175, 2018
1082018
LazySVD: Even faster SVD decomposition yet without agonizing pain
Z Allen-Zhu, Y Li
Advances in Neural Information Processing Systems, 974-982, 2016
832016
Neon2: Finding local minima via first-order oracles
Z Allen-Zhu, Y Li
Advances in Neural Information Processing Systems, 3716-3726, 2018
742018
First efficient convergence for streaming k-pca: a global, gap-free, and near-optimal rate
Z Allen-Zhu, Y Li
2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS …, 2017
612017
Much faster algorithms for matrix scaling
Z Allen-Zhu, Y Li, R Oliveira, A Wigderson
2017 IEEE 58th Annual Symposium on Foundations of Computer Science (FOCS …, 2017
602017
On the convergence rate of training recurrent neural networks
Z Allen-Zhu, Y Li, Z Song
Advances in neural information processing systems, 6676-6688, 2019
522019
Algorithmic framework for model-based deep reinforcement learning with theoretical guarantees
Y Luo, H Xu, Y Li, Y Tian, T Darrell, T Ma
arXiv preprint arXiv:1807.03858, 2018
512018
Towards explaining the regularization effect of initial large learning rate in training neural networks
Y Li, C Wei, T Ma
Advances in Neural Information Processing Systems, 11674-11685, 2019
422019
What Can ResNet Learn Efficiently, Going Beyond Kernels?
Z Allen-Zhu, Y Li
Advances in Neural Information Processing Systems, 9017-9028, 2019
362019
Near-optimal design of experiments via regret minimization
Z Allen-Zhu, Y Li, A Singh, Y Wang
International Conference on Machine Learning, 126-135, 2017
362017
Doubly accelerated methods for faster CCA and generalized eigendecomposition
Z Allen-Zhu, Y Li
International Conference on Machine Learning, 98-106, 2017
332017
Linear convergence of a frank-wolfe type algorithm over trace-norm balls
Z Allen-Zhu, E Hazan, W Hu, Y Li
Advances in Neural Information Processing Systems, 6191-6200, 2017
282017
The system can't perform the operation now. Try again later.
Articles 1–20