Yuan Cao
Yuan Cao
Google Brain
Verified email at
Cited by
Cited by
Google's neural machine translation system: Bridging the gap between human and machine translation
Y Wu, M Schuster, Z Chen, QV Le, M Norouzi, W Macherey, M Krikun, ...
arXiv preprint arXiv:1609.08144, 2016
Simvlm: Simple visual language model pretraining with weak supervision
Z Wang, J Yu, AW Yu, Z Dai, Y Tsvetkov, Y Cao
arXiv preprint arXiv:2108.10904, 2021
Massively multilingual neural machine translation in the wild: Findings and challenges
N Arivazhagan, A Bapna, O Firat, D Lepikhin, M Johnson, M Krikun, ...
arXiv preprint arXiv:1907.05019, 2019
Hierarchical generative modeling for controllable speech synthesis
WN Hsu, Y Zhang, RJ Weiss, H Zen, Y Wu, Y Wang, Y Cao, Y Jia, Z Chen, ...
arXiv preprint arXiv:1810.07217, 2018
Gmail smart compose: Real-time assisted writing
MX Chen, BN Lee, G Bansal, Y Cao, S Zhang, J Lu, J Tsay, Y Wang, ...
Proceedings of the 25th ACM SIGKDD International Conference on Knowledge …, 2019
Lingvo: a modular and scalable framework for sequence-to-sequence modeling
J Shen, P Nguyen, Y Wu, Z Chen, MX Chen, Y Jia, A Kannan, T Sainath, ...
arXiv preprint arXiv:1902.08295, 2019
Leveraging weakly supervised data to improve end-to-end speech-to-text translation
Y Jia, M Johnson, W Macherey, RJ Weiss, Y Cao, CC Chiu, N Ari, ...
ICASSP 2019-2019 IEEE International Conference on Acoustics, Speech and …, 2019
Fully-hierarchical fine-grained prosody modeling for interpretable speech synthesis
G Sun, Y Zhang, RJ Weiss, Y Cao, H Zen, Y Wu
ICASSP 2020-2020 IEEE international conference on acoustics, speech and …, 2020
Training deeper neural machine translation models with transparent attention
A Bapna, MX Chen, O Firat, Y Cao, Y Wu
arXiv preprint arXiv:1808.07561, 2018
Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and autoregressive prosody prior
G Sun, Y Zhang, RJ Weiss, Y Cao, H Zen, A Rosenberg, B Ramabhadran, ...
ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and …, 2020
Your gan is secretly an energy-based model and you should use discriminator driven latent sampling
T Che, R Zhang, J Sohl-Dickstein, H Larochelle, L Paull, Y Cao, Y Bengio
Advances in Neural Information Processing Systems 33, 12275-12287, 2020
Gradient vaccine: Investigating and improving multi-task optimization in massively multilingual models
Z Wang, Y Tsvetkov, O Firat, Y Cao
arXiv preprint arXiv:2010.05874, 2020
Leveraging monolingual data with self-supervision for multilingual neural machine translation
A Siddhant, A Bapna, Y Cao, O Firat, M Chen, S Kudugunta, ...
arXiv preprint arXiv:2005.04816, 2020
React: Synergizing reasoning and acting in language models
S Yao, J Zhao, D Yu, N Du, I Shafran, K Narasimhan, Y Cao
arXiv preprint arXiv:2210.03629, 2022
Towards zero-label language learning
Z Wang, AW Yu, O Firat, Y Cao
arXiv preprint arXiv:2109.09193, 2021
Joshua 4.0: Packing, PRO, and paraphrases
J Ganitkevitch, Y Cao, J Weese, M Post, C Callison-Burch
Proceedings of the Seventh Workshop on Statistical Machine Translation, 283-291, 2012
Joshua 5.0: Sparser, better, faster, server
M Post, J Ganitkevitch, L Orland, J Weese, Y Cao, C Callison-Burch
Proceedings of the Eighth Workshop on Statistical Machine Translation, 206-212, 2013
Hallucinated n-best lists for discriminative language modeling
K Sagae, M Lehr, E Prud'hommeaux, P Xu, N Glenn, D Karakos, ...
2012 IEEE International Conference on Acoustics, Speech and Signal …, 2012
Effective sequence-to-sequence dialogue state tracking
J Zhao, M Mahdieh, Y Zhang, Y Cao, Y Wu
arXiv preprint arXiv:2108.13990, 2021
Joshua 6: A phrase-based and hierarchical statistical machine translation system.
M Post, Y Cao, G Kumar
Prague Bull. Math. Linguistics 104, 5-16, 2015
The system can't perform the operation now. Try again later.
Articles 1–20