User profiles for You Yang

Yang You

Presidential Young Professor, National University of Singapore
Verified email at comp.nus.edu.sg
Cited by 6626

Large batch optimization for deep learning: Training bert in 76 minutes

Y You, J Li, S Reddi, J Hseu, S Kumar… - arXiv preprint arXiv …, 2019 - arxiv.org
Training large deep neural networks on massive datasets is computationally very challenging.
There has been recent surge in interest in using large batch stochastic optimization …

Recent Advances in the Chemical Synthesis of C-Glycosides

Y Yang, B Yu - Chemical reviews, 2017 - ACS Publications
Advances in the chemical synthesis of C-pyranosides/furanosides are summarized,
covering the literature from 2000 to 2016. The majority of the methods take advantage of the …

Rational design of electrode materials for advanced supercapacitors: from lab research to commercialization

J Huang, Y Xie, Y You, J Yuan, Q Xu… - Advanced Functional …, 2023 - Wiley Online Library
Supercapacitors can harvest electrical energy from intermittent sources and transfer it quickly,
but their specific energy must be raised if they are applied to efficiently power wearable …

Crosstalk among jasmonate, salicylate and ethylene signaling pathways in plant disease and immune responses

YX Yang, GJ Ahammed, C Wu, S Fan… - Current Protein and …, 2015 - ingentaconnect.com
Phytohormone crosstalk is crucial for plant defenses against pathogens and insects in
which salicylic acid (SA), jasmonic acid (JA), and ethylene (ET) play key roles. These low …

A stabilized conforming nodal integration for Galerkin mesh‐free methods

JS Chen, CT Wu, S Yoon, Y You - International journal for …, 2001 - Wiley Online Library
Abstract Domain integration by Gauss quadrature in the Galerkin mesh‐free methods adds
considerable complexity to solution procedures. Direct nodal integration, on the other hand, …

Large batch training of convolutional networks

Y You, I Gitman, B Ginsburg - arXiv preprint arXiv:1708.03888, 2017 - arxiv.org
… The logic behind linear LR scaling is straight-forward: if you increase B by k while keeping
the number of epochs unchanged, you will do k fewer steps. So it seems natural to increase …

Open-sora: Democratizing efficient video production for all

Z Zheng, X Peng, T Yang, C Shen, S Li, H Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Vision and language are the two foundational senses for humans, and they build up our
cognitive ability and intelligence. While significant breakthroughs have been made in AI …

Imagenet training in minutes

Y You, Z Zhang, CJ Hsieh, J Demmel… - Proceedings of the 47th …, 2018 - dl.acm.org
In this paper, we investigate large scale computers' capability of speeding up deep neural
networks (DNN) training. Our approach is to use large batch size, powered by the Layer-wise …

Super gas barrier and selectivity of graphene oxide-polymer multilayer thin films.

YH Yang, L Bolling, MA Priolo… - … Materials (Deerfield Beach …, 2012 - europepmc.org
Super gas barrier thin films, fabricated with layer-by-layer assembly of polyethylenimine and
graphene oxide, exhibit significantly reduced oxygen and carbon dioxide transmission rates …

[PDF][PDF] Scaling sgd batch size to 32k for imagenet training

Y You, I Gitman, B Ginsburg - arXiv preprint arXiv:1708.03888, 2017 - arxiv.org
1 arXiv: 1708.03888 v1 [cs. CV] 13 Aug 2017 layers based on the norm of the weights (|| w||)
and the norm of the gradients (||∇ w||). By using LARS LR, we can scale the batch size to …