A widely deployed method for reducing the training time of deep neural
networks is to normalize activations at each layer. Although various
Use your arXiv email address to see your arXiv papers in GroundAI.
By signing up you accept our content policy
Already have an account? Sign in
No a member yet? Create an account