From post

Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10, 000-Layer Vanilla Convolutional Neural Networks.

, , , , и . ICML, том 80 из Proceedings of Machine Learning Research, стр. 5389-5398. PMLR, (2018)

Please choose a person to relate this publication to

To differ between persons with the same name, the academic degree and the title of an important publication will be displayed.

 

Другие публикации лиц с тем же именем

Neural Tangents: Fast and Easy Infinite Neural Networks in Python., , , , , , и . ICLR, OpenReview.net, (2020)Beyond Human Data: Scaling Self-Training for Problem-Solving with Language Models., , , , , , , , , и 31 other автор(ы). Trans. Mach. Learn. Res., (2024)Training Language Models on the Knowledge Graph: Insights on Hallucinations and Their Detectability., , , , , , , , , и 21 other автор(ы). CoRR, (2024)Precise Learning Curves and Higher-Order Scalings for Dot-product Kernel Regression., , , , и . NeurIPS, (2022)Disentangling Trainability and Generalization in Deep Neural Networks., , и . ICML, том 119 из Proceedings of Machine Learning Research, стр. 10462-10472. PMLR, (2020)The Surprising Simplicity of the Early-Time Learning Dynamics of Neural Networks., , , и . NeurIPS, (2020)Finite Versus Infinite Neural Networks: an Empirical Study., , , , , , и . NeurIPS, (2020)Small-scale proxies for large-scale Transformer training instabilities., , , , , , , , , и 6 other автор(ы). ICLR, OpenReview.net, (2024)Bayesian Deep Convolutional Networks with Many Channels are Gaussian Processes, , , , , , , , и . (2018)cite arxiv:1810.05148Comment: Published as a conference paper at ICLR 2019.Frontier Language Models are not Robust to Adversarial Arithmetic, or "What do I need to say so you agree 2+2=5?, , , , , , , , , и 20 other автор(ы). CoRR, (2023)