Follow
Roman Novak
Roman Novak
Google Deepmind
Verified email at polytechnique.edu
Title
Cited by
Cited by
Year
Deep neural networks as gaussian processes
J Lee, Y Bahri, R Novak, SS Schoenholz, J Pennington, J Sohl-Dickstein
arXiv preprint arXiv:1711.00165, 2017
11562017
Wide neural networks of any depth evolve as linear models under gradient descent
J Lee, L Xiao, S Schoenholz, Y Bahri, R Novak, J Sohl-Dickstein, ...
Advances in neural information processing systems 32, 2019
9942019
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
A Srivastava, A Rastogi, A Rao, AAM Shoeb, A Abid, A Fisch, AR Brown, ...
arXiv preprint arXiv:2206.04615, 2022
7442022
Sensitivity and generalization in neural networks: an empirical study
R Novak, Y Bahri, DA Abolafia, J Pennington, J Sohl-Dickstein
arXiv preprint arXiv:1802.08760, 2018
4642018
Bayesian deep convolutional networks with many channels are gaussian processes
R Novak, L Xiao, J Lee, Y Bahri, G Yang, J Hron, DA Abolafia, ...
arXiv preprint arXiv:1810.05148, 2018
3332018
Neural tangents: Fast and easy infinite neural networks in python
R Novak, L Xiao, J Hron, J Lee, AA Alemi, J Sohl-Dickstein, ...
arXiv preprint arXiv:1912.02803, 2019
2332019
Finite versus infinite neural networks: an empirical study
J Lee, S Schoenholz, J Pennington, B Adlam, L Xiao, R Novak, ...
Advances in Neural Information Processing Systems 33, 15156-15172, 2020
1842020
Dataset distillation with infinitely wide convolutional networks
T Nguyen, R Novak, L Xiao, J Lee
Advances in Neural Information Processing Systems 34, 5186-5198, 2021
1822021
Infinite attention: NNGP and NTK for deep attention networks
J Hron, Y Bahri, J Sohl-Dickstein, R Novak
International Conference on Machine Learning, 4376-4386, 2020
1052020
Fast finite width neural tangent kernel
R Novak, J Sohl-Dickstein, SS Schoenholz
International Conference on Machine Learning, 17018-17044, 2022
492022
On the infinite width limit of neural networks with a standard parameterization
J Sohl-Dickstein, R Novak, SS Schoenholz, J Lee
arXiv preprint arXiv:2001.07301, 2020
482020
Exploring the neural algorithm of artistic style
Y Nikulin, R Novak
arXiv preprint arXiv:1602.07188, 2016
332016
Improving the neural algorithm of artistic style
R Novak, Y Nikulin
arXiv preprint arXiv:1605.04603, 2016
312016
Iterative refinement for machine translation
R Novak, M Auli, D Grangier
arXiv preprint arXiv:1610.06602, 2016
292016
Exact posterior distributions of wide Bayesian neural networks
J Hron, Y Bahri, R Novak, J Pennington, J Sohl-Dickstein
arXiv preprint arXiv:2006.10541, 2020
282020
Beyond human data: Scaling self-training for problem-solving with language models
A Singh, JD Co-Reyes, R Agarwal, A Anand, P Patil, PJ Liu, J Harrison, ...
arXiv preprint arXiv:2312.06585, 2023
192023
Fast neural kernel embeddings for general activations
I Han, A Zandieh, J Lee, R Novak, L Xiao, A Karbasi
Advances in neural information processing systems 35, 35657-35671, 2022
112022
Small-scale proxies for large-scale transformer training instabilities
M Wortsman, PJ Liu, L Xiao, K Everett, A Alemi, B Adlam, JD Co-Reyes, ...
arXiv preprint arXiv:2309.14322, 2023
92023
Wide Bayesian neural networks have a simple weight posterior: theory and accelerated sampling
J Hron, R Novak, J Pennington, J Sohl-Dickstein
International conference on machine learning, 8926-8945, 2022
52022
Frontier Language Models are not Robust to Adversarial Arithmetic, or" What do I need to say so you agree 2+ 2= 5?
CD Freeman, L Culp, A Parisi, ML Bileschi, GF Elsayed, A Rizkowsky, ...
arXiv preprint arXiv:2311.07587, 2023
2023
The system can't perform the operation now. Try again later.
Articles 1–20