Hugo Future Imperfect Slim

Data Scaling Laws in NMT: The Effect of Noise and Architecture Under Review
Yamini Bansal, Behrooz Ghorbani, Ankush Garg, Biao Zhang, Colin Cherry, Maxim Krikun, Behnam Neyshabur, Orhan Firat

Revisiting Model Stitching to Compare Neural Representations NeurIPS 2021
Yamini Bansal, Preetum Nakkiran, Boaz Barak

For self-supervised learning, Rationality implies generalization, provably ICLR 2021
Yamini Bansal*, Gal Kaplun*, Boaz Barak
Blog       Talk

Distributional Generalization: A New Kind of Generalization ICML Workshop on Overparameterization: Pitfalls and Opportunities Preetum Nakkiran*, Yamini Bansal*
Talk

Improving the Reconstruction of Disentangled Representation Learners via Multi-Stage Modelling Under Review
Akash Srivastava*, Yamini Bansal*, Yukun Ding*, Cole Hurwitz*, Kai Xu, Prasanna Sattigeri, Bernhard Egger, David D. Cox, Josh Tenenbaum, Dan Gutfreund

Deep Double Descent: Where Bigger Models and More Data Hurt ICLR 2020
Preetum Nakkiran, Gal Kaplun*, Yamini Bansal*, Tristan Yang, Boaz Barak, Ilya Sutskever
Blog       Shorter Blog

Minnorm training: an algorithm for training over-parameterized deep neural networks Manuscript
Yamini Bansal, Madhu Advani, David Cox, Andrew Saxe

On the information bottleneck theory of deep learning ICLR 2018
Andrew Saxe, Yamini Bansal, Joel Dapello, Madhu Advani, Artemy Kolchinsky, Brendan Tracey, David Cox


* co-authorship