Hossein Mobahi

Hossein Mobahi

I am a Research Scientist at Google DeepMind. I joined Google Research in 2016. Before that, I was a Postdoctoral Researcher at the Computer Science and Artificial Intelligence Laboratory (CSAIL) at MIT, where I had the privilege of working with Professor Bill Freeman and Dr. John Fisher. I earned my PhD in Computer Science from the University of Illinois at Urbana-Champaign (UIUC), where I was fortunate to be supervised by Professor Yi Ma. My broad interest lies in Artificial Intelligence, with a specific focus on the intersection of Machine Learning and Optimization. My research is often guided by mathematical principles, aiming to develop practically successful methods that offer greater clarity in their understanding or improved performance. I am a co-creator of the Sharpness Aware Minimization (SAM) method. My work also includes contributions to the theory of self-distillation, optimization by the continuation method, and understanding the loss surface of neural networks. I have a long-standing interest in radically new approaches to creating neural architectures, particularly those inspired by biology and the human brain, as well as curriculum learning strategies that aim to enhance training efficiency and generalization through more effective data presentation.
Authored Publications
Sort By
  • Title
  • Title, descending
  • Year
  • Year, descending
    Google
Sharpness-Aware Minimization Improves Language Model Generalization
Yi Tay
Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) (2022), pp. 7360-7371
Sharpness-aware Minimization for Efficiently Improving Generalization
Pierre Foret
Ariel Kleiner
Behnam Neyshabur
ICLR Spotlight (2021)
Methods and Analysis of The First Competition in Predicting Generalization of Deep Learning
Yiding Jiang
Parth Natekar
Manik Sharma
Sumukh K. Aithal
Dhruva Kashyap
Natarajan Subramanyam
Carlos Lassance
Daniel M. Roy
Gintare Karolina Dziugaite
Suriya Gunasekar
Isabelle Guyon
Pierre Foret
Scott Yak i
Behnam Neyshabur
Samy Bengio
Proceedings of the NeurIPS 2020 Competition and Demonstration Track, PMLR (2021)
NeurIPS 2020 Competition: Predicting Generalization in Deep Learning
Yiding Jiang
Pierre Foret
Scott Yak
Daniel M. Roy
Gintare Karolina Dziugaite
Samy Bengio
Suriya Gunasekar
Isabelle Guyon
Behnam Neyshabur
arXiv (2020)
Self-Distillation Amplifies Regularization in Hilbert Space
Mehrdad Farajtabar
Peter Bartlett
Neural Information Processing Systems (NeurIPS) (2020)
Fantastic Generalization Measures and Where to Find Them
Yiding Jiang
Behnam Neyshabur
Dilip Krishnan
Samy Bengio
ICLR (2020)
A Margin-Based Measure of Generalization for Deep Networks
Yiding Jiang
Dilip Krishnan
Samy Bengio
ICLR (2019)
Large Margin Deep Networks for Classification
Gamaleldin Fathy Elsayed
Dilip Krishnan
Samy Bengio
NeurIPS (2018)
Homotopy Analysis for Tensor PCA
Anima Anandkumar
Rong Ge
Conference on Learning Theory (2017), pp. 79-104