Our mission is to spread useful AI effectively around the world.
About the team
The Google Cloud AI research team tackles underexplored, real-world challenges for Google Cloud customers. We work on a range of inspiring problems based on Google Cloud customer needs, identifying research topics that maximize both scientific and real-world impact. As such, we collaborate closely with product teams to put our research results in the hands of our customers, and publish the findings in top ML venues.
Innovations coming out of the Google Cloud AI research team help explain the behavior of sophisticated machine learning models and improve current capabilities by enabling more efficient use of data. Additionally, we proactively identify market needs, and work with customers to identify specific use cases where innovation is needed (e.g. recommendation systems, document understanding, infectious disease modeling).
Research areas
Team focus summaries
Highlighted projects
A novel graph-based architecture with Rich Attention Transformer designed for form document understanding. FormNet recovers local syntactic information and achieves SOTA performance on public benchmarks.
A model agnostic approach that allows users to specify input-output relationships that their system should obey and integrates them with a DNN in a way that can be modified at inference time.
A novel continual learning framework that avoids catastrophic forgetting and maintains high accuracy without having to retain past training data.
An anomaly detection method using an ensemble of one-class classifiers and a self-supervised data representation and refinement process to achieve robust results on a completely uncurated dataset.
A framework that locally distills a block box model into an interpretable model of our choice (e.g. shallow decision tree or linear regression) without sacrificing performance.
A new deep learning model for time-series that beats other algorithms by a large margin and provides useful explanations in various forms.
A new deep learning method for tabular data that improves over other DNN and ensemble decision tree models on many datasets and provides interpretable insights.
A framework that enables quantitatively compute the importance of each training sample for the model, which yields better quality estimates than the alternatives with significant time savings.
Featured publications
Association for Computational Linguistics (ACL) (2021)
Transactions on Machine Learning Research (TMLR) (2022)
International Journal of Forecasting (2021)
International Conference on Learning Representations (ICLR) (2021)