Google Research

ProtoAttend: Attention-based prototypical learning

Journal of Machine Learning Research (2020)

Abstract

We propose a novel inherently interpretable machine learning method that bases decisions on few relevant examples that we call prototypes. Our method, ProtoAttend, can be integrated into a wide range of neural network architectures including pre-trained models. It utilizes an attention mechanism that relates the encoded representations to samples in order to determine prototypes. Without sacrificing ac- curacy of the original model, ProtoAttend yields superior results in: sample-based interpretability, confidence estimation and distribution mismatch detection.

Research Areas

Learn more about how we do research

We maintain a portfolio of research projects, providing individuals and teams the freedom to emphasize specific types of work