Google Research

Appearance-and-Relation Networks for Video Classification

  • Limin Wang
  • Wei Li
  • Wen Li
  • Luc Van Gool
arXiv (2017)


Spatiotemporal feature learning in videos is a fundamental and difficult problem in computer vision. This paper presents a new architecture, termed as Appearance-and-Relation Networks (ARTNets), to learn video representation in an end-to-end manner. ARTNet is constructed by stacking multiple generic building blocks, called as SMART, whose goal is to simultaneously model appearance and relation from RGB input in a separate and explicit manner. Specifically, SMART block decouples the problem of spatiotemporal feature learning into an appearance branch for spatial modeling and a relation branch for temporal modeling. Appearance branch is implemented based on the linear combination of pixels or filter responses in each frame, while relation branch is designed based on the multiplicative interactions between pixels or filter responses across multiple frames. We perform experiments on three action recognition benchmarks: Kinetics, UCF101, and HMDB51, demonstrating that SMART block obtains an evident improvement over 3D convolution for spatiotemporal feature learning. Under the same training setting, ARTNets achieve superior performance on these three datasets to the existing state-of-the-art methods. The code is at

Research Areas

Learn more about how we do research

We maintain a portfolio of research projects, providing individuals and teams the freedom to emphasize specific types of work