Skip to content

Implementation of Appearance and Relation Network (ARTNET) for the task of action recognition in video classification

Notifications You must be signed in to change notification settings

maryam-fatima/Implementation-of-ARTNet-in-python

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 

Repository files navigation

Implementation of Appearance and Relation Network (ARTNET) for the task of action recognition in video classification.

Spatiotemporal feature learning in videos is a fundamental problem in computer vision. Appearance-and-Relation Network(ARTNet), learns video representation in an end-to-end manner. ARTNets are constructed by stacking multiple generic building blocks, called as SMART, whose goal is to simultaneously model appearance and relation from RGB input in a separate and explicit manner. SMART blocks decouple the spatiotemporal learning module into an appearance branch for spatial modeling and a relation branch for temporal modeling. The appearance branch is implemented based on the linear combination of pixels or filter responses in each frame. While the relation branch is designed based on the multiplicative interactions between pixels or filter responses across multiple frames.

SMART Block

SMART block

Appearence and Relation Networks

ARTNets

Reference:

Wang, L., Li, W., Li, W., & Van Gool, L. (2018). Appearance-and-relation networks for video classification. In Proceedings of the IEEE conference on computer vision and pattern recognition (pp. 1430-1439).

About

Implementation of Appearance and Relation Network (ARTNET) for the task of action recognition in video classification

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published