Self-Supervised Keypoint Discovery in Behavioral Videos

We propose a method for learning the posture and structure of agents from unlabelled behavioral videos. Starting from the observation that behaving agents are generally the main sources of movement in behavioral videos, our method, Behavioral Keypoint Discovery (B-KinD), uses an encoder-decoder arch...

Full description

Saved in:
Bibliographic Details
Published in:2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Vol. 2022; pp. 2161 - 2170
Main Authors: Sun, Jennifer J., Ryou, Serim, Goldshmid, Roni H., Weissbourd, Brandon, Dabiri, John O., Anderson, David J., Kennedy, Ann, Yue, Yisong, Perona, Pietro
Format: Conference Proceeding Journal Article
Language:English
Published: IEEE 01-06-2022
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We propose a method for learning the posture and structure of agents from unlabelled behavioral videos. Starting from the observation that behaving agents are generally the main sources of movement in behavioral videos, our method, Behavioral Keypoint Discovery (B-KinD), uses an encoder-decoder architecture with a geometric bottleneck to reconstruct the spatiotemporal difference between video frames. By focusing only on regions of movement, our approach works directly on input videos without requiring manual annotations. Experiments on a variety of agent types (mouse, fly, human, jellyfish, and trees) demonstrate the generality of our approach and reveal that our discovered keypoints represent semantically meaningful body parts, which achieve state-of-the-art performance on key-point regression among self-supervised methods. Additionally, B-KinD achieve comparable performance to supervised keypoints on downstream tasks, such as behavior classification, suggesting that our method can dramatically reduce model training costs vis-a-vis supervised methods.
Bibliography:Equal contribution.
Current affiliation: Samsung Advanced Institute of Technology
ISSN:1063-6919
2575-7075
DOI:10.1109/CVPR52688.2022.00221