ReActNet: Temporal Localization of Repetitive Activities in Real-World Videos

10/14/2019
by   Giorgos Karvounas, et al.
31

We address the problem of temporal localization of repetitive activities in a video, i.e., the problem of identifying all segments of a video that contain some sort of repetitive or periodic motion. To do so, the proposed method represents a video by the matrix of pairwise frame distances. These distances are computed on frame representations obtained with a convolutional neural network. On top of this representation, we design, implement and evaluate ReActNet, a lightweight convolutional neural network that classifies a given frame as belonging (or not) to a repetitive video segment. An important property of the employed representation is that it can handle repetitive segments of arbitrary number and duration. Furthermore, the proposed training process requires a relatively small number of annotated videos. Our method raises several of the limiting assumptions of existing approaches regarding the contents of the video and the types of the observed repetitive activities. Experimental results on recent, publicly available datasets validate our design choices, verify the generalization potential of ReActNet and demonstrate its superior performance in comparison to the current state of the art.

READ FULL TEXT

page 1

page 3

page 4

page 7

page 8

research
05/17/2020

Detecting Forged Facial Videos using convolutional neural network

In this paper, we propose to detect forged videos, of faces, in online v...
research
10/18/2021

Graph Convolution Neural Network For Weakly Supervised Abnormality Localization In Long Capsule Endoscopy Videos

Temporal activity localization in long videos is an important problem. T...
research
09/14/2022

Real-world Video Anomaly Detection by Extracting Salient Features in Videos

We propose a lightweight and accurate method for detecting anomalies in ...
research
05/06/2019

Spatio-Temporal Action Localization in a Weakly Supervised Setting

Enabling computational systems with the ability to localize actions in v...
research
08/20/2019

ViSiL: Fine-grained Spatio-Temporal Video Similarity Learning

In this paper we introduce ViSiL, a Video Similarity Learning architectu...
research
06/15/2022

Structured Video Tokens @ Ego4D PNR Temporal Localization Challenge 2022

This technical report describes the SViT approach for the Ego4D Point of...
research
09/15/2019

Mining Minimal Map-Segments for Visual Place Classifiers

In visual place recognition (VPR), map segmentation (MS) is a preprocess...

Please sign up or login with your details

Forgot password? Click here to reset