Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Learning Transferable Self-attentive Representations for Action Recognition in Untrimmed Videos with Weak Supervision (1902.07370v1)

Published 20 Feb 2019 in cs.CV

Abstract: Action recognition in videos has attracted a lot of attention in the past decade. In order to learn robust models, previous methods usually assume videos are trimmed as short sequences and require ground-truth annotations of each video frame/sequence, which is quite costly and time-consuming. In this paper, given only video-level annotations, we propose a novel weakly supervised framework to simultaneously locate action frames as well as recognize actions in untrimmed videos. Our proposed framework consists of two major components. First, for action frame localization, we take advantage of the self-attention mechanism to weight each frame, such that the influence of background frames can be effectively eliminated. Second, considering that there are trimmed videos publicly available and also they contain useful information to leverage, we present an additional module to transfer the knowledge from trimmed videos for improving the classification performance in untrimmed ones. Extensive experiments are conducted on two benchmark datasets (i.e., THUMOS14 and ActivityNet1.3), and experimental results clearly corroborate the efficacy of our method.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Xiao-Yu Zhang (23 papers)
  2. Haichao Shi (9 papers)
  3. Changsheng Li (37 papers)
  4. Kai Zheng (134 papers)
  5. Xiaobin Zhu (21 papers)
  6. Lixin Duan (51 papers)
Citations (58)