Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

HAKE: A Knowledge Engine Foundation for Human Activity Understanding (2202.06851v2)

Published 14 Feb 2022 in cs.CV, cs.AI, and cs.LG

Abstract: Human activity understanding is of widespread interest in artificial intelligence and spans diverse applications like health care and behavior analysis. Although there have been advances in deep learning, it remains challenging. The object recognition-like solutions usually try to map pixels to semantics directly, but activity patterns are much different from object patterns, thus hindering success. In this work, we propose a novel paradigm to reformulate this task in two stages: first mapping pixels to an intermediate space spanned by atomic activity primitives, then programming detected primitives with interpretable logic rules to infer semantics. To afford a representative primitive space, we build a knowledge base including 26+ M primitive labels and logic rules from human priors or automatic discovering. Our framework, the Human Activity Knowledge Engine (HAKE), exhibits superior generalization ability and performance upon canonical methods on challenging benchmarks. Code and data are available at http://hake-mvig.cn/.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (9)
  1. Yong-Lu Li (47 papers)
  2. Xinpeng Liu (19 papers)
  3. Xiaoqian Wu (8 papers)
  4. Yizhuo Li (21 papers)
  5. Zuoyu Qiu (2 papers)
  6. Liang Xu (117 papers)
  7. Yue Xu (79 papers)
  8. Hao-Shu Fang (38 papers)
  9. Cewu Lu (203 papers)
Citations (32)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com