• 综合
  • 标题
  • 关键词
  • 摘要
  • 学者
  • 期刊-刊名
  • 期刊-ISSN
  • 会议名称
搜索

作者:

Yu, Hongyang (Yu, Hongyang.) | Zhang, Xinfeng (Zhang, Xinfeng.) | Wang, Yaowei (Wang, Yaowei.) | Huang, Qingming (Huang, Qingming.) | Yin, Baocai (Yin, Baocai.)

收录:

EI Scopus SCIE

摘要:

Accident detection in surveillance or dashcam videos is a common task in the field of traffic accident analysis by using videos. However, as accidents occur sparsely and randomly in the real world, the data records are more scarce than the training data for standard detection tasks such as object detection or instance detection. Moreover, the limited and diverse accident data makes it more difficult to model the accident pattern for fine-grained accident detection tasks analyzing the accident in detail. Extra prior information should be introduced in the tasks such as the common vision feature which could offer relatively effective information for many vision tasks. The big model could generate the common vision feature by training on abundant data and consuming a lot of computing time and resources. Even though the accident video data is special, the big model could also extract common vision features. Thus, in this paper, we propose to apply knowledge distillation to fine-grained accident detection which analyzes the spatial temporal existence and severity for solving the issues of complex computing (distillation to the small model) and keeping good performance under limited accident data. Knowledge distillation could offer extra general vision feature information from the pre-trained big model. Common knowledge distillation guides the student network to learn the same representations from the teacher network by logit mimicking or feature imitation. However, single-level distillation could only focus on one aspect of mimicking classification logit or deep features. Multiple tasks with different focuses are required for fine-grained accident detection, such as multiple accident classification, temporal-spatial accident region detection, and accident severity estimation. Thus in this paper, multiple-level distillation is proposed for the different modules to generate the unified video feature concerning all the tasks in fine-grained accident detection analysis. The various experimental results on a fine-grained accident detection dataset which provides more detailed annotations of accidents demonstrate that our method could effectively model the video feature for multiple tasks.

关键词:

knowledge distillation Video accident detection multiple-level distillation fine-grained accident detection

作者机构:

  • [ 1 ] [Yu, Hongyang]Peng Cheng Lab, Shenzhen 518066, Peoples R China
  • [ 2 ] [Wang, Yaowei]Peng Cheng Lab, Shenzhen 518066, Peoples R China
  • [ 3 ] [Yin, Baocai]Peng Cheng Lab, Shenzhen 518066, Peoples R China
  • [ 4 ] [Zhang, Xinfeng]Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 100039, Peoples R China
  • [ 5 ] [Huang, Qingming]Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 100039, Peoples R China
  • [ 6 ] [Yin, Baocai]Beijing Univ Technol, Fac Informat Technol, Beijing 100124, Peoples R China

通讯作者信息:

  • [Zhang, Xinfeng]Univ Chinese Acad Sci, Sch Comp Sci & Technol, Beijing 100039, Peoples R China;;

查看成果更多字段

相关关键词:

来源 :

IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY

ISSN: 1051-8215

年份: 2024

期: 6

卷: 34

页码: 4445-4457

8 . 4 0 0

JCR@2022

被引次数:

WoS核心集被引频次:

SCOPUS被引频次: 3

ESI高被引论文在榜: 0 展开所有

万方被引频次:

中文被引频次:

近30日浏览量: 2

归属院系:

在线人数/总访问数:643/4958314
地址:北京工业大学图书馆(北京市朝阳区平乐园100号 邮编:100124) 联系我们:010-67392185
版权所有:北京工业大学图书馆 站点建设与维护:北京爱琴海乐之技术有限公司