78 231

Cited 0 times in

Cited 10 times in

An Efficient Human Instance-Guided Framework for Video Action Recognition

DC Field Value Language
dc.date.accessioned2023-02-10T00:48:51Z-
dc.date.available2023-02-10T00:48:51Z-
dc.date.issued2021-12-
dc.identifier.urihttps://ir.ymlib.yonsei.ac.kr/handle/22282913/192386-
dc.description.abstractIn recent years, human action recognition has been studied by many computer vision researchers. Recent studies have attempted to use two-stream networks using appearance and motion features, but most of these approaches focused on clip-level video action recognition. In contrast to traditional methods which generally used entire images, we propose a new human instance-level video action recognition framework. In this framework, we represent the instance-level features using human boxes and keypoints, and our action region features are used as the inputs of the temporal action head network, which makes our framework more discriminative. We also propose novel temporal action head networks consisting of various modules, which reflect various temporal dynamics well. In the experiment, the proposed models achieve comparable performance with the state-of-the-art approaches on two challenging datasets. Furthermore, we evaluate the proposed features and networks to verify the effectiveness of them. Finally, we analyze the confusion matrix and visualize the recognized actions at human instance level when there are several people.-
dc.description.statementOfResponsibilityopen-
dc.languageEnglish-
dc.publisherMDPI-
dc.relation.isPartOfSENSORS-
dc.rightsCC BY-NC-ND 2.0 KR-
dc.subject.MESHHuman Activities*-
dc.subject.MESHHumans-
dc.subject.MESHMotion-
dc.subject.MESHNeural Networks, Computer*-
dc.subject.MESHRecognition, Psychology-
dc.subject.MESHVision, Ocular-
dc.titleAn Efficient Human Instance-Guided Framework for Video Action Recognition-
dc.typeArticle-
dc.contributor.collegeCollege of Medicine (의과대학)-
dc.contributor.departmentDept. of Radiology (영상의학교실)-
dc.contributor.googleauthorInwoong Lee-
dc.contributor.googleauthorDoyoung Kim-
dc.contributor.googleauthorDongyoon Wee-
dc.contributor.googleauthorSanghoon Lee-
dc.identifier.doi10.3390/s21248309-
dc.relation.journalcodeJ03219-
dc.identifier.eissn1424-8220-
dc.identifier.pmid34960404-
dc.subject.keywordconvolutional neural network-
dc.subject.keywordhuman action recognition-
dc.subject.keywordhuman detection-
dc.subject.keywordmultiple human tracking-
dc.subject.keywordtemporal sequence analysis-
dc.citation.volume21-
dc.citation.number24-
dc.citation.startPage8309-
dc.identifier.bibliographicCitationSENSORS, Vol.21(24) : 8309, 2021-12-
Appears in Collections:
1. College of Medicine (의과대학) > Dept. of Radiology (영상의학교실) > 1. Journal Papers

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.