0 54

Cited 0 times in

Visual modalities-based multimodal fusion for surgical phase recognition

DC Field Value Language
dc.contributor.author박성현-
dc.contributor.author형우진-
dc.date.accessioned2024-03-22T06:19:47Z-
dc.date.available2024-03-22T06:19:47Z-
dc.date.issued2023-09-
dc.identifier.issn0010-4825-
dc.identifier.urihttps://ir.ymlib.yonsei.ac.kr/handle/22282913/198527-
dc.description.abstractSurgical workflow analysis is essential to help optimize surgery by encouraging efficient communication and the use of resources. However, the performance of phase recognition is limited by the use of information related to the presence of surgical instruments. To address the problem, we propose visual modality-based multimodal fusion for surgical phase recognition to overcome the limited diversity of information such as the presence of instruments. Using the proposed methods, we extracted a visual kinematics-based index related to using instruments, such as movement and their interrelations during surgery. In addition, we improved recognition performance using an effective convolutional neural network (CNN)-based fusion method for visual features and a visual kinematics-based index (VKI). The visual kinematics-based index improves the understanding of a surgical procedure since information is related to instrument interaction. Furthermore, these indices can be extracted in any environment, such as laparoscopic surgery, and help obtain complementary information for system kinematics log errors. The proposed methodology was applied to two multimodal datasets, a virtual reality (VR) simulator-based dataset (PETRAW) and a private distal gastrectomy surgery dataset, to verify that it can help improve recognition performance in clinical environments. We also explored the influence of a visual kinematics-based index to recognize each surgical workflow by the instrument's existence and the instrument's trajectory. Through the experimental results of a distal gastrectomy video dataset, we validated the effectiveness of our proposed fusion approach in surgical phase recognition. The relatively simple yet index-incorporated fusion we propose can yield significant performance improvements over only CNN-based training and exhibits effective training results compared to fusion based on Transformers, which require a large amount of pre-trained data. © 2023 Elsevier Ltd-
dc.description.statementOfResponsibilityrestriction-
dc.languageEnglish-
dc.publisherElsevier-
dc.relation.isPartOfCOMPUTERS IN BIOLOGY AND MEDICINE-
dc.rightsCC BY-NC-ND 2.0 KR-
dc.titleVisual modalities-based multimodal fusion for surgical phase recognition-
dc.typeArticle-
dc.contributor.collegeCollege of Medicine (의과대학)-
dc.contributor.departmentDept. of Surgery (외과학교실)-
dc.contributor.googleauthorBogyu Park-
dc.contributor.googleauthorHyeongyu Chi-
dc.contributor.googleauthorBokyung Park-
dc.contributor.googleauthorJiwon Lee-
dc.contributor.googleauthorHye Su Jin-
dc.contributor.googleauthorSunghyun Park-
dc.contributor.googleauthorWoo Jin Hyung-
dc.contributor.googleauthorMin-Kook Choi-
dc.identifier.doi10.1016/j.compbiomed.2023.107453-
dc.contributor.localIdA06210-
dc.contributor.localIdA04382-
dc.relation.journalcodeJ00638-
dc.identifier.eissn1879-0534-
dc.identifier.pmid37774560-
dc.identifier.urlhttps://www.sciencedirect.com/science/article/pii/S0010482523009186-
dc.subject.keywordMultimodal learning-
dc.subject.keywordSurgical phase recognition-
dc.subject.keywordSurgical workflow-
dc.subject.keywordVisual kinematics-based index-
dc.contributor.alternativeNamePark, Sung Hyun-
dc.contributor.affiliatedAuthor박성현-
dc.contributor.affiliatedAuthor형우진-
dc.citation.volume166-
dc.citation.startPage107453-
dc.identifier.bibliographicCitationCOMPUTERS IN BIOLOGY AND MEDICINE, Vol.166 : 107453, 2023-09-
Appears in Collections:
1. College of Medicine (의과대학) > Dept. of Surgery (외과학교실) > 1. Journal Papers

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.