13 206

Cited 1 times in

Information Extraction from Clinical Texts with Generative Pre-trained Transformer Models

DC Field Value Language
dc.contributor.author김남오-
dc.contributor.author김민수-
dc.date.accessioned2025-05-02T00:10:43Z-
dc.date.available2025-05-02T00:10:43Z-
dc.date.issued2025-02-
dc.identifier.urihttps://ir.ymlib.yonsei.ac.kr/handle/22282913/205303-
dc.description.abstractPurpose: Processing and analyzing clinical texts are challenging due to its unstructured nature. This study compared the performance of GPT (Generative Pre-trained Transformer)-3.5 and GPT-4 for extracting information from clinical text. Materials and Methods: Three types of clinical texts, containing patient characteristics, medical history, and clinical test results extracted from case reports in open-access journals were utilized as input. Simple prompts containing queries for information extraction were then applied to both models using the Greedy Approach as the decoding strategy. When GPT models underperformed in certain tasks, we applied alternative decoding strategies or incorporated prompts with task-specific definitions. The outputs generated by GPT models were evaluated as True or False to determine the accuracy of information extraction. Results: Clinical texts containing patient characteristics (60 texts), medical history (50 texts), and clinical test results (25 texts) were extracted from 60 case reports. GPT models could extract information accurately with simple prompts to extract straightforward information from clinical texts. Regarding sex, GPT-4 demonstrated a significantly higher accuracy rate (95%) compared to GPT-3.5 (70%). GPT-3.5 (78%) outperformed GPT-4 (57%) in extracting body mass index (BMI). Utilizing alternative decoding strategies to sex and BMI did not practically improve the performance of the two models. In GPT-4, the revised prompts, including definitions of each sex category or the BMI formula, rectified all incorrect responses regarding sex and BMI generated during the main workflow. Conclusion: GPT models could perform adequately with simple prompts for extracting straightforward information. For complex tasks, incorporating task-specific definitions into the prompts is a suitable strategy than relying solely on simple prompts. Therefore, researchers and clinicians should use their expertise to create effective prompts and monitor LLM outcomes when extracting complex information from clinical texts.-
dc.description.statementOfResponsibilityopen-
dc.languageEnglish-
dc.publisherIvyspring International Publisher-
dc.relation.isPartOfINTERNATIONAL JOURNAL OF MEDICAL SCIENCES-
dc.rightsCC BY-NC-ND 2.0 KR-
dc.subject.MESHData Mining* / methods-
dc.subject.MESHFemale-
dc.subject.MESHHumans-
dc.subject.MESHMale-
dc.subject.MESHNatural Language Processing-
dc.titleInformation Extraction from Clinical Texts with Generative Pre-trained Transformer Models-
dc.typeArticle-
dc.contributor.collegeCollege of Medicine (의과대학)-
dc.contributor.departmentDept. of Anesthesiology and Pain Medicine (마취통증의학교실)-
dc.contributor.googleauthorMin-Soo Kim-
dc.contributor.googleauthorPhilip Chung-
dc.contributor.googleauthorNima Aghaeepour-
dc.contributor.googleauthorNamo Kim-
dc.identifier.doi10.7150/ijms.103332-
dc.contributor.localIdA00356-
dc.contributor.localIdA00463-
dc.relation.journalcodeJ02917-
dc.identifier.eissn1449-1907-
dc.identifier.pmid40027192-
dc.subject.keywordAccess to Information-
dc.subject.keywordMedical Informatics.-
dc.subject.keywordMedical Records-
dc.subject.keywordNatural Language Processing-
dc.contributor.alternativeNameKim, Namo-
dc.contributor.affiliatedAuthor김남오-
dc.contributor.affiliatedAuthor김민수-
dc.citation.volume22-
dc.citation.number5-
dc.citation.startPage1015-
dc.citation.endPage1028-
dc.identifier.bibliographicCitationINTERNATIONAL JOURNAL OF MEDICAL SCIENCES, Vol.22(5) : 1015-1028, 2025-02-
Appears in Collections:
1. College of Medicine (의과대학) > Dept. of Anesthesiology and Pain Medicine (마취통증의학교실) > 1. Journal Papers

qrcode

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.