Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients
The introduction of pre-trained language models in natural language processing (NLP) based on deep learning and the availability of electronic health records (EHRs) presents a great opportunity to transfer the “knowledge” learned from data in the general domain to enable the analysis of unstructured...
Main Authors: | , , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Frontiers Media S.A.
2021-01-01
|
Series: | Frontiers in Psychiatry |
Subjects: | |
Online Access: | https://www.frontiersin.org/articles/10.3389/fpsyt.2020.533949/full |
id |
doaj-e5c7d71fd0664381b13b5c077cb91efc |
---|---|
record_format |
Article |
spelling |
doaj-e5c7d71fd0664381b13b5c077cb91efc2021-01-15T05:00:30ZengFrontiers Media S.A.Frontiers in Psychiatry1664-06402021-01-011110.3389/fpsyt.2020.533949533949Deep Learning-Based Natural Language Processing for Screening Psychiatric PatientsHong-Jie Dai0Hong-Jie Dai1Hong-Jie Dai2Chu-Hsien Su3You-Qian Lee4You-Chen Zhang5Chen-Kai Wang6Chian-Jue Kuo7Chian-Jue Kuo8Chi-Shin Wu9Intelligent System Laboratory, Department of Electrical Engineering, College of Electrical Engineering and Computer Science, National Kaohsiung University of Science and Technology, Kaohsiung, TaiwanSchool of Post-Baccalaureate Medicine, Kaohsiung Medical University, Kaohsiung, TaiwanNational Institute of Cancer Research, National Health Research Institutes, Tainan, TaiwanDepartment of Psychiatry, National Taiwan University Hospital, Taipei, TaiwanIntelligent System Laboratory, Department of Electrical Engineering, College of Electrical Engineering and Computer Science, National Kaohsiung University of Science and Technology, Kaohsiung, TaiwanIntelligent System Laboratory, Department of Electrical Engineering, College of Electrical Engineering and Computer Science, National Kaohsiung University of Science and Technology, Kaohsiung, TaiwanBig Data Laboratory, Chunghwa Telecom Laboratories, Taoyuan, TaiwanTaipei City Psychiatric Center, Taipei City Hospital, Taipei, TaiwanDepartment of Psychiatry, School of Medicine, College of Medicine, Taipei Medical University, Taipei, TaiwanDepartment of Psychiatry, National Taiwan University Hospital, Taipei, TaiwanThe introduction of pre-trained language models in natural language processing (NLP) based on deep learning and the availability of electronic health records (EHRs) presents a great opportunity to transfer the “knowledge” learned from data in the general domain to enable the analysis of unstructured textual data in clinical domains. This study explored the feasibility of applying NLP to a small EHR dataset to investigate the power of transfer learning to facilitate the process of patient screening in psychiatry. A total of 500 patients were randomly selected from a medical center database. Three annotators with clinical experience reviewed the notes to make diagnoses for major/minor depression, bipolar disorder, schizophrenia, and dementia to form a small and highly imbalanced corpus. Several state-of-the-art NLP methods based on deep learning along with pre-trained models based on shallow or deep transfer learning were adapted to develop models to classify the aforementioned diseases. We hypothesized that the models that rely on transferred knowledge would be expected to outperform the models learned from scratch. The experimental results demonstrated that the models with the pre-trained techniques outperformed the models without transferred knowledge by micro-avg. and macro-avg. F-scores of 0.11 and 0.28, respectively. Our results also suggested that the use of the feature dependency strategy to build multi-labeling models instead of problem transformation is superior considering its higher performance and simplicity in the training process.https://www.frontiersin.org/articles/10.3389/fpsyt.2020.533949/fulldeep learningnatural language processingtext classificationpatient screeningpsychiatric diagnoses |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Hong-Jie Dai Hong-Jie Dai Hong-Jie Dai Chu-Hsien Su You-Qian Lee You-Chen Zhang Chen-Kai Wang Chian-Jue Kuo Chian-Jue Kuo Chi-Shin Wu |
spellingShingle |
Hong-Jie Dai Hong-Jie Dai Hong-Jie Dai Chu-Hsien Su You-Qian Lee You-Chen Zhang Chen-Kai Wang Chian-Jue Kuo Chian-Jue Kuo Chi-Shin Wu Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients Frontiers in Psychiatry deep learning natural language processing text classification patient screening psychiatric diagnoses |
author_facet |
Hong-Jie Dai Hong-Jie Dai Hong-Jie Dai Chu-Hsien Su You-Qian Lee You-Chen Zhang Chen-Kai Wang Chian-Jue Kuo Chian-Jue Kuo Chi-Shin Wu |
author_sort |
Hong-Jie Dai |
title |
Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients |
title_short |
Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients |
title_full |
Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients |
title_fullStr |
Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients |
title_full_unstemmed |
Deep Learning-Based Natural Language Processing for Screening Psychiatric Patients |
title_sort |
deep learning-based natural language processing for screening psychiatric patients |
publisher |
Frontiers Media S.A. |
series |
Frontiers in Psychiatry |
issn |
1664-0640 |
publishDate |
2021-01-01 |
description |
The introduction of pre-trained language models in natural language processing (NLP) based on deep learning and the availability of electronic health records (EHRs) presents a great opportunity to transfer the “knowledge” learned from data in the general domain to enable the analysis of unstructured textual data in clinical domains. This study explored the feasibility of applying NLP to a small EHR dataset to investigate the power of transfer learning to facilitate the process of patient screening in psychiatry. A total of 500 patients were randomly selected from a medical center database. Three annotators with clinical experience reviewed the notes to make diagnoses for major/minor depression, bipolar disorder, schizophrenia, and dementia to form a small and highly imbalanced corpus. Several state-of-the-art NLP methods based on deep learning along with pre-trained models based on shallow or deep transfer learning were adapted to develop models to classify the aforementioned diseases. We hypothesized that the models that rely on transferred knowledge would be expected to outperform the models learned from scratch. The experimental results demonstrated that the models with the pre-trained techniques outperformed the models without transferred knowledge by micro-avg. and macro-avg. F-scores of 0.11 and 0.28, respectively. Our results also suggested that the use of the feature dependency strategy to build multi-labeling models instead of problem transformation is superior considering its higher performance and simplicity in the training process. |
topic |
deep learning natural language processing text classification patient screening psychiatric diagnoses |
url |
https://www.frontiersin.org/articles/10.3389/fpsyt.2020.533949/full |
work_keys_str_mv |
AT hongjiedai deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT hongjiedai deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT hongjiedai deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT chuhsiensu deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT youqianlee deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT youchenzhang deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT chenkaiwang deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT chianjuekuo deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT chianjuekuo deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients AT chishinwu deeplearningbasednaturallanguageprocessingforscreeningpsychiatricpatients |
_version_ |
1724337669859704832 |