Psychiatry Investig Search

CLOSE


Psychiatry Investig > Volume 22(11); 2025 > Article
Lim, Kim, Cha, Lee, Choi, Kang, Park, Seo, and Lee: Optimizing Suicide Risk Prediction in Korea: A Comparison of Model Performance Using Resampling Methods and Machine Learning Algorithms

Abstract

Objective

Machine learning (ML) can assist in predicting suicide risk and identifying associated risk factors. Various resampling methods and algorithms must be applied to develop an ML prediction model with better performance. In this study, we developed an optimal Korean suicide prediction model by applying five ML algorithms, unsampled data, and two resampling methods.

Methods

In this study, data from the Korea National Health and Nutrition Examination Survey for 2017, 2019, and 2021 were integrated and analyzed to predict suicidal ideation in subjects aged ≥19 years. Logistic regression, random forest (RF), k-nearest neighbor, gradient boosting, and adaptive boosting were used as ML algorithms. Undersampling and oversampling are used as resampling methods to solve data imbalance problems.

Results

Among the study participants, 16,947 (95.14%) and 866 (4.86%) belonged to the control and suicidal ideation groups, respectively. Among the 15 ML models, the RF model exhibited excellent performance (sensitivity=0.781, area under the curve=0.870) in an algorithm trained with undersampled data.

Conclusion

Developing an optimized Korean suicide prediction model through additional validation based on the ML model developed in this study will help predict suicide risk factors caused by the interaction of individual, social, and environmental factors.

INTRODUCTION

Machine learning (ML) is a field of artificial intelligence (AI) in which software learns based on data, identifies patterns, and automates the creation of models for data analysis [1]. ML-based prediction models are becoming increasingly useful in mental healthcare. A previous study combined linguistic features and social media behavioral data to predict the occurrence of depression with 70% accuracy [2]. In a study targeting older individuals, a prediction model that integrated clinical variables, such as apathy, increased antidepressant resistance, and cognitive and magnetic resonance imaging data, predicted the diagnosis and treatment response of late-life depression [3,4]. Big data analysis combined with ML helps in the early detection and diagnosis of mental illness [5,6] and provides personalized treatment methods by predicting treatment responses [5-8]. Suicide occurs through a complex combination of personal, social, and environmental factors, with prevention being the only effective measure. Considering these characteristics, developing an ML prediction model will help detect and predict risk factors for suicide. Previous suicide-related studies have analyzed clinical and general population-based prospective and retrospective data using general statistical and sampling methods [9,10]. The development of an ML-based prediction model complements the limitations of previous studies. It is an efficient way to develop measures to reduce the high suicide rate, which is a major social issue in South Korea [5].
Representative risk factors for suicide include individual psychological factors [11-15], physical diseases [13,15,16], social and environmental factors [17,18], and lifestyle habits [19,20]. Thus far, prospective and retrospective studies have analyzed risk factors for suicide. Retrospective studies provide limited and insufficient information, because the clinical characteristics of risk factors for suicide depend on self-reports or medical records [21]. Prospective studies require a long follow-up period and consent from the subjects; therefore, it is difficult to study suicide deaths and the subjects are limited to suicide attempters [9,10,21]. ML presents a model based on a prospective prediction model free from the dependence on retrospective information [5]. Furthermore, ML research focuses on prediction (the explanatory power of the model) rather than inference (hypothesis testing), and simultaneously handles many predictors [5]. ML can be used to develop models that predict suicide risk by comprehensively considering various suicide-related factors.
Previous studies have used ML to predict suicide risk and identify risk factors. An ML model to predict suicide attempts among subjects with mood disorders was developed with an accuracy of 64.7%-72% [22]. The ML model was developed to identify suicide risk factors and predict suicide attempts by applying the random forest (RF) algorithm to sociodemographic data and psychiatric scales [23]. In a meta-analysis of 87 studies that used ML to develop a prediction model for suicidal behavior, the AI/ML method predicted suicidal behavior with >90% accuracy [24]. As suicide is due to a complex combination of social and environmental factors as well as individual characteristics [25], it is difficult to apply overseas ML models. Therefore, it is necessary to develop an ML prediction model that uses large-scale domestic data.
In a previous study, suicide risk was predicted using data from the Korea National Health and Nutrition Examination Survey (KNHANES). However, only one ML algorithm was used, making it impossible to compare the model performance with other ML algorithms [26,27]. Another study using KNHANES data used six ML algorithms, but only one data resampling method was used, making it impossible to compare the model performance with other resampling data [28]. In ML studies predicting suicide using population-based datasets, there is a class imbalance problem, in which there are fewer suicidal cases than non-suicidal cases [29]. Therefore, a method to resample the dataset is required to improve model performance [29]. The dataset could be adjusted through resampling techniques to address the issue of class imbalance in the data distribution. Resampling can be performed as follows: replicating data from the minority class, increasing the number of samples by creating virtual samples, or removing some data from the majority class. This study aimed to develop an optimal Korean suicide prediction model by applying various ML algorithms and data resampling using the KNHANES data.

METHODS

Data

The Korea Centers for Disease Control and Prevention has been administering the KNHANES since 1998 to assess Korean individuals’ general health and nutritional status. This study used data from the 7th (2016-2018) and the 8th (2019-2021) KNHANES [30-32]. Among the KNHANES data, those from 2017, 2019, and 2021 were investigated and analyzed to predict whether participants aged 19 years or older had suicidal ideation. The data used in this study were obtained from the KNHANES and Nutrition Examination Survey website (https://knhanes.kdca.go.kr/knhanes/main.do).

Subjects

This study included 17,813 adults aged ≥19 years. In the KNHANES, those who answered “yes” to the question “Have you ever seriously thought about suicide in the past year?” were classified into the suicidal ideation group, and those who answered “no” were classified into the non-suicidal ideation group.

Exploring risk factors for suicidal ideation

Statistical analysis

KNHANES data used in this study were extracted using two-stage stratified cluster sampling. In data analysis, the complex structure of the sampling procedure was incorporated using weights. A comparative analysis was performed between the suicidal and non-suicidal ideation groups. For continuous variables, weighted means (standard errors) were calculated, and the means between the groups were compared using an independent t-test. Categorical variables were compared by calculating the number of subjects and weighted proportions and applying the Rao-Scott chi-squared test. For statistical analysis, the PROC SURVEY procedure in SAS ver. 9.4 (SAS Institute) was used, and statistical significance was set at p<0.05.

ML analysis

Data processing and set assignment

Of the 17,813 participants, 866 (4.8%) belonged to the suicidal ideation group, representing an imbalanced class. Compared with a class that is infrequently present in the dataset, the prevalent class is called the majority class [33]. To solve this problem, the raw data were separated in a 7:3 ratio for training and testing, and undersampling and oversampling were performed using the training set. Undersampling uses majority classes (non-suicidal ideation group) by randomly selecting samples, whereas oversampling uses the synthetic minority oversampling technique (SMOTE). SMOTE is a novel approach to oversampling data in the case of the present imbalanced data [33]. To compare the test performances of the models trained with each dataset, the test set was split from the raw data, and the test performance was calculated using the same test data (Figure 1).

Input variables

We used the following variables to develop the ML algorithms: sociodemographic characteristics (age, sex, town, medical care, household income, educational status, housing type, marital status, and job position), health behaviors (subjective health status, stress awareness, depressive mood: depressive mood was determined based on whether they have depressed continuously for more than 2 weeks in past year, weight change in the past 1 year, walking, smoking, and drinking), and presence of chronic diseases (hypertension, dyslipidemia, stroke, osteoarthritis, rheumatic arthritis, osteoporosis, diabetes mellitus, renal failure, and liver cirrhosis).

Feature selection

Logistic regression (LR), RF, k-nearest neighbors (kNN), gradient boosting (GB), and adaptive boosting (AB) were used to select features related to suicidal ideation. The feature importance was measured using permutation importance. Permutation importance is a test method that measures how much the model’s prediction error increases after permutation, which is a method of calculating feature by shuffling a random set of ordered items into a different order. This is interpreted as the feature with the highest value being more important, and the value decreasing as less important. This was implemented using the scikit-learn package (https://scikit-learn.org/stable/).

ML analysis

All ML models were implemented using Python and the scikit-learn package. An ML approach was used to identify the suicidal ideation group and the risk factors contributing to suicidal ideation. LR, RF, kNN, GB, and AB algorithms were used in the study. LR is a regularization technique that minimizes the error between predictions and actual observations by imposing a penalty on variables that contribute little to the results. This study uses the ridge (L2) and lasso (L1) penalty terms. The L2 penalty reduces the coefficients of low-contribution variables to approximately zero by adding a penalty corresponding to the square of the coefficient. The L1 penalty adds a penalty corresponding to the sum of the absolute values, making the coefficients of the variables with low contributions zero [34]. RF is an ensemble algorithm consisting of multiple decision trees. A decision tree creates a model with a tree structure according to decision rules that determine the direction of branching from the features (variables). The impurities in each feature were measured, and the feature with the greatest decrease was selected. The RF algorithm is an extension of the bootstrap aggregation (bagging) method for ensembling multiple independent decision trees using bagging and characteristic randomness [35]. kNN is a method for grouping data points for classification or prediction based on similarity measurement techniques. Among the kNNs measured with the distance function that measures similarity, this algorithm assigns the most similar class and classifies it according to the majority vote of the neighbors [36]. GB and AB are ML methods based on boosting. Unlike bagging, boosting is a sequential learning method that sequentially learns several simple tree models, such as weak learners. This method assigns weights to the error data of the first learned model so that the second learner can perform better classification. This was repeated sequentially to provide a model with a strong performance. The difference between GB and AB is the method of applying weights. GB assigns weights using a gradient descent, whereas AB assigns more weights to incorrect samples [37].

Performance evaluation

Each model was evaluated using a test set. The performance of each model was evaluated using the F1 score, precision, sensitivity, specificity, and area under the curve (AUC) (Supplementary Table 1). Positive results correspond to the suicidal ideation group. The sensitivity is the ratio of people predicted to be positive by the model to those who are actually positive (true positive [TP]/positive [P]). The specificity is the ratio of people predicted to be negative by the model to those who are actually negative (true negative [TN]/negative [N]). Precision is the ratio of the actual positives to those predicted by the model to be positive (TP/predicted positive [PP]). The F1 score is the harmonic mean of precision and sensitivity and is an indicator that evaluates model performance when the data target is unbalanced. The AUC was calculated using a receiver operating characteristic (ROC) curve. An AUC value close to 0 indicates poor model prediction, whereas a value close to 1 indicates better model prediction [38].
All participants provided written informed consent to participate in the KNHANES. The Institutional Review Board of the Gyeongsang National University Changwon Hospital in Korea approved the use of publicly available data for statistical analysis (IRB No. 2024-01-020).

RESULTS

Characteristics of the participants

Among the study participants, 16,947 were in the non-suicidal ideation group and 866 were in the suicidal ideation group. Walking was more common in the non-suicidal ideation group than in the suicidal ideation group. Old age, low household income, low educational status, single detached house or tenement, separated or unmarried status, unemployment, medical care, poor subjective health status, hypertension, dyslipidemia, stroke, osteoarthritis, diabetes mellitus, stress awareness, depressive mood, weight change in the past 1 year, current smoking, and high-risk drinking were more frequent in the suicidal ideation group than in the non-suicidal ideation group (Table 1).

Selected features from the ML models

The features selected using the unsampled data are listed in Supplementary Table 2. Each algorithm adopted 10-40 features to construct a predictive model. The AB algorithm selected the lowest number (n=10), whereas the RF algorithm selected the highest (n=40). Supplementary Table 3 summarizes the features selected using undersampling data. The AB algorithm selected the least number (n=17) and RF selected the greatest number (n=40). Supplementary Table 4 summarizes the results of variable selection using the oversampled data. The AB algorithm selected the lowest number (n=31), whereas the LR, RF, and GB algorithms selected the highest number (n=39). Several common variables associated with suicidal ideation were identified by using the predictive model developed in this study. These included age, self-reported health status, stress awareness, depressive mood, weight changes over the past year, current smoking habits, being in the lowest household income quartile, having at least a university degree, and being a permanent employee.

Predictive performance of the ML models

The performance metrics of the models for predicting suicidal ideation using unsampled, undersampled, and oversampled data are presented in Table 2. For unsampled data, GB had the highest AUC (0.876), lowest sensitivity (0.092), and a specificity of 0.995. For the undersampled data, RF had the highest AUC (0.870) and sensitivity (0.781). For the oversampled data, GB had the highest AUC (0.872), sensitivity of 0.312, and specificity of 0.973. Figure 2 shows the ROC curves of LR, RF, kNN, GB, and AB for predicting suicidal ideation. The ROC curve using unsampled data showed that GB had the highest AUC (0.876), followed by AB (0.872). The ROC curve with under-sampled data showed that RF had the highest AUC (0.870), followed by LR (0.869) and GB (0.869). The ROC curve with oversampled data showed that GB had the highest AUC (0.872), followed by AB (0.867).
Figure 3 illustrates the confusion matrices when the five models with unsampled, undersampled, and oversampled data were applied to the test data. Among the ML models trained with unsampled data, the AB model had the highest sensitivity (55/260=0.212). Among the ML models trained with the undersampled data, the RF model had the highest sensitivity (203/260=0.781), whereas with the oversampled data, the LR model had the highest sensitivity (128/260=0.492). According to the comprehensive performance comparison, the RF model performed best in predicting suicidal ideation using undersampled data (sensitivity=0.781, AUC=0.870).

DISCUSSION

In this study, data representing adults from the general Korean population were analyzed using five ML algorithms and three data sampling methods. Fifteen prediction models are compared to develop an optimal suicide prediction model. As the purpose of this study was to predict suicidal ideation in adults in the general population, it was necessary to develop a model with high sensitivity and AUC. Among the algorithms trained using undersampling data, the RF model, with 87% accuracy (AUC=0.870, 95% confidence interval: 0.841-0.898) and a sensitivity of 0.781, was able to classify the suicidal ideation and non-suicidal ideation groups better than the other models. Previous studies that developed models to predict the risk of suicidal ideation using ML suggested that many input variables or various ML-based algorithms are required to develop a better-performing model [28,39]. Our study complements the limitations of data sampling methods and ML algorithms used in previous suicide-related studies. First, we developed a prediction model with better performance than the existing models (accuracy: 71.9% to 81.0%) by applying five ML algorithms using a nationally representative dataset containing various variables related to personal, social, and environmental factors. Second, we compared the performances of the prediction models by applying various data resampling methods, such as undersampling and oversampling.
In previous studies that predicted suicidal ideators and suicide attempters using KNHANES data, one ML algorithm and one data resampling method were used [26,27]. Previous studies used only one ML algorithm (RF), but we used five ML algorithms and compared their performances with those of other ML algorithms. In our study, the GB algorithm trained with oversampled and unsampled data showed better AUC (0.872 and 0.876, respectively) than the other models; however, its sensitivity (0.312 and 0.092, respectively) was relatively low. When we compared the model performances of various ML algorithms, the one with the best performance was the RF model using undersampled data (sensitivity=0.781, AUC=0.870).
This study used two types of data resampling (undersampling and oversampling), whereas previous ML studies used only one type of resampling data [26,27]. Among all participants, there was a significant difference in the numbers between the suicidal ideation group (n=866) and the non-suicidal ideation group (n=16,947). When such imbalanced data are used directly, they tend to be biased toward the majority class. Therefore, data resampling was necessary. 29,40 Previous studies that utilized ML employed a type of oversampling approach, SMOTE, to attenuate the class imbalance problem [27] or the undersampling method [26]. However, despite the data resampling of the prediction model, there are limitations in the evaluation of actual data affected by a biased class ratio [27]. Similarly, the prediction model in a previous study, trained with the RF algorithm by undersampling, had a positive predictive value (PPV) of 0.787 in the test set. However, when applied to the entire population, PPV was as low as 0.462 [26]. Because the previous study used resampled data as a test set, it is possible that the PPV was high and did not reflect the actual phenomenon. In our study, the test set was split from the raw data, and the resampled data were applied only to the training set to develop a model that could evaluate the actual data.
In a previous study predicting suicidal ideation, six ML algorithms were used. However, only one data resampling (oversampling) method was used [28]. It was possible that the performance of the ML algorithm improved due to the lower prevalence of the minority class in the test set (AUC=0.794 to 0.877) [28]. In our study, the performance of the ML algorithm with unsampled data showed a high AUC (0.876) despite a low sensitivity (0.092). Therefore, various data resampling methods were applied, and the performance of the prediction model was evaluated using the sensitivity and AUC. In this study, the sensitivity and AUC of the prediction model trained with the RF algorithm using unsampled, undersampled, and oversampled data were as follows: AUC=0.731, sensitivity=0.173; AUC=0.870, sensitivity=0.781; AUC=0.840, sensitivity=0.162. As a result of this study, the data resampling method used to improve the performance of the suicide prediction model was undersampling.
This study has several limitations. First, we used data from the KNHANES on suicidal ideation and psychological status with simple questions, which may have affected the model’s performance. Second, because we developed an ML model using cross-sectional data, it was difficult to determine whether similar results could be obtained using longitudinal data. Third, additional analyses are required to compare the performance of the prediction models with those of other ML algorithms, such as artificial neural networks and support vector machines. Despite these limitations, this study is the first to develop an ML model for predicting suicidal ideation in Korean adults by comparing the performances of five ML algorithms and two data resampling methods. The prediction model developed in this study will help predict complex suicide risk factors arising from the interaction of individual, social, and environmental factors and develop interventions to improve them. The results of this study are significant for a sustainable Korean suicidal ideation prediction model that can be developed through ML using integrated data collected annually.

Supplementary Materials

The Supplement is available with this article at https://doi.org/10.30773/pi.2025.0187.
Supplementary Table 1.
Measuring performance
pi-2025-0187-Supplementary-Table-1.pdf
Supplementary Table 2.
Features associated with suicidal ideation selected using machine-learning algorithms that use unsampled data
pi-2025-0187-Supplementary-Table-2.pdf
Supplementary Table 3.
Features associated with suicidal ideation selected using machine-learning algorithms that use undersampled data
pi-2025-0187-Supplementary-Table-3.pdf
Supplementary Table 4.
Features associated with suicidal ideation selected using machine-learning algorithms that use oversampled data
pi-2025-0187-Supplementary-Table-4.pdf

Notes

Availability of Data and Material

The data are available at https://knhanes.kdca.go.kr/knhanes/main.do.

Conflicts of Interest

The authors have no potential conflicts of interest to disclose.

Author Contributions

Conceptualization: Eunji Lim, Dongyun Lee. Data curation: Sung Hyo Seo. Formal analysis: Eunji Lim, Nuree Kang. Funding acquisition: Dongyun Lee. Investigation: Eunji Lim, Sung Hyo Seo, Dongyun Lee. Methodology: Nuree Kang, Soyoung Park. Project administration: Dongyun Lee. Resources: Jae-Won Choi, Dongyun Lee. Supervision: Sung Hyo Seo, Dongyun Lee. Visualization: Eunji Lim, Sung Hyo Seo. Writing— original draft: Eunji Lim. Writing—review & editing: Bong-Jo Kim, Boseok Cha, So-Jin Lee, Sung Hyo Seo, Dongyun Lee.

Funding Statement

This work was supported by the Biomedical Research Institute Fund (GNUCHBRIF-2025-0005) from Gyeongsang National University Hospital in 2025 and a research grant from Gyeongsang National University in 2023.

Acknowledgments

The authors express their sincere gratitude to all the participants for their participation in this study.

Figure 1.
Development of the prediction model. SMOTE, synthetic minority over-sampling technique.
pi-2025-0187f1.jpg
Figure 2.
ROC curves for five machine-learning models in identifying suicidal ideation. A: Unsampling data. B: Undersampling data. C: Oversampling data. ROC, receiver operating characteristic; AUC, area under the curve.
pi-2025-0187f2.jpg
Figure 3.
Confusion matrix of five machine-learning models. A: Unsampling data. B: Undersampling data. C: Oversampling data. LR, logistic regression; RF, random forest; kNN, k-nearest neighbors; GB, gradient boosting; AB, adaptive boosting.
pi-2025-0187f3.jpg
Table 1.
Baseline characteristics of the participants
Total subjects (N=17,813) Non-suicidal ideation (N=16,947) Suicidal ideation (N=866) p
Sex <0.001
 Male 7,898 (49.9) 7,553 (50.2) 345 (42.4)
 Female 9,915 (50.1) 9,394 (49.8) 521 (57.6)
Age, year 47.8±0.3 47.7±0.2 50.3±0.9 0.002
Household income, quartile <0.001
 Lowest 3,486 (14.7) 3,134 (13.8) 352 (33.2)
 2nd 4,401 (23.9) 4,182 (23.8) 219 (25.5)
 3rd 4,721 (28.9) 4,568 (29.2) 153 (21.7)
 Highest 5,205 (32.6) 5,063 (33.2) 142 (19.5)
Educational statu <0.001
 ≤Elementary school 4,154 (17.0) 3,807 (16.3) 347 (31.1)
 Middle school 1,677 (7.8) 1,578 (7.7) 99 (10.1)
 High school 5,523 (33.6) 5,261 (33.5) 262 (34.8)
 ≥University graduate 6,459 (41.6) 6,301 (42.5) 158 (24.1)
Town, urban 3,576 (15.8) 3,374 (15.7) 202 (18.8) 0.115
Housing type <0.001
 Single detached house 6,079 (29.8) 5,710 (29.4) 369 (38.3)
 Apartment 9,495 (57.2) 9,118 (57.7) 377 (46.8)
 Tenement 1,752 (10.3) 1,651 (10.2) 101 (12.8)
 Others 487 (2.6) 468 (2.7) 19 (2.1)
Marital status <0.001
 Married, live together 12,040 (65.1) 11,581 (65.8) 459 (49.8)
 Married, separated 2,637 (10.9) 2,406 (10.5) 231 (20.8)
 Not married 3,136 (24.0) 2,960 (23.8) 176 (29.4)
Job position <0.001
 Permanent worker 3,571 (24.0) 3,500 (24.7) 71 (9.1)
 Temporary worker 3,872 (22.0) 3,687 (22.0) 185 (22.0)
 Self-employed 2,427 (13.6) 2,344 (13.8) 83 (9.4)
 Others 7,943 (40.4) 7,416 (39.5) 527 (59.5)
Medical care 821 (3.8) 698 (3.4) 123 (12.6) <0.001
Subjective health status <0.001
 Good 12,715 (69.6) 4,998 (31.3) 100 (12.1)
 Bad 5,098 (30.4) 11,949 (68.7) 766 (87.9)
Hypertension 4,391 (19.3) 4,120 (19.0) 271 (24.5) <0.001
Dyslipidemia 3,089 (14.4) 2,901 (14.2) 188 (17.8) 0.014
Stroke 303 (1.3) 274 (1.2) 29 (3.4) <0.001
OA 1,875 (7.6) 1,719 (7.3) 156 (13.2) <0.001
RA 271 (1.2) 248 (1.2) 23 (1.5) 0.258
Osteoporosis 1,197 (4.4) 1,086 (4.2) 111 (8.9) <0.001
DM 1,793 (8.1) 1,663 (7.9) 130 (13.8) <0.001
Renal failure 79 (0.4) 72 (0.4) 7 (0.8) 0.076
Liver cirrhosis 49 (0.2) 44 (0.2) 5 (0.5) 0.074
Stress awareness 4,723 (28.1) 4,141 (26.1) 582 (69.4) <0.001
Depressive mood 2,089 (11.1) 1,508 (8.5) 581 (67.3) <0.001
Subjective body type recognition 7,879 (45.2) 7,471 (45.1) 408 (47.5) 0.242
Weight change in past 1 year 6,674 (39.8) 6,251 (39.2) 423 (52.4) <0.001
Walking 6,708 (38.6) 6,434 (38.9) 274 (32.5) 0.005
Current smoker 3,033 (19.8) 2,812 (19.4) 221 (29.6) <0.001
High risk drinking 1,989 (12.7) 1,860 (12.5) 129 (17.3) 0.001

Data are presented as mean±standard deviation or number (%). OA, osteoarthritis; RA, rheumatic arthritis; DM, diabetes mellitus.

Table 2.
Performance of the prediction model for suicidal ideation using machine-learning algorithms trained on unsampling data, undersampling data, oversampling data
Data resampling Models F1 score Precision Sensitivity Specificity AUC (95% CI)
Unsampling LR 0.156 0.511 0.092 0.995 0.868 (0.840-0.896)
RF 0.243 0.405 0.173 0.987 0.731 (0.696-0.767)
KNN 0.157 0.188 0.135 0.970 0.552 (0.516-0.589)
GB 0.156 0.511 0.092 0.995 0.876 (0.848-0.903)
AB 0.294 0.482 0.212 0.988 0.872 (0.844-0.900)
Under sampling LR 0.316 0.201 0.738 0.850 0.869 (0.841-0.897)
RF 0.301 0.187 0.781 0.826 0.870 (0.841-0.898)
KNN 0.208 0.162 0.292 0.923 0.714 (0.678-0.750)
GB 0.313 0.198 0.742 0.846 0.869 (0.841-0.897)
AB 0.314 0.200 0.735 0.849 0.859 (0.830-0.888)
Over sampling LR 0.346 0.267 0.492 0.931 0.849 (0.819-0.878)
RF 0.240 0.467 0.162 0.991 0.840 (0.810-0.871)
KNN 0.169 0.173 0.165 0.960 0.635 (0.598-0.672)
GB 0.338 0.370 0.312 0.973 0.872 (0.845-0.900)
AB 0.378 0.393 0.365 0.971 0.867 (0.838-0.895)

AUC, area under the receiver operating characteristic curve; CI, confidence interval; LR, logistic regression; RF, random forest; kNN, k-nearest neighbors; GB, gradient boosting; AB, adaptive boosting.

REFERENCES

1. Janiesch C, Zschech P, Heinrich K. Machine learning and deep learning. Electron Mark 2021;31:685-695.
crossref pdf
2. De Choudhury M, Gamon M, Counts S, Horvitz E. Predicting depression via social media. Proc Int AAAI Conf Web Soc Media 2013;7:128-137.
crossref pdf
3. Grzenda A, Speier W, Siddarth P, Pant A, Krause-Sorio B, Narr K, et al. Machine learning prediction of treatment outcome in late-life depression. Front Psychiatry 2021;12:738494
crossref pmid pmc
4. Patel MJ, Andreescu C, Price JC, Edelman KL, Reynolds CF 3rd, Aizenstein HJ. Machine learning approaches for integrating clinical and imaging features in late-life depression classification and response prediction. Int J Geriatr Psychiatry 2015;30:1056-1067.
crossref pmid pmc
5. Chekroud AM, Bondar J, Delgadillo J, Doherty G, Wasil A, Fokkema M, et al. The promise of machine learning in predicting treatment outcomes in psychiatry. World Psychiatry 2021;20:154-170.
crossref pmid pmc pdf
6. Squires M, Tao X, Elangovan S, Gururajan R, Zhou X, Acharya UR, et al. Deep learning and machine learning in psychiatry: a survey of current progress in depression detection, diagnosis and treatment. Brain Inform 2023;10:10
crossref pmid pmc pdf
7. Nielsen AN, Barch DM, Petersen SE, Schlaggar BL, Greene DJ. Machine learning with neuroimaging: evaluating its applications in psychiatry. Biol Psychiatry Cogn Neurosci Neuroimaging 2020;5:791-798.
crossref pmid pmc
8. Tai AMY, Albuquerque A, Carmona NE, Subramanieapillai M, Cha DS, Sheko M, et al. Machine learning and big data: implications for disease modeling and therapeutic discovery in psychiatry. Artif Intell Med 2019;99:101704
crossref pmid
9. Li X, Mu F, Liu D, Zhu J, Yue S, Liu M, et al. Predictors of suicidal ideation, suicide attempt and suicide death among people with major depressive disorder: a systematic review and meta-analysis of cohort studies. J Affect Disord 2022;302:332-351.
crossref pmid
10. Wenzel A, Berchick ER, Tenhave T, Halberstadt S, Brown GK, Beck AT. Predictors of suicide relative to other deaths in patients with suicide attempts and suicide ideation: a 30-year prospective study. J Affect Disord 2011;132:375-382.
crossref pmid
11. Bickford D, Morin RT, Nelson JC, Mackin RS. Determinants of suicide-related ideation in late life depression: associations with perceived stress. Clin Gerontol 2020;43:37-45.
crossref pmid pmc
12. Goodwin R, Olfson M. Self-perception of poor health and suicidal ideation in medical patients. Psychol Med 2002;32:1293-1299.
crossref pmid
13. Joshi P, Song HB, Lee SA. Association of chronic disease prevalence and quality of life with suicide-related ideation and suicide attempt among Korean adults. Indian J Psychiatry 2017;59:352-358.
crossref pmid pmc
14. Park JI, Yang JC, Han C, Park TW, Chung SK. Suicidal ideation among Korean elderly: risk factors and population attributable fractions. Psychiatry 2016;79:262-281.
crossref pmid
15. Zhu J, Xu L, Sun L, Li J, Qin W, Ding G, et al. Chronic disease, disability, psychological distress and suicide ideation among rural elderly: results from a population survey in Shandong. Int J Environ Res Public Health 2018;15:1604
crossref pmid pmc
16. Song A, Koh EJ, Lee WY, Chang S, Lim J, Choi M, et al. Suicide risk of chronic diseases and comorbidities: a Korean case-control study. J Affect Disord 2024;349:431-437.
crossref
17. Schneider B, Grebner K, Schnabel A, Hampel H, Georgi K, Seidler A. Impact of employment status and work-related factors on risk of completed suicide: a case-control psychological autopsy study. Psychiatry Res 2011;190:265-270.
crossref pmid
18. Tassone VK, Duffy SF, Dunnett S, Boparai JK, Zuluaga Cuartas V, Jung H, et al. Decreased odds of depressive symptoms and suicidal ideation with higher education, depending on sex and employment status. PLoS One 2024;19:e0299817
crossref pmid pmc
19. Kim H, Jeon HJ, Bae JN, Cho MJ, Cho SJ, Lee H, et al. Association of body mass index with suicide behaviors, perceived stress, and life dissatisfaction in the Korean general population. Psychiatry Investig 2018;15:272-278.
crossref pmid pmc pdf
20. Poorolajal J, Darvishi N. Smoking and suicide: a meta-analysis. PLoS One 2016;11:e0156348
crossref pmid pmc
21. Euser AM, Zoccali C, Jager KJ, Dekker FW. Cohort studies: prospective versus retrospective. Nephron Clin Pract 2009;113:c214-c217.
crossref pmid pdf
22. Passos IC, Mwangi B, Cao B, Hamilton JE, Wu MJ, Zhang XY, et al. Identifying a clinical signature of suicidality among patients with mood disorders: a pilot study using a machine learning approach. J Affect Disord 2016;193:109-116.
crossref
23. Shen Y, Zhang W, Chan BSM, Zhang Y, Meng F, Kennon EA, et al. Detecting risk of suicide attempts among Chinese medical college students using a machine learning algorithm. J Affect Disord 2020;273:18-23.
crossref pmid
24. Bernert RA, Hilberg AM, Melia R, Kim JP, Shah NH, Abnousi F. Artificial intelligence and suicide prevention: a systematic review of machine learning investigations. Int J Environ Res Public Health 2020;17:5929
crossref pmid pmc
25. World Health Organization. National suicide prevention strategies [Internet] Available at: https://apps.who.int/iris/bitstream/handle/10665/279765/9789241515016-eng.pdf?ua=1. Accessed October 15, 2024.

26. Ryu S, Lee H, Lee DK, Park K. Use of a machine learning algorithm to predict individuals with suicide ideation in the general population. Psychiatry Investig 2018;15:1030-1036.
crossref pmid pmc pdf
27. Ryu S, Lee H, Lee DK, Kim SW, Kim CE. Detection of suicide attempters among suicide ideators using machine learning. Psychiatry Investig 2019;16:588-593.
crossref pmid pmc pdf
28. Oh B, Yun JY, Yeo EC, Kim DH, Kim J, Cho BJ. Prediction of suicidal ideation among Korean adults using machine learning: a cross-sectional study. Psychiatry Investig 2020;17:331-340.
crossref pmid pmc pdf
29. Li DC, Liu CW, Hu SC. A learning method for the class imbalance problem with medical data sets. Comput Biol Med 2010;40:509-518.
crossref pmid
30. Korea Disease Control and Prevention Agency. Korea national health and nutrition examination survey (KNHANES) [Internet] Available at: https://knhanes.kdca.go.kr/knhanes/postSendPage.do?url=/rawDataDwnld/rawDataDwnld.do&postparam=%7B%22menuId%22:%221003%22%7D. Accessed October 15, 2024.

31. Korea Disease Control and Prevention Agency. Korea national health and nutrition examination survey (KNHANES) [Internet] Available at: https://knhanes.kdca.go.kr/knhanes/postSendPage.do?url=/rawDataDwnld/rawDataDwnld.do&postparam=%7B%22menuId%22:%2210031001%22%7D. Accessed October 15, 2024.

32. Korea Disease Control and Prevention Agency. Korea national health and nutrition examination survey (KNHANES) [Internet] Available at: https://knhanes.kdca.go.kr/knhanes/postSendPage.do?url=/dataAnlsGd/utztnGd.do&postparam=%7B%22menuId%22:%2210031002%22%7D. Accessed October 15, 2024.

33. Pradipta GA, Wardoyo R, Musdholifah A, Sanjaya INH, Ismail M. SMOTE for handling imbalanced data problem: a review [Internet] Available at: http://doi.org/10.1109/ICIC54025.2021.9632912". Accessed October 15, 2024.

34. Bisong E. Logistic regression. In: Bisong E, editor. Building machine learning and deep learning models on Google Cloud platform: a comprehensive guide for beginners. Berkeley: Apress, 2019, p. 243-250.

35. Breiman L. Random forests. Mach Learn 2001;45:5-32.
crossref pdf
36. Kramer O. K-nearest neighbors. In: Kramer O, editor. Dimensionality reduction with unsupervised nearest neighbors. Berlin, Heidelberg: Springer, 2013, p. 13-23.

37. Natekin A, Knoll A. Gradient boosting machines, a tutorial. Front Neurorobot 2013;7:21
crossref pmid pmc
38. Saroj RK, Yadav PK, Singh R, Chilyabanyama ON. Machine learning algorithms for understanding the determinants of under-five mortality. BioData Min 2022;15:20
crossref pmid pmc pdf
39. Walsh CG, Ribeiro JD, Franklin JC. Predicting suicide attempts in adolescents with longitudinal clinical data and machine learning. J Child Psychol Psychiatry 2018;59:1261-1270.
crossref pmid pdf
40. Rendón E, Alejo R, Castorena C, Isidro-Ortega FJ, Granda-Gutiérrez EE. Data sampling methods to deal with the big data multi-class imbalance problem. Appl Sci 2020;10:1276
crossref


ABOUT
AUTHOR INFORMATION
ARTICLE CATEGORY

Browse all articles >

BROWSE ARTICLES
Editorial Office
#522, G-five Central Plaza, 27 Seochojungang-ro 24-gil, Seocho-gu, Seoul 06601, Korea
Tel: +82-2-537-6171  Fax: +82-2-537-6174    E-mail: psychiatryinvest@gmail.com                

Copyright © 2026 by Korean Neuropsychiatric Association.

Developed in M2PI

Close layer
prev next