Search
On-Site Program Calendar
Browse By Day
Browse By Time
Browse By Person
Browse By Room
Browse By Unit
Browse By Session Type
Search Tips
Change Preferences / Time Zone
Sign In
Bluesky
Threads
X (Twitter)
YouTube
This study aims to solve the class imbalance problem that occurs when applying machine learning technology to large-scale datasets by evaluating the performance of various resampling methods (synthetic minority class oversampling technique (SMOTE), adaptive synthetic sampling (ADASYN), SMOTE combined with edited nearest neighbors (SMOTE+ENN), and SMOTE combined with Tomek links (SMOTE+Tomek) with ensemble models such as Extra Trees, Random Forest, XGBoost, and CatBoost. The optimal combination was found to be SMOTE and CatBoost. Additionally, explainable AI (XAI) techniques, particularly Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP), were applied to identify key predictive variables, providing both local and global insights.