상세 보기
- 송채린;
- 이나현;
- 곽일엽
WEB OF SCIENCE
0SCOPUS
0초록
While deep learning has achieved revolutionary success in image and natural language processing, traditional gradient boosting-based machine learning (ML) models still dominate in the biomedical domain for tabular data. This study systematically evaluates the performance and efficiency of three ML models (XGBoost, LightGBM, CatBoost) and four deep learning (DL) models on five public biomedical datasets, applying identical preprocessing and hyperparameter tuning. Experimental results show that for small to medium-sized datasets (under 10,000 samples), ML models consistently demonstrated superior performance and speed. On large-scale datasets (over 200,000 samples), DL models showed comparable performance but with significantly decreased efficiency as the number of features increased. In conclusion, gradient boosting-based ML models remain a robust choice for most biomedical tabular problems, while Transformer-based DL models may offer limited benefits only when applied to very large datasets with sufficient computational resources.
키워드
- 제목
- 의생물학 Tabular 데이터에서 딥러닝과 전통적 머신러닝의 성능 비교
- 제목 (타언어)
- Benchmarking Deep Learning vs. Traditional Machine Learning on Biomedical Tabular Data
- 저자
- 송채린; 이나현; 곽일엽
- 발행일
- 2025-10
- 유형
- Y
- 권
- 27
- 호
- 5
- 페이지
- 1501 ~ 1515