상세 보기
- Seo, Seungmo;
- Youn, Jongsu;
- Jung, Seungjin;
- Kwak, Minji;
- Jin, Heegon;
- ... Choi, Jongwon;
- 외 1명
WEB OF SCIENCE
0SCOPUS
0초록
When the target domain is well defined, only a small portion of a deep neural network is often sufficient to achieve satisfactory performance. However, traditional model compression method typically requires access to the original dataset and additional labeled annotations, an approach that is both costly and often restricted by privacy concerns. In this study, we propose a novel slimmable adaptation framework designed for source-free unsupervised domain adaptation (SFUDA). Specifically, we introduce a block transformable network that can be slimmed into multiple sub-models without requiring any additional fine-tuning or retraining. Furthermore, we develop an online multi-path stabilization strategy that enhances the performance of these slimmed sub-models using only unlabeled target data. Comprehensive experiments across various scenarios confirm that the proposed method enables the slimmable model to adapt effectively to diverse domain shifts. Our approach surpasses existing unsupervised domain adaptation and model compression techniques that rely on labeled data.
키워드
- 제목
- Online Adaptive Slimmable Network for Source-Free Unsupervised Domain Adaptation
- 저자
- Seo, Seungmo; Youn, Jongsu; Jung, Seungjin; Kwak, Minji; Jin, Heegon; Jung, Hojoon; Choi, Jongwon
- 발행일
- 2026
- 유형
- Article
- 저널명
- IEEE Access
- 권
- 14
- 페이지
- 26459 ~ 26475