Deep Disentangled Metric Learning
Citations

WEB OF SCIENCE

0
Citations

SCOPUS

1

초록

Proxy-based metric learning has enhanced semantic similarity with class representatives and exhibited noteworthy performance in deep metric learning tasks. While these methods alleviate computational demands by learning instance-to-class relationships rather than instance-to-instance relationships, they often limit features to be class-specific, thereby degrading generalization performance for unseen class. In this paper, we introduce a novel perspective called Disentangled Deep Metric Learning (DDML), grounded in the framework of information bottleneck, which applies class-agnostic regularization to existing DML methods. Unlike conventional NormSoftmax methods, which primarily emphasize distinct class-specific features, our DDML enables a diverse feature representation by seamlessly transitioning between class-specific features with the aid of class-agnostic features. It smooths decision boundaries, allowing unseen classes to have stable semantic representations in the embedding space. To achieve this, we learn disentangled representations of both class-specific and class-agnostic features in the context of DML. Our method easily integrates into existing proxy-based algorithms, consistently delivering improved performance. Copyright © 2025, Association for the Advancement of Artificia Intelligence (www.aaai.org). All rights reserved.

제목
Deep Disentangled Metric Learning
저자
Park, JinheePark, JisooNa, DagyeongKwon, Junseok
DOI
10.1609/aaai.v39i19.34184
발행일
2025-04
유형
Conference paper
저널명
Proceedings of the AAAI Conference on Artificial Intelligence
39
19
페이지
19830 ~ 19838