상세 보기
- Lee, Go-Eun;
- Kim, Seon Ho;
- Cho, Jungchan;
- Choi, Sang Tae;
- Choi, Sang-Il
WEB OF SCIENCE
19SCOPUS
21초록
We propose a novel text-guided cross-position attention module which aims at applying a multi-modality of text and image to position attention in medical image segmentation. To match the dimension of the text feature to that of the image feature map, we multiply learnable parameters by text features and combine the multi-modal semantics via cross-attention. It allows a model to learn the dependency between various characteristics of text and image. Our proposed model demonstrates superior performance compared to other medical models using image-only data or image-text data. Furthermore, we utilize our module as a region of interest (RoI) generator to classify the inflammation of the sacroiliac joints. The RoIs obtained from the model contribute to improve the performance of classification models. © The Author(s), under exclusive license to Springer Nature Switzerland AG 2023.
키워드
- 제목
- Text-Guided Cross-Position Attention for Segmentation: Case of Medical Image
- 저자
- Lee, Go-Eun; Kim, Seon Ho; Cho, Jungchan; Choi, Sang Tae; Choi, Sang-Il
- 발행일
- 2023
- 유형
- Proceedings Paper
- 권
- 14224 LNCS
- 페이지
- 537 ~ 546