dc.contributor.author | Ozturk, Saban | |
dc.contributor.author | Celik, Emin | |
dc.contributor.author | Cukur, Tolga | |
dc.date.accessioned | 2024-03-12T19:29:13Z | |
dc.date.available | 2024-03-12T19:29:13Z | |
dc.date.issued | 2023 | |
dc.identifier.issn | 0020-0255 | |
dc.identifier.issn | 1872-6291 | |
dc.identifier.uri | https://doi.org/10.1016/j.ins.2023.118938 | |
dc.identifier.uri | https://hdl.handle.net/20.500.12450/2232 | |
dc.description.abstract | The increasing utilization of medical imaging technology with digital storage capabilities has facilitated the compilation of large-scale data repositories. Fast access to image samples with similar appearance to suspected cases in these repositories can help establish a consulting system for healthcare professionals, and improve diagnostic procedures while minimizing processing delays. However, manual querying of large repositories is labor intensive. Content-based image retrieval (CBIR) offers an automated solution based on quantitative assessment of image simi-larity based on image features in a latent space. Since conventional methods based on hand-crafted features typically show poor generalization performance, learning-based CBIR methods have received attention recently. A common framework in this domain involves classifier-guided models that are trained to detect different image classes. Similarity assessments are then per-formed on the features captured by the intermediate stages of the trained models. While classifier -guided methods are powerful in inter-class discrimination, they are suboptimally sensitive to within-class differences in image features. An alternative framework instead performs task -agnostic training to learn an embedding space that enforces the representational discriminabil-ity of images. Within this representational-learning framework, a powerful method is triplet-wise learning that addresses the deficiencies of point-wise and pair-wise learning in characterizing the similarity relationships between image classes. However, the traditional triplet loss enforces separation between only a subset of image samples within the triplet via a manually-set constant margin value, so it can lead to suboptimal segregation of opponent classes and limited general-ization performance. To address these limitations, we introduce a triplet-learning method for automated querying of medical image repositories based on a novel Opponent Class Adaptive Margin (OCAM) loss. To maintain optimally discriminative representations, OCAM considers relationships among all image pairs within the triplet and utilizes an adaptive margin value that is automatically selected per dataset and during the course of training iterations. CBIR performance of OCAM is compared against state-of-the-art loss functions for representational learning on three public databases (gastrointestinal disease, skin lesion, lung disease). On average, OCAM shows an mAP performance of 86.30% in the KVASIR dataset, 70.30% in the ISIC 2019 dataset, and 85.57% in the X-RAY dataset. Comprehensive experiments in each application domain demonstrate the superior performance of OCAM against competing triplet-wise methods at 1.52%, classifier -guided methods at 2.29%, and non-triplet representational-learning methods at 4.56%. | en_US |
dc.language.iso | eng | en_US |
dc.publisher | Elsevier Science Inc | en_US |
dc.relation.ispartof | Information Sciences | en_US |
dc.rights | info:eu-repo/semantics/openAccess | en_US |
dc.subject | CBIR | en_US |
dc.subject | Medical image retrieval | en_US |
dc.subject | Triplet | en_US |
dc.subject | Representational learning | en_US |
dc.subject | Hashing | en_US |
dc.title | Content-based medical image retrieval with opponent class adaptive margin loss | en_US |
dc.type | article | en_US |
dc.department | Amasya Üniversitesi | en_US |
dc.authorid | Çukur, Tolga/0000-0002-2296-851X | |
dc.authorid | OZTURK, Saban/0000-0003-2371-8173 | |
dc.identifier.volume | 637 | en_US |
dc.relation.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
dc.identifier.scopus | 2-s2.0-85153475334 | en_US |
dc.identifier.doi | 10.1016/j.ins.2023.118938 | |
dc.department-temp | [Ozturk, Saban; Celik, Emin; Cukur, Tolga] Bilkent Univ, Dept Elect & Elect Engn, TR-06800 Ankara, Turkiye; [Ozturk, Saban; Celik, Emin; Cukur, Tolga] Bilkent Univ, Natl Magnet Resonance Res Ctr, TR-06800 Ankara, Turkiye; [Ozturk, Saban] Amasya Univ, Dept Elect & Elect Engn, TR-05001 Amasya, Turkiye; [Cukur, Tolga] Bilkent Univ, Sabuncu Brain Res Ctr, Neurosci Program, TR-06800 Ankara, Turkiye | en_US |
dc.identifier.wos | WOS:000990675100001 | en_US |
dc.authorwosid | Çukur, Tolga/Z-5452-2019 | |