Comparing Six Deep Learning Feature Extractors for CBIR Tasks

This article evaluates six deep-learning feature extractors for content-based image retrieval (CBIR), spanning both self-supervised and supervised approaches. It analyzes DINOv1, DINOv2, and DreamSim as ImageNet-pretrained self-supervised models, and contrasts them with SwinTransformer and two ResNet50 variants—one trained on RadImageNet and another on fractal geometry renderings. By extending earlier studies, the comparison highlights how backbone choice, training data, and pretraining strategies impact performance across medical and synthetic imaging tasks.


This content originally appeared on HackerNoon and was authored by Image Recognition

Abstract and 1. Introduction

  1. Materials and Methods

    2.1 Vector Database and Indexing

    2.2 Feature Extractors

    2.3 Dataset and Pre-processing

    2.4 Search and Retrieval

    2.5 Re-ranking retrieval and evaluation

  2. Evaluation and 3.1 Search and Retrieval

    3.2 Re-ranking

  3. Discussion

    4.1 Dataset and 4.2 Re-ranking

    4.3 Embeddings

    4.4 Volume-based, Region-based and Localized Retrieval and 4.5 Localization-ratio

  4. Conclusion, Acknowledgement, and References

2.2 Feature Extractors

We extend the analysis of Khun Jush et al. [2023] by adding two ResNet50 embeddings and evaluating the performance of six different slice embedding extractors for CBIR tasks. All the feature extractors are based on deep-learning-based models.

\ Table 1: Mapping of the original TS classes to 29 coarse anatomical regions.

\ Self-supervised Models: We employed three self-supervised models pre-trained on ImageNet [Deng et al., 2009]. DINOv1 [Caron et al., 2021], that demonstrated learning efficient image representations from unlabeled data using self-distillation. DINOv2 [Oquab et al., 2023], is built upon DINOv1 [Caron et al., 2021], and this model scales the pre-training process by combining an improved training dataset, patchwise objectives during training and introducing a new regularization technique, which gives rise to superior performance on segmentation tasks. DreamSim [Fu et al., 2023], built upon the foundation of DINOv1 [Caron et al., 2021], fine-tunes the model using synthetic data triplets specifically designed to be cognitively impenetrable with human judgments. For the self-supervised models, we used the best-performing backbone reported by the developers of the models.

\ Supervised Models: We included a SwinTransformer model [Liu et al., 2021] and a ResNet50 model [He et al., 2016] trained in a supervised manner using the RadImageNet dataset [Mei et al., 2022] that includes 5 million annotated 2D CT, MRI, and ultrasound images of musculoskeletal, neurologic, oncologic, gastrointestinal, endocrine, and pulmonary pathology. Furthermore, a ResNet50 model pre-trained on rendered images of fractal geometries was included based on [Kataoka et al., 2022]. These training images are formula-derived, non-natural, and do not require any human annotation.

\

:::info Authors:

(1) Farnaz Khun Jush, Bayer AG, Berlin, Germany (farnaz.khunjush@bayer.com);

(2) Steffen Vogler, Bayer AG, Berlin, Germany (steffen.vogler@bayer.com);

(3) Tuan Truong, Bayer AG, Berlin, Germany (tuan.truong@bayer.com);

(4) Matthias Lenga, Bayer AG, Berlin, Germany (matthias.lenga@bayer.com).

:::


:::info This paper is available on arxiv under CC BY 4.0 DEED license.

:::

\


This content originally appeared on HackerNoon and was authored by Image Recognition


Print Share Comment Cite Upload Translate Updates
APA

Image Recognition | Sciencx (2025-08-27T07:00:16+00:00) Comparing Six Deep Learning Feature Extractors for CBIR Tasks. Retrieved from https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/

MLA
" » Comparing Six Deep Learning Feature Extractors for CBIR Tasks." Image Recognition | Sciencx - Wednesday August 27, 2025, https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/
HARVARD
Image Recognition | Sciencx Wednesday August 27, 2025 » Comparing Six Deep Learning Feature Extractors for CBIR Tasks., viewed ,<https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/>
VANCOUVER
Image Recognition | Sciencx - » Comparing Six Deep Learning Feature Extractors for CBIR Tasks. [Internet]. [Accessed ]. Available from: https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/
CHICAGO
" » Comparing Six Deep Learning Feature Extractors for CBIR Tasks." Image Recognition | Sciencx - Accessed . https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/
IEEE
" » Comparing Six Deep Learning Feature Extractors for CBIR Tasks." Image Recognition | Sciencx [Online]. Available: https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/. [Accessed: ]
rf:citation
» Comparing Six Deep Learning Feature Extractors for CBIR Tasks | Image Recognition | Sciencx | https://www.scien.cx/2025/08/27/comparing-six-deep-learning-feature-extractors-for-cbir-tasks/ |

Please log in to upload a file.




There are no updates yet.
Click the Upload button above to add an update.

You must be logged in to translate posts. Please log in or register.