Next Article in Journal
The Impact of Diabetes Education on Continuous Glucose Monitoring in SUS-Dependent Patients in a Northeastern Brazilian City
Previous Article in Journal
Optimizing Childhood Obesity Management: The Role of Edmonton Obesity Staging System in Personalized Care Pathways
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using Super-Resolution for Enhancing Visual Perception and Segmentation Performance in Veterinary Cytology

1
ACC Cyfronet AGH, Nawojki 11, 30-950 Kraków, Poland
2
AGH University of Science and Technology, al. Mickiewicza 30, 30-059 Kraków, Poland
3
University of Life Sciences, al. Akademicka 13, 20-950 Lublin, Poland
*
Authors to whom correspondence should be addressed.
Life 2024, 14(3), 321; https://doi.org/10.3390/life14030321
Submission received: 6 December 2023 / Revised: 18 February 2024 / Accepted: 23 February 2024 / Published: 28 February 2024
(This article belongs to the Section Animal Science)

Abstract

:
The primary objective of this research was to enhance the quality of semantic segmentation in cytology images by incorporating super-resolution (SR) architectures. An additional contribution was the development of a novel dataset aimed at improving imaging quality in the presence of inaccurate focus. Our experimental results demonstrate that the integration of SR techniques into the segmentation pipeline can lead to a significant improvement of up to 25% in the mean average precision (mAP) metric. These findings suggest that leveraging SR architectures holds great promise for advancing the state-of-the-art in cytology image analysis.

1. Introduction, Motivation

Veterinary cytology is a specialized field within veterinary medicine that focuses on the microscopic examination of cells and cellular structures in animals. This discipline plays a crucial role in diagnosing and monitoring various diseases and conditions in animals. By analyzing the morphology, size, and arrangement of cells obtained through techniques like fine-needle aspiration, veterinarians can identify abnormalities such as cancerous growths or infections, aiding in the development of appropriate treatment plans.
The process of veterinary cytology can be time consuming and labor intensive, requiring skilled professionals to examine numerous cell samples under a microscope. Artificial intelligence can significantly expedite this process by automating cell image analysis, identifying and categorizing abnormalities, and providing veterinarians with quicker, more accurate diagnostic results. This not only saves valuable time but also enhances the efficiency of veterinary care, ultimately improving the outcomes for animals in need of diagnosis and treatment.
In recent years, deep learning-based solutions have emerged as a prominent topic in the field of Information Technology. Novel approaches are being developed and implemented daily to optimize, enhance, and facilitate various aspects of our lives. This growing trend is also evident in the medical domain, including veterinary medicine. It is important to note that deploying models for healthcare applications entails significant responsibility and necessitates rigorous testing and monitoring to mitigate any risks associated with artificial intelligence (AI) predictions.
Our research team is dedicated to developing solutions that assist veterinarians in making faster and more accurate diagnoses for their animal patients. Our prior work has focused on age classification [1], object segmentation [2], and detection [3] in the context of cytology imaging for canines. In this study, we continue our investigation into AI applications in the veterinary field, specifically exploring the combination of Super-Resolution (SR) and Semantic Segmentation techniques. By building upon previous research, we aim to further advance the state-of-the-art in veterinary image analysis and improve diagnostic outcomes.

2. State-of-the-Art, Reason for Conducting the Research

The acquisition of cytology images is a multifaceted process that involves the preparation of tissue samples using staining methods such as Diff-Quik, followed by the selection of suitable areas by a veterinary expert and image capture via a microscope-mounted camera. This study aims to address the challenges associated with obtaining images of inadequate focus or suboptimal quality for examination purposes. The integration of super-resolution (SR) techniques into the image analysis process represents a promising avenue for enhancing the quality and utility of cytology images, thereby facilitating more accurate and reliable diagnostics.
Research Questions:
Can deep learning-based architectures enhance the quality and resolution of cytology images, thereby facilitating improved image quality assessments? To what extent can such enhancements aid pathologists in diagnosing challenging or average-quality cases? Does the improvement of image quality augment the performance of semantic segmentation architectures in detecting carcinogenic cells within preparations?

2.1. Advancements in Super-Resolution and Semantic Segmentation

The field of medical imaging has seen significant advancements through the integration of deep learning techniques, particularly in super-resolution (SR) and semantic segmentation. He et al.’s development of deep residual learning for image recognition [4] laid the groundwork for subsequent innovations in image analysis, including the application of SR to improve image quality. Similarly, the introduction of U-Net by Ronneberger et al. [5] revolutionized biomedical image segmentation, offering a powerful tool for detailed tissue and cell analysis.

2.2. Dual Super-Resolution Learning for Semantic Segmentation

Building on these foundations, Wang et al. [6] proposed a dual super-resolution learning framework specifically designed to enhance segmentation accuracy without additional computational costs. This approach directly addresses the challenges faced in cytology image analysis, where the precision of segmentation is paramount. By integrating SR techniques into the segmentation process, this method not only improves image quality but also enhances the efficiency of diagnostic models.

2.3. Effectiveness of Super-Resolution in Medical Imaging

The effectiveness of SR techniques in medical imaging has been further demonstrated by Ledig et al. [7], who developed a generative adversarial network for photo-realistic super-resolution. This advancement underscores the potential of SR to significantly improve the resolution and clarity of medical images, facilitating more accurate diagnoses. Moreover, Dong et al. [8], Zhang et al. [9], Pereira et al. [10], Pham et al. [11], Goodfellow et al. [12] and Wang et al. [13] have contributed to the field by developing deep learning models that enhance the resolution of images, proving essential for detailed medical analysis and diagnosis.

2.4. Super-Resolution for Enhanced Diagnostic Accuracy

The integration of SR with semantic segmentation has shown promise in enhancing diagnostic accuracy. Studies by Chen et al. [14] and Zhou et al. [15] have demonstrated the potential of deep-learning models to improve semantic segmentation, which, when combined with high-resolution images obtained through SR techniques, can significantly aid in the detection and classification of diseases. Furthermore, the work by Isola et al. [16] on image-to-image translation with conditional adversarial networks has opened new avenues for applying SR in medical imaging, suggesting that enhancing image quality can substantially improve the performance of segmentation models.

2.5. Recent Progress and Integrative Approaches in Super-Resolution

Recent studies have underscored the transformative impact of super-resolution techniques across various domains of medical imaging, offering novel perspectives and methodological advances. For instance, Wang et al. [17] introduced a smarter microscope system that leverages E-CNN-based super-resolution to enhance single cell analysis, marking a significant step forward in cytology imaging. Similarly, Zhang et al. [18] explored the diagnostic potential of super-resolution in visualizing biliary structures, demonstrating its utility in clinical settings. Caputa et al. [19] specifically addressed veterinary cytology by integrating super-resolution to improve visual perception and segmentation performance, which is directly relevant to our research objectives. These studies, alongside contributions from Li et al. [20], Ma et al. [21], and Yuqian et al. [22], illustrate a collective move towards enhancing diagnostic accuracy and image quality through advanced computational techniques. The synthesis of super-resolution and deep learning not only promises to refine diagnostic capabilities in cytology but also sets a new benchmark for precision in medical image analysis.

3. Contribution, New Algorithm, Constructed System

The primary contribution of this study is the incorporation of a Super-Resolution module into the machine-learning pipeline, with the aim of enhancing the accuracy of segmentation models (Figure 1). This potential application for improving image quality emerged as a result of various distortions that may occur during the acquisition of cytological preparation images (Table 1). Our research is specifically focused on addressing poor sharpness distortions.
The objective of this study is to develop a machine learning model capable of enhancing the quality of images affected by improper focus settings on the microscope’s adjustment knob. To evaluate the effectiveness of this approach, the enhanced images will be compared to properly created images. The development of a dedicated dataset is a prerequisite for this evaluation [23].
In the context of veterinary examinations, an animal patient undergoes evaluation when visible skin alterations are observed. A tissue sample is subsequently obtained and examined under a microscope, during which an image is generated. In instances where the image quality is sub optimal or the microscope lens focus is improperly set, the decision block (binary image classifier) routes the image to the Super-Resolution model. Following this enhancement process, the segmentation model identifies objects within the image, and a diagnosis is proposed.

3.1. Novel Dataset

The majority of datasets for Super-Resolution (SR) tasks are artificially generated, employing image downscaling and interpolation techniques. However, the nature of the distortion we aim to address is distinct from these methods. Recognizing this led to the development of an experimental dataset [23] in collaboration with a veterinary expert, which is elaborated upon in the subsequent chapter.
The following algorithm was proposed for the acquisition of samples:
1.
Identify the diagnostic region within the cytological preparation.
2.
Adjust the microscope lens focus to obtain a high-quality image.
3.
Intentionally alter the microscope’s adjustment knob to degrade the image quality and sharpness, thereby simulating the real-world distortion.
This approach enabled the generation of a dataset that more accurately represents the specific type of distortion we aim to mitigate, providing a more suitable foundation for model training and evaluation as in Figure 2.
Following this procedure we collected 1192 high resolution (2592 × 1944) samples with their corresponding distorted versions.

3.2. Proposed New Super-Resolution Metric

In this study, an additional metric based on frequency analysis is employed as a novel approach to assess segmentation performance. This method involves grouping the energy computed using a two-dimensional Discrete Fourier Transform (DFT-2D) by frequency. This operation enables the observation of the total energy within each frequency range, providing insights into how machine learning models affect high frequencies in the image, which contribute to visual sharpness. The procedure for this approach is as follows (Figure A1):
1.
Open an image in YCbCr mode and use only the luminance channel.
2.
Compute the two-dimensional discrete Fourier Transform and shift the zero-frequency component to the center of the spectrum.
3.
Calculate the absolute sum of all magnitudes for a chosen set of ring-shaped masks and display the results in a bar plot.
This method employs visual representation to analyze the frequencies present within an image. Primary advantage is that it is not based on singular metrics such as SSIM [24] or LPIPS [25], as it captures changes in distribution subsequent to the application of filters or models.

3.3. Research Formula for Specific Medical Use Case with Unknown Degradation

During the course of the research, addressing the super-resolution (SR) task when the nature of degradation is unknown was a significant concern. The following strategies were assessed:
1.
Use pre-trained models on various datasets;
2.
Investigate known degradations, such as bicubic interpolation on our medical dataset;
3.
Develop a dedicated super-resolution dataset exhibiting the same degradation intended to be mitigated.
The most favorable results were achieved using the third approach; however, it is crucial to consider feedback from domain experts in the medical field. It was discovered that applying super-resolution to images introduced artifacts that would not typically be present in cytology images. Veterinary specialists tended to favor lower-quality images over sharper ones, as the artifacts introduced by SR models hindered the diagnostic process. One of the key conclusions drawn is that a sharper image does not necessarily equate to a superior model. Therefore, in medical applications metrics cannot be the only evaluator for performance assessment.

4. Experiments and Results

This chapter provides a detailed look at the experiments conducted during this research study. Each subchapter explains the different stages of the study in a more accessible way, while still maintaining scientific accuracy.

4.1. Comparison of Possible Distortions in Cytology Imaging

Table 1 presents a list of potential distortions that may occur during image creation by veterinary experts. These hypothetical scenarios may require the application of Super-Resolution models as a preprocessing step to recover the images to their desired quality.

4.2. Comparison of Different Image Upsampling Methods Using Scale Factor 2

This section presents a visual analysis of various classical image upsampling techniques when applied with a scale factor of 2. The objective is to evaluate the performance of each method in terms of image quality, preservation of structural details, and overall effectiveness in enhancing the resolution of the original image (Figure A2).

4.3. Pretrained Segmentation Model Approach

This section contains the results of inferencing pretrained Super-Resolution models and measuring their impact on semantic segmentation task on our cytology dataset.
In this comparative analysis, three state-of-the-art image upsampling architectures were selected for evaluation: SwinIR, BSRGAN, and RealSRGAN. The goal was to assess each model’s performance in terms of image quality and impact on segmentation metrics.
For segmentation evaluation, a deep learning model based on the Cascade Mask R-CNN [26] architecture was selected. The ResNeSt101 [27], which employs skip connections (i.e., input values bypass the current layer without any modifications, and are then summed with the modified input), was used for feature extraction. The model was initialized with weights pre-trained on the MS COCO dataset [28] and subsequently fine-tuned on cytology images.

4.3.1. Impact of Bicubic Interpolation on Segmentation Inference

The results presented in Table 2 reveal an expected trend. As the bicubic interpolation scaling factor increases, which corresponds to a greater loss of information, both segmentation and super-resolution metrics are negatively affected. A higher scaling factor leads to increased confusion between objects. For instance, with bicubic interpolation using a scaling factor of five, almost no cells are accurately recognized for the two cancer types, as illustrated in Figure A3.
Figure A4 presents the relationship between the segmentation Average Precision (segmAP) and Peak Signal-to-Noise Ratio (PSNR) metrics. A decreasing trend in the ratio between these two metrics is observed when the scaling factor increases. This indicates that, in some cases, a linear correlation exists between the performance metrics of these two distinct computer vision tasks. As the scaling factor increases, the quality of the image and segmentation performance both tend to degrade.

4.3.2. Impact of the Pre-Trained Super-Resolution Models on Segmentation Inference

In this phase, a comparative analysis is conducted to evaluate the performance of the segmentation model on the original dataset opposite to bicubic interpolation with scaling factors of two and five. The primary objective is to investigate whether employing pre-trained models can enhance the accuracy of the segmentation process.
The application of various super-resolution (SR) architectures on the original data did not yield any improvements in segmentation quality. Nevertheless, the minimal loss in mean average precision (mAP) indicates that the model is proficient in identifying cancer cells that have undergone enhancement through the SR process (Table 3).
When employing image enhancement techniques for damaged data with decimation and bicubic interpolation, the results, as presented in Table A1 and Table A2, are found to be worse. This suggests that the utilization of SR introduces additional noise to the data, consequently leading to poorer performance by the segmentation model. This outcome was anticipated, given that the model was pre-trained on original data and was subsequently required to handle processed data during the inference phase.
The findings indicate that the naive application of SR models to images does not yield improvements in segmentation quality. However, it does enhance the perceptual quality of the image, as depicted in Figure A5. In certain instances, it also results in an increase in SR metrics, as demonstrated in Table 3.
The obtained results were subsequently reviewed in consultation with a veterinary expert. It was determined that, although the images appeared sharper following the application of the super-resolution model, the presence of certain artifacts rendered them less reliable than the original images. The enhanced image quality did not contribute to improved diagnostic accuracy, as these artifacts introduced elements that would not typically be found in cytology images.

4.3.3. Training Semantic Segmentation Model on Super-Resolution Medical Data

In this experiment, we trained and tested the segmentation on data processed in various ways. As demonstrated in Table 4, we downsampled the images using decimation and bicubic interpolation and then upsampled using super-resolution architectures. While this research may not hold practical significance from a medical standpoint, as manipulating original data is generally discouraged, it does reveal that the optimal results for cancer cell recognition are obtained when utilizing undistorted, original data.

4.3.4. Dedicated Dataset Experiments for Super-Resolution

Ultimately, the experiments were carried out on a dedicated dataset, with the SwinIR architecture selected for training [29].
During the exploration of the dataset, the data distribution was analyzed. The histograms presented in Figure A8 showcase the PSNR values for both bicubic interpolation and our dedicated dataset, in comparison to high-resolution original data. The spectrum of our dataset is wider and there are images that would be considered of a good quality in terms of pixel loss. In contrast, bicubic interpolation exhibits less diversity, limiting its applicability to the restoration of specific distortion types.
The dedicated dataset encompasses various forms of degradation that are likely to be encountered in cytology images.
The initial experiment exhibited a substantial improvement in the PSNR metric upon training the SwinIR model on our dataset, as displayed in Table A3. This improvement is also evident in the inferred images after training, presented in Figure A6.
The transformer model underwent training for approximately 1000 epochs, utilizing four NVIDIA V-100 GPUs from the ACK Cyfronet Prometheus supercomputer [30]. Default parameters tailored for the classical super-resolution task were employed. The upsampling factor was set to 2.
Experiments involving training on our dataset demonstrated a notable improvement when compared to the bicubic dataset, which is commonly utilized in super-resolution tasks.
The second experiment focused on examining the influence of the selected spectrum. We investigated whether a narrow or wide spectrum of our dataset would yield superior results. Figure A7 illustrates the three distinct segments of the dataset that were employed in this experiment.
The experiment reveals that training on the widest spectrum (as illustrated in Figure A7) leads to the the most favorable results for both narrow and wide spectrum test datasets, as presented in Table A4. This finding suggests that the model effectively learns to reconstruct images when the training data encompasses a diverse range of distortions with varying types and levels (Figure 3).

4.3.5. Improving Segmentation Metrics Results with Super-Resolution

For end-to-end experiments aimed at enhancing segmentation using super-resolution (SR) architectures, a new subset of the dataset was employed. This subset comprised both semantic segmentation annotations and distorted images representing the exact same area.
The results depicted in the subsequent table exhibit promising outcomes. In certain methodologies, the results show improvements when compared to the low-quality dataset. For instance, with the BSRGAN architecture, employing a 4x upsampling technique followed by subsampling to the required resolution, the improvement reaches up to 25% when compared to the results that would have been obtained using a low-quality dataset. As anticipated, the most optimal results are achieved when training the segmentation on high-quality images, where the segm_mAP_75 increase is up to 64%. Interesting aspect of the following experiments is that using super-resolution as a first step, before subsampling to desired annotations size leads to better results than the opposite operation. This is because subsampling as a first step loses even more information that can not be restored during upsampling step, leading to lower segmentation results.
The intriguing outcome of the subsequent experiments in Table 5 suggests that implementing super-resolution as an initial step, prior to the subsampling process to achieve the desired annotation size, yields superior results compared to the inverse sequence of operations. This observed phenomena could be primarily attributed to the fact that the adoption of subsampling as a preliminary step entails an excessive loss of information. This loss of detail, once occurred, is unable to be fully recuperated during the upsampling process, culminating in compromised segmentation outcomes. Hence, the order of these processing steps plays a critical role in maximizing the data fidelity and overall accuracy of the segmentation results.

5. Conclusions

The presented research, which encompasses two fields of Computer Vision—Super-Resolution and Semantic Segmentation—underscores the possibility of enhancing the quality of medical images for their interpretation and analysis. The primary challenges identified during the investigation include a scarcity of data and differing perceptions between programmers and medical experts.
The first challenge stems from the nature of the formulated problem. Restoring an image from an unknown degradation is a daunting task, particularly in medical imaging. Consequently, a unique dataset was created for this study, enabling improvements in both super-resolution metrics and human perception. The second challenge arises from the specific use case provided; veterinary experts analyze medical images differently from individuals unfamiliar with cytology. The realization that sharper, high-resolution images are not always preferable for diagnosis due to artifacts observable after applying AI models was not immediately evident.
This research primary achievements include formulating a potential procedure for addressing unknown degradation, such as incorrectly set sharpness on a microscope. The steps taken during the experimental phase could potentially be applied to other domain-specific use cases. Another valuable aspect is the identification and analysis of possible distortions in medical imaging. We facilitated a better understanding of the problem’s nature and its uniqueness compared to standard Super-Resolution tasks.
Finally as expected, the substantial increase in the PSNR measure during SwinIR architecture training (Table A3) and the visual perception improvement shown in Figure A6 are noteworthy. The remarkable improvement of up to 25% in certain experimental scenarios, with respect to the segm_mAP metric change, is also worth mentioning. This underscores the potential of the applied methods in enhancing the performance of image segmentation tasks in medical imaging.

Author Contributions

Conceptualization, J.C., J.G. and M.W.; methodology, J.G.; software, J.C., J.G. and M.K.; validation, J.C., J.G. and M.K.; formal analysis, P.R. and M.W.; investigation, S.M., R.F., A.Ś., E.J., S.K., A.D.-B., M.P., K.W. and M.W.; data curation, D.Ł.; writing—original draft preparation, J.C.; writing—review and editing, J.C. and M.W.; visualization, J.C.; supervision, J.C., P.R., K.W. and M.W.; project administration, P.R.; funding acquisition, K.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available in a publicly accessible repository. The data presented in this study are openly available at https://github.com/jakubcaputa/super-resolution-dataset (accessed on 22 February 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. 2D-DFT magnitude bar plot for original image and two bicubic interpolation degradations, example with 10 ring-shaped masks.
Figure A1. 2D-DFT magnitude bar plot for original image and two bicubic interpolation degradations, example with 10 ring-shaped masks.
Life 14 00321 g0a1
Figure A2. Comparison of different image upsampling methods using Scale 2. Magnification 400×.
Figure A2. Comparison of different image upsampling methods using Scale 2. Magnification 400×.
Life 14 00321 g0a2
Figure A3. True Positive cell detection numbers for different scaling factors.
Figure A3. True Positive cell detection numbers for different scaling factors.
Life 14 00321 g0a3
Figure A4. Correlation between segmAP and PSNR metrics for different scaling factors.
Figure A4. Correlation between segmAP and PSNR metrics for different scaling factors.
Life 14 00321 g0a4
Figure A5. Example results of chosen architectures (cropped part of the image). Magnification 400×.
Figure A5. Example results of chosen architectures (cropped part of the image). Magnification 400×.
Life 14 00321 g0a5
Figure A6. Comparison of results for training on dedicated and bicubic datasets. Magnification 400×.
Figure A6. Comparison of results for training on dedicated and bicubic datasets. Magnification 400×.
Life 14 00321 g0a6
Table A1. Results for inference on the test set for bicubic interpolation with Factor 2.
Table A1. Results for inference on the test set for bicubic interpolation with Factor 2.
Data Set Bicubic2Segm_mAPAvg_PrecisionAvg RecallPSNRSSIMLPIPS
original0.4260.5660.63334.690.920.0084
realSRGAN0.3790.5590.62434.490.920.073
BSRGAN0.3600.5360.60733.770.900.090
SwinIR0.3450.5340.60633.540.910.087
SwinIR_large0.3390.5410.61333.700.910.091
Figure A7. Histogram presenting chosen spectrums.
Figure A7. Histogram presenting chosen spectrums.
Life 14 00321 g0a7
Table A2. Results for inference on the test set for bicubic interpolation with Factor 5.
Table A2. Results for inference on the test set for bicubic interpolation with Factor 5.
Data Set Bicubic5Segm_mAPAvg_PrecisionAvg RecallPSNRSSIMLPIPS
original0.1160.3110.41931.580.720.377
realSRGAN0.0920.3000.39731.380.690.252
BSRGAN0.0930.2760.38131.460.700.282
SwinIR0.0710.2500.37431.290.690.301
SwinIR_large0.0550.2270.35731.170.670.348
Table A3. Dedicated and bicubic results.
Table A3. Dedicated and bicubic results.
Train Data SetTest Data SetPSNR
bicubicdedicated25.84
dedicateddedicated29.99
Table A4. Spectrum experiments PSNR value.
Table A4. Spectrum experiments PSNR value.
Trained
ValidatedSpectrumNarrowMiddleWide
Narrow30.4130.1429.73
Middle30.6430.0629.69
Wide30.7629.4029.90
Figure A8. Comparison of histograms for datasets.
Figure A8. Comparison of histograms for datasets.
Life 14 00321 g0a8

References

  1. Mazurek, S.; Wielgosz, M.; Caputa, J.; Frączek, R.; Karwatowski, M.; Grzeszczyk, J.; Łukasik, D.; Śmiech, A.; Russek, P.; Jamro, E.; et al. Canine age classification using Deep Learning as a step toward preventive medicine in animals. In Proceedings of the 17th Conference on Computer Science and Intelligence Systems (FedCSIS), Sofia, Bulgaria, 4–7 September 2022; pp. 169–172. [Google Scholar] [CrossRef]
  2. Grzeszczyk, J.; Karwatowski, M.; Łukasik, D.; Wielgosz, M.; Russek, P.; Mazurek, S.; Caputa, J.; Frączek, R.; Śmiech, A.; Jamro, E.; et al. Segmentation of the Veterinary Cytological Images for Fast Neoplastic Tumors Diagnosis. arXiv 2023, arXiv:2305.04332. [Google Scholar]
  3. Caputa, J.; Łukasik, D.; Wielgosz, M.; Karwatowski, M.; Frączek, R.; Russek, P.; Wiatr, K. Fast Pre-Diagnosis of Neoplastic Changes in Cytology Images Using Machine Learning. Appl. Sci. 2021, 11, 7181. [Google Scholar] [CrossRef]
  4. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 29 June–1 July 2016; pp. 770–778. [Google Scholar]
  5. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  6. Wang, L.; Li, D.; Zhu, Y.; Tian, L.; Shan, Y. Dual Super-Resolution Learning for Semantic Segmentation. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3773–3782. [Google Scholar]
  7. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  8. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV. Springer: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
  9. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  10. Pereira, M.; dos Santos, J. How Effective Is Super-Resolution to Improve Dense Labelling of Coarse Resolution Imagery? In Proceedings of the 32nd SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Rio de Janeiro, Brazil, 28–30 October 2019; pp. 202–209. [Google Scholar]
  11. Pham, C.H.; Tor-Díez, C.; Meunier, H.; Bednarek, N.; Fablet, R.; Passat, N.; Rousseau, F. Simultaneous Super-Resolution and Segmentation Using a Generative Adversarial Network: Application to Neonatal Brain MRI. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 991–994. [Google Scholar]
  12. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 2014, 27, 139–144. [Google Scholar] [CrossRef]
  13. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Loy, C.C. ESRGAN: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  15. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  16. Isola, P.; Zhu, J.Y.; Zhou, T.; Efros, A.A. Image-to-Image Translation with Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 1125–1134. [Google Scholar]
  17. Wang, W.; Yang, L.; Sun, H.; Peng, X.; Yuan, J.; Zhong, W.; Chen, J.; He, X.; Ye, L.; Zeng, Y.; et al. Cellular Nucleus Image-based Smarter Microscope System for Single Cell Analysis. Biosens. Bioelectron. 2024, 250, 116052. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, W.; Chai, N.; Wu, Q.; Linghu, E. The New Criteria for Differential Diagnosis of Indeterminate Biliary Stricture under Super Minimally Invasive Peroral Cholangioscopy. Chin. Med. J. 2024, 137, 255–256. [Google Scholar] [CrossRef] [PubMed]
  19. Caputa, J.; Wielgosz, M.; Łukasik, D.; Russek, P.; Grzeszczyk, J.; Karwatowski, M.; Mazurek, S.; Frączek, R.; Śmiech, A.; Wiatr, K.; et al. Using Super-Resolution for Enhancing Visual Perception and Segmentation Performance in Veterinary Cytology. arXiv 2023, arXiv:2306.11848. [Google Scholar]
  20. Li, H.; Jia, Y.; Zhu, H.; Han, B.; Du, J.; Liu, Y. Multi-level Feature Extraction and Reconstruction for 3D MRI Image Super-Resolution. Comput. Biol. Med. 2024, 108151. [Google Scholar] [CrossRef] [PubMed]
  21. Ma, Y.; Zhou, W.; Ma, R.; Wang, E.; Yang, S.; Tang, Y.; Zhang, X.-P.; Guan, X. DOVE: Doodled Vessel Enhancement for Photoacoustic Angiography Super-Resolution. Med. Image Anal. 2024, 103106. [Google Scholar] [CrossRef] [PubMed]
  22. Yuqian, W.; Aksenov, S.V. Convolutional Neural Networks for Brain Lesion Segmentation in MRI Images. In Proceedings of the Modern Problems in Mechanical Engineering: Collection of Articles from the XVI International Scientific and Technical Conference, Tomsk, Russia, 27 November–1 December 2023; Tomsk Polytechnic University: Tomsk, Russia, 2023; pp. 385–386. [Google Scholar]
  23. Super-Resolution Data Set. Available online: https://github.com/jakubcaputa/super-resolution-dataset (accessed on 22 February 2024).
  24. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  26. Cai, Z.; Vasconcelos, N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1483–1498. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, H.; Wu, C.; Zhang, Z.; Zhu, Y.; Lin, H.; Zhang, Z.; Sun, Y.; He, T.; Mueller, J.; Manmatha, R.; et al. ResNeSt: Split-Attention Networks. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), New Orleans, LA, USA, 19–20 June 2020. [Google Scholar]
  28. Lin, T.Y.; Maire, M.; Belongie, S.; Bourdev, L.; Girshick, R.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L.; Dollár, P. Microsoft COCO: Common Objects in Context. In Computer Vision—ECCV 2014, Proceedings of the 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Springer International Publishing: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  29. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  30. Prometheus Supercomputer. Available online: https://kdm.cyfronet.pl/portal/Prometheus:Podstawy (accessed on 22 February 2024).
Figure 1. Proposed working system scheme.
Figure 1. Proposed working system scheme.
Life 14 00321 g001
Figure 2. Examples of high and low sharpness images from the dataset. Magnification 400×.
Figure 2. Examples of high and low sharpness images from the dataset. Magnification 400×.
Life 14 00321 g002
Figure 3. Example result for training on a wide spectrum. Magnification 400×.
Figure 3. Example result for training on a wide spectrum. Magnification 400×.
Life 14 00321 g003
Table 1. Potential distortions in veterinary image creation.
Table 1. Potential distortions in veterinary image creation.
Distortion
Type
ImageDescription
correct imageLife 14 00321 i001Image properly created
dark lightingLife 14 00321 i002microscope bulb is not turned on or the room where the image is created is dark, the resulting image may suffer from low contrast and poor illumination.
closed apertureLife 14 00321 i003Responsible for the amount of the light that comes to a focus in the image plane
closed condensorLife 14 00321 i004An improperly adjusted condenser, which is responsible for providing evenly distributed illumination
dark outsideLife 14 00321 i005The image was not directly at the lens leading to dark edges
bad sharpnessLife 14 00321 i006The microscope screw set inaccurately or the focus is set for the background of the image
Table 2. Comparison of super-resolution and segmentation metrics using pretrained segmentation model for inference.
Table 2. Comparison of super-resolution and segmentation metrics using pretrained segmentation model for inference.
FactorSegm_mAPAvg_PrecisionAvg RecallPSNRSSIMLPIPS
Original0.4390.6230.679-10
20.3920.5980.66334.980.930.07
30.2760.4920.57332.620.820.17
40.1950.4440.54032.190.790.26
50.1130.3480.45631.680.740.35
Table 3. Results for inference on the test set for the original dataset.
Table 3. Results for inference on the test set for the original dataset.
Data Set OriginalSegm_mAPAvg_PrecisionAvg_RecallPSNRSSIMLPIPS
Original0.4570.5960.654---
RealSRGAN0.4460.5870.54036.160.960.050
BSRGAN0.4300.5720.63633.540.930.092
SwinIR0.4210.5670.63035.100.960.057
SwinIR_large0.4570.5840.64337.180.9770.041
Table 4. Training semantic segmentation on interpolated and super-resolution data.
Table 4. Training semantic segmentation on interpolated and super-resolution data.
Data SetData Set ExplanationSegm_mAPAvg_PrecisionAvg_Recall
originalOriginal cancer inflammation dataset0.4940.4940.584
bicubic2Original dataset decimated and interpolated using scaling factor two (half of the pixels left after decimation)0.4650.4650.567
bicubic4Original dataset decimated and interpolated using scaling factor four (quarter of the pixels left after decimation)0.4230.4230.527
realSRGAN ×4Original dataset upscaled using realSRGAN and resized to original size0.4870.4870.587
BSRGAN ×4Original dataset upscaled using BSRGAM and resized to original size0.4780.4780.587
SwinIR ×4Original dataset upscaled using SwinIR and resized to original size0.4820.4820.579
Table 5. End-to-end pipeline tests BSRGAN.
Table 5. End-to-end pipeline tests BSRGAN.
Experiment
Summary
Segm_
mAP_50
Segm_
mAP_75
Segm_mAP_50
Percent_Change
Segm_mAP_75
Percent_Change
high quality dataset0.3140.25538.94%63.46%
low quality dataset0.2260.1560.00%0.00%
subsampling ×2, SR ×20.1980.120−12.39%−23.08%
SR ×2, subsampling ×20.2330.1103.10%−29.49%
subsampling ×4, SR ×40.2270.1550.44%−0.64%
SR ×4, subsampling ×40.2790.19523.45%25.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Caputa, J.; Wielgosz, M.; Łukasik, D.; Russek, P.; Grzeszczyk, J.; Karwatowski, M.; Mazurek, S.; Frączek, R.; Śmiech, A.; Jamro, E.; et al. Using Super-Resolution for Enhancing Visual Perception and Segmentation Performance in Veterinary Cytology. Life 2024, 14, 321. https://doi.org/10.3390/life14030321

AMA Style

Caputa J, Wielgosz M, Łukasik D, Russek P, Grzeszczyk J, Karwatowski M, Mazurek S, Frączek R, Śmiech A, Jamro E, et al. Using Super-Resolution for Enhancing Visual Perception and Segmentation Performance in Veterinary Cytology. Life. 2024; 14(3):321. https://doi.org/10.3390/life14030321

Chicago/Turabian Style

Caputa, Jakub, Maciej Wielgosz, Daria Łukasik, Paweł Russek, Jakub Grzeszczyk, Michał Karwatowski, Szymon Mazurek, Rafał Frączek, Anna Śmiech, Ernest Jamro, and et al. 2024. "Using Super-Resolution for Enhancing Visual Perception and Segmentation Performance in Veterinary Cytology" Life 14, no. 3: 321. https://doi.org/10.3390/life14030321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop