Comparative Analysis of Deep Convolutional Neural Networks for Detecting Medical Image Deepfakes

Authors: Abdel Rahman Alsabbagh, Omar Al-Kadi

Published: 2024-01-08 16:37:22+00:00

AI Summary

This research evaluates 13 state-of-the-art Deep Convolutional Neural Networks (DCNNs) for detecting deepfakes in medical images. ResNet50V2 shows superior precision and specificity, while DenseNet169 excels in accuracy, recall, and F1-score; MobileNetV3Large offers a competitive balance of speed and performance.

Abstract

Generative Adversarial Networks (GANs) have exhibited noteworthy advancements across various applications, including medical imaging. While numerous state-of-the-art Deep Convolutional Neural Network (DCNN) architectures are renowned for their proficient feature extraction, this paper investigates their efficacy in the context of medical image deepfake detection. The primary objective is to effectively distinguish real from tampered or manipulated medical images by employing a comprehensive evaluation of 13 state-of-the-art DCNNs. Performance is assessed across diverse evaluation metrics, encompassing considerations of time efficiency and computational resource requirements. Our findings reveal that ResNet50V2 excels in precision and specificity, whereas DenseNet169 is distinguished by its accuracy, recall, and F1-score. We investigate the specific scenarios in which one model would be more favorable than another. Additionally, MobileNetV3Large offers competitive performance, emerging as the swiftest among the considered DCNN models while maintaining a relatively small parameter count. We also assess the latent space separability quality across the examined DCNNs, showing superiority in both the DenseNet and EfficientNet model families and entailing a higher understanding of medical image deepfakes. The experimental analysis in this research contributes valuable insights to the field of deepfake image detection in the medical imaging domain.


Key findings
ResNet50V2 exhibited the highest precision and specificity. DenseNet169 achieved the best accuracy, recall, and F1-score. MobileNetV3Large provided a good balance between performance and computational efficiency.
Approach
The study uses 13 pre-trained DCNNs, modified by adding a global average pooling layer and fully connected layers for binary classification. The models were trained on a dataset of real and CT-GAN generated medical images, using binary cross-entropy loss and the Adam optimizer. Performance was evaluated across multiple metrics.
Datasets
LIDC-IDRI (real medical images), CT-GAN generated images (deepfakes)
Model(s)
ConvNeXtTiny, DenseNet121, DenseNet169, DenseNet201, EfficientNetB4, EfficientNetV2S, InceptionV3, MobileNetV3Large, RegNetX040, RegNetY040, ResNet50V2, VGG19, Xception
Author countries
Jordan, Saudi Arabia