CT - Based Organ Anomaly Detection Using U-Net and Convolutional Neural Networks Hybrid Technique
Keywords:
Deep learning, CNN, Autoencoder, CT Scan, U-Net, Medical Image Segmentation, Anomaly DetectionAbstract
Deep learning-based medical image analysis advanced significantly with transformer-based architectures and self-configuring segmentation frameworks. This paper presents an implementation-focused approach for CT-based organ anomaly detection using deep learning techniques. The proposed system integrates the U-Net framework for accurate and automated organ segmentation, ensuring adaptive performance across different CT datasets without manual tuning. For feature extraction, convolutional neural networks (CNNs) and transformer-based models are employed to capture both local spatial patterns and global contextual relationships in CT images. Anomaly detection is performed using classification-based methods to distinguish normal and abnormal regions, along with reconstruction-based techniques such as autoencoders to identify deviations through reconstruction error. The integration of these methods improves robustness and accuracy in detecting organ abnormalities. The primary objective of the system is to enhance diagnostic precision, reduce radiologist workload, and enable automated identification of abnormal regions in CT scans, contributing to efficient computer-aided diagnosis systems in healthcare applications. Accurate detection of organ anomalies from Computed Tomography (CT) scans is critical for early diagnosis and treatment planning. This paper presents a hybrid deep learning framework combining U-Net for organ segmentation and Convolutional Neural Networks (CNNs) for anomaly classification. The proposed approach first segments the target organ using a U-Net architecture and subsequently classifies the segmented region into normal or abnormal categories using a CNN model. Experimental results demonstrate improved accuracy and robustness compared to standalone classification models.
Downloads
References
F. Isensee et al., “nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation,” Nature Methods, 2021–2022.
J. Chen et al., “TransUNet: Transformers make strong encoders for medical image segmentation,” 2021–2022.
A. Hatamizadeh et al., “UNETR: Transformers for 3D medical image segmentation,” 2022.
X. Li et al., “Swin UNETR: Swin Transformers for semantic segmentation of brain tumors in MRI images,” 2022.
Z. Zhou et al., “U-Net++: A nested U-Net architecture for medical image segmentation,” 2018–2022.
X. Chen et al., “Deep CNN-based classification for CT image analysis using ResNet and DenseNet,” 2022.
W. Wang et al., “Unsupervised anomaly detection using autoencoders in medical imaging,” 2022.
Y. Zhang et al., “Self-supervised learning for medical image analysis,” 2022.
S. Bakas et al., “Advancing The Cancer Genome Atlas glioma MRI collections with expert segmentation labels and radiomic features,” 2018–2022.
Multiple Authors, “Hybrid CNN-Transformer architectures for medical image analysis: A review,” 2022.
Olaf Ronneberger, Philipp Fischer, and Thomas Brox, “U-Net: Convolutional Networks for Biomedical Image Segmentation,” in Proc. Int. Conf. Medical Image Computing and Computer-Assisted Intervention (MICCAI), 2015, pp. 234–241.
Alex Krizhevsky, Ilya Sutskever, and Geoffrey Hinton, “ImageNet Classification with Deep Convolutional Neural Networks,” in Proc. Advances in Neural Information Processing Systems (NeurIPS), 2012, pp. 1097–1105.
Geert Litjens et al., “A Survey on Deep Learning in Medical Image Analysis,” Medical Image Analysis, vol. 42, pp. 60–88, 2017.
Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun, “Deep Residual Learning for Image Recognition,” in Proc. IEEE Conf. Computer Vision and Pattern Recognition (CVPR), 2016, pp. 770–778.
Downloads
Published
How to Cite
Issue
Section
License

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
All papers should be submitted electronically. All submitted manuscripts must be original work that is not under submission at another journal or under consideration for publication in another form, such as a monograph or chapter of a book. Authors of submitted papers are obligated not to submit their paper for publication elsewhere until an editorial decision is rendered on their submission. Further, authors of accepted papers are prohibited from publishing the results in other publications that appear before the paper is published in the Journal unless they receive approval for doing so from the Editor-In-Chief.
IJISAE open access articles are licensed under a Creative Commons Attribution-ShareAlike 4.0 International License. This license lets the audience to give appropriate credit, provide a link to the license, and indicate if changes were made and if they remix, transform, or build upon the material, they must distribute contributions under the same license as the original.


