International Journal of Big Data Intelligent Technology, 2024, 5(1); doi: 10.38007/IJBDIT.2024.050116.
Yunhe Li, Mei Yang, Tao Bian, Wenhui Xia and Haitao Wu
Shenzhen Chenzhuo Technology Company, Shenzhen 518055, China
In this paper, a 4-fold super-resolution analysis scheme for CT images based on HAT (Hybrid Attention Transformer) and self-integrated fusion method is proposed. The scheme realizes the effective capture and fusion of global and local information of CT images by constructing the Dense Residual Transformer Super-Resolution Analysis Architecture (DRTSR), which fuses the self-attention mechanism of Transformer and the local perception capability of Convolutional Neural Network (CNN). Meanwhile, the introduction of the self-integrated fusion method reduces the possible bias of a single model and enhances the stability and accuracy of the results by training multiple HAT models and intelligently fusing them in the testing phase. The experimental results show that the method in this paper demonstrates significant advantages in a number of evaluation indexes, and the generated super-resolution images perform better in terms of naturalness and realism, and the details and sharpness are significantly better than other methods. This work provides higher quality image data support for CT image applications and advances the field of image super-resolution analysis.
Computed Tomography, Super-resolution, Hybrid Attention Transformer, Self-integrated Fusion Method
Yunhe Li, Mei Yang, Tao Bian, Wenhui Xia and Haitao Wu. Super-resolution analysis of CT images based on HAT and self-integrated fusion method. International Journal of Big Data Intelligent Technology (2024), Vol. 5, Issue 1: 153-160. https://doi.org/10.38007/IJBDIT.2024.050116.
[1] Zhou, S. Kevin, et al. "A review of deep learning in medical imaging: Imaging traits, technology trends, case studies with progress highlights, and future promises." Proceedings of the IEEE 109.5 (2021): 820-838.
[2] Chen, Hongchi, et al. "Deep learning-based algorithms for low-dose CT imaging: A review." European Journal of Radiology (2024): 111355.
[3] Goodfellow, Ian, et al. "Generative adversarial networks." Communications of the ACM 63.11 (2020): 139-144.
[4] Wang X, Yu K, Wu S, et al. Esrgan: Enhanced super-resolution generative adversarial networks[C]//Proceedings of the European conference on computer vision (ECCV) workshops. 2018: 0-0.
[5] Jiang, Yuchen, et al. "Quo vadis artificial intelligence?." Discover Artificial Intelligence 2.1 (2022): 4.
[6] Han, Kai, et al. "Transformer in transformer." Advances in neural information processing systems 34 (2021): 15908-15919.
[7] Liang, Jingyun, et al. "Swinir: Image restoration using swin transformer." Proceedings of the IEEE/CVF international conference on computer vision. 2021.
[8] Liu, Yihao, et al. "Evaluating the generalization ability of super-resolution networks." IEEE Transactions on pattern analysis and machine intelligence (2023).
[9] Chen, Xiangyu, et al. "Hat: Hybrid attention transformer for image restoration." arXiv preprint arXiv:2309.05239 (2023).
[10] Zhu, Lianghui, et al. "Vision mamba: Efficient visual representation learning with bidirectional state space model." arXiv preprint arXiv:2401.09417 (2024).
[11] Demir, Ugur, and Gozde Unal. "Patch-based image inpainting with generative adversarial networks." arXiv preprint arXiv:1803.07422 (2018).
[12] Wen, Long, et al. "A new transfer learning based on VGG-19 network for fault diagnosis." 2019 IEEE 23rd international conference on computer supported cooperative work in design (CSCWD). IEEE, 2019.
[13] Eldarova, E. L. M. I. R. A., V. A. L. E. R. Y. Starovoitov, and K. A. Z. I. Z. A. T. Iskakov. "Comparative analysis of universal methods no reference quality assessment of digital images." (2021).
[14] Mittal, Anish, Anush Krishna Moorthy, and Alan Conrad Bovik. "No-reference image quality assessment in the spatial domain." IEEE Transactions on image processing 21.12 (2012): 4695-4708.
[15] Li, Yunhe, et al. "4× Super‐resolution of unsupervised CT images based on GAN." IET Image Processing 17.8 (2023): 2362-2374.