Архитектуры глубокого обучения для сегментации мультифазных КТ-изображений
Работая с сайтом, я даю свое согласие на использование файлов cookie. Это необходимо для нормального функционирования сайта, показа целевой рекламы и анализа трафика. Статистика использования сайта обрабатывается системой Яндекс.Метрика
Научный журнал Моделирование, оптимизация и информационные технологииThe scientific journal Modeling, Optimization and Information Technology
Online media
issn 2310-6018

Deep learning architectures for multiphase CT image segmentation

Samsonenko S.V.,  idKashirina I.L.

UDC 004.8
DOI: 10.26102/2310-6018/2026.54.3.012

  • Abstract
  • List of references
  • About authors

The article provides a comprehensive systematic analysis of modern deep learning architectures for automatic segmentation of multiphase CT images. The specific features of multiphase data are considered in detail, the main of which are spatial mismatches (offsets) between phases caused by patient movements and the different nature of the accumulation of contrast agent in pathological tissues at different phases. These features make direct adaptation of classical segmentation methods ineffective and require the development of specialized architectures. The article traces the evolution of approaches: from basic convolutional networks (U-Net, 3D U-Net, nnU-Net) and hybrid models (TransUNet, UNETR) combining convolutions and transformers to specialized solutions. Special attention is paid to models with mechanisms of cross-attention between phases, such as PA-ResSeg, M3Net and MULLET, which allow for implicit alignment of features and adaptive merging of information from different phases without explicit registration (alignment) of images. The paper also analyzes the comparative advantages of various data fusion strategies from different phases (early, late, cross-interaction), discusses issues of computational efficiency and availability of open datasets. Key trends and promising areas of development of the field have been identified, including the use of fundamental models (MedSAM, VoxTell) and modal-agnostic learning. It is concluded that further progress in the field of multiphase segmentation of CT images is associated with the creation of computationally efficient architectures capable of integration into the real clinical process to support diagnostic solutions.

1. Wu L., Wang H., Chen Y., et al. Beyond radiologist-level liver lesion detection on multi-phase contrast-enhanced CT images by deep learning. iScience. 2023;26(11). https://doi.org/10.1016/j.isci.2023.108183

2. Rudenko A.V., Rudenko M.A., Kashirina I.L. The use of artificial neural networks to search for objects in medical images. Modeling, Optimization and Information Technology. 2024;12(3). (In Russ.). https://doi.org/10.26102/2310-6018/2024.46.3.013

3. Xu Y., Cai M., Lin L., et al. PA-ResSeg: A phase attention residual network for liver tumor segmentation from multiphase CT images. Medical Physics. 2021;48(7):3752–3766. https://doi.org/10.1002/mp.14922

4. Ronneberger O., Fischer Ph., Brox Th. U-Net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015: 18th International Conference: Proceedings: Part III, 05–09 October 2015, Munich, Germany. Cham: Springer; 2015. P. 234–241. https://doi.org/10.1007/978-3-319-24574-4_28

5. Çiçek Ö., Abdulkadir A., Lienkamp S.S., Brox Th., Ronneberger O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In: Medical Image Computing and Computer-Assisted Intervention – MICCAI 2016: 19th International Conference: Proceedings: Part II, 17–21 October 2016, Athens, Greece. Cham: Springer; 2016. P. 424–432. https://doi.org/10.1007/978-3-319-46723-8_49

6. Milletari F., Navab N., Ahmadi S.-A. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV), 25–28 October 2016, Stanford, CA, USA. IEEE; 2016. P. 565–571. https://doi.org/10.1109/3DV.2016.79

7. Isensee F., Jaeger P.F., Kohl S.A.A., Petersen J., Maier-Hein K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nature Methods. 2021;18:203–211. https://doi.org/10.1038/s41592-020-01008-z

8. Kulikov A.A., Kashirina I.L., Savkina E.F. Segmentation of liver volumetric lesions in multiphase CT images using the nnU-Net framework. Modeling, Optimization and Information Technology. 2025;13(1). (In Russ.). https://doi.org/10.26102/2310-6018/2025.48.1.040

9. Dosovitskiy A., Beyer L., Kolesnikov A., et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In: 9th International Conference on Learning Representations, ICLR 2021, 03–07 May 2021, Virtual Event, Austria. 2021. https://doi.org/10.48550/arXiv.2010.11929

10. Zheng S., Lu J., Zhao H., et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 20–25 June 2021, Nashville, TN, USA. IEEE; 2021. P. 6877–6886. https://doi.org/10.1109/CVPR46437.2021.00681

11. Chen J., Lu Y., Yu Q., et al. TransUNet: Transformers make strong encoders for medical image segmentation. arXiv. URL: https://doi.org/10.48550/arXiv.2102.04306 [Accessed 15th January 2026].

12. Hatamizadeh A., Tan Y., Nath V., et al. UNETR: Transformers for 3D medical image segmentation. In: Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 03–08 January 2022, Waikoloa, HI, USA. IEEE; 2022. P. 1748–1758. https://doi.org/10.1109/WACV51458.2022.00181

13. Hatamizadeh A., Nath V., Tang Y., et al. Swin UNETR: Swin transformers for semantic segmentation of brain tumors in MRI images. In: Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: 7th International Workshop, BrainLes 2021, Held in Conjunction with MICCAI 2021: Revised Selected Papers: Part I, 27 September 2021, Virtual Event. Cham: Springer; 2022. P. 272–284. https://doi.org/10.1007/978-3-031-08999-2_22

14. Oktay O., Schlemper J., Folgoc L.L., et al. Attention U-Net: Learning where to look for the pancreas. arXiv. URL: https://doi.org/10.48550/arXiv.1804.03999 [Accessed 15th January 2026].

15. Alirr O.I. Dual attention U-Net for liver tumor segmentation in CT images. International Journal of Computers, Communications & Control. 2024;19(2).

16. Dostovalova A.M., Gorshenin A.K., Starichkova J.V., Arzamasov K.M. Comparative analysis of modifications of U-Net neuronal network architectures in medical image segmentation. Digital Diagnostics. 2024;5(4):833–853. https://doi.org/10.17816/DD629866

17. Ma J., He Y., Li F., et al. Segment anything in medical images. Nature Communications. 2024;15. https://doi.org/10.1038/s41467-024-44824-z

18. Rokuss M., Langenberg M., Kirchhoff Y., et al. VoxTell: Free-text promptable universal 3D medical image segmentation. arXiv. URL: https://doi.org/10.48550/arXiv.2511.11450 [Accessed 15th January 2026].

19. Starichkova Yu.V., Pitinov A.V., Gazanova N.Sh. Development of a method for registration of multiphase CT images using affine transformations. In: Medelectronics–2024. Medical electronics and new medical technologies: Proceedings of the XIV International Scientific and Technical Conference, 05–06 December 2024, Minsk, Belarus. Minsk; 2024. P. 237–240. (In Russ.).

20. Liu F., Cai J., Huo Y., et al. JSSR: A joint synthesis, segmentation, and registration system for 3D multi-modal image alignment of large-scale pathological CT scans. In: Computer Vision – ECCV 2020: 16th European Conference: Proceedings: Part XIII, 23–28 August 2020, Glasgow, UK. Cham: Springer; 2020. P. 257–274. https://doi.org/10.1007/978-3-030-58601-0_16

21. Zhou Y., Li Y., Zhang Zh., et al. Hyper-Pairing Network for Multi-Phase Pancreatic Ductal Adenocarcinoma Segmentation. arXiv. URL: https://doi.org/10.48550/arXiv.1905.00367 [Accessed 20th January 2026].

22. Qu T., Wang X., Fang Ch., et al. M3Net: A multi-scale multi-view framework for multi-phase pancreas segmentation based on cross-phase non-local attention. Medical Image Analysis. 2022;75. https://doi.org/10.1016/j.media.2021.102232

23. Lteif D., Appapogu D., Bargal S.A., Plummer B.A., Kolachalama V.B. Anatomy-guided, modality-agnostic segmentation of neuroimaging abnormalities. Human Brain Mapping. 2025;46(14). https://doi.org/10.1002/hbm.70329

24. Antonelli M., Reinke A., Bakas S., et al. The Medical Segmentation Decathlon. Nature Communications. 2022;13. https://doi.org/10.1038/s41467-022-30695-9

25. Wu X., Su H., Hua Y., et al. A multi-phase CT dataset for automated differential diagnosis of liver tumors. Scientific Data. 2026;13. https://doi.org/10.1038/s41597-025-06343-4

26. Bartnik K., Bartczak T., Krzyziński M., et al. WAW-TACE: A hepatocellular carcinoma multiphase CT dataset with segmentations, radiomics features, and clinical data. Radiology: Artificial Intelligence. 2024;6(6).

27. Elbatel M., Yi Q., Huang X., et al. Triphasic-aided Liver Lesion Segmentation in Non-contrast CT (TriALS) Challenge. In: Medical Image Computing and Computer Assisted Intervention 2025 (MICCAI). https://doi.org/10.5281/zenodo.15087646

Samsonenko Stanislav Vladimirovich

MIREA – Russian Technological University

Moscow, Russian Federation

Kashirina Irina Leonidovna
Doctor of Engineering Sciences, Professor

WoS | Scopus | ORCID | eLibrary |

MIREA – Russian Technological University
Voronezh State University

Voronezh, Russian Federation

Keywords: hybrid architectures, image segmentation, attention mechanisms, multiphase CT, feature fusion, medical imaging, deep learning, computer vision, PA-ResSeg, m3Net

For citation: Samsonenko S.V., Kashirina I.L. Deep learning architectures for multiphase CT image segmentation. Modeling, Optimization and Information Technology. 2026;14(3). URL: https://moitvivt.ru/ru/journal/pdf?id=2211 DOI: 10.26102/2310-6018/2026.54.3.012 (In Russ).

16

Full text in PDF

Received 02.02.2026

Revised 16.03.2026

Accepted 24.03.2026