ObjectiveTo systematically summarize recent advancements in the application of artificial intelligence (AI) in key components of radiotherapy (RT), explore the integration of technical innovations with clinical practice, and identify current limitations in real-world implementation. MethodsA comprehensive analysis of representative studies from recent years was conducted, focusing on the technical implementation and clinical effectiveness of AI in image reconstruction, automatic delineation of target volumes and organs at risk, intelligent treatment planning, and prediction of RT-related toxicities. Particular attention was given to deep learning models, multimodal data integration, and their roles in enhancing decision-making processes. ResultsAI-based low-dose image enhancement techniques had significantly improved image quality. Automated segmentation methods had increased the efficiency and consistency of contouring. Both knowledge-driven and data-driven planning systems had addressed the limitations of traditional experience-dependent approaches, contributing to higher quality and reproducibility in treatment plans. Additionally, toxicity prediction models that incorporated multimodal data enabled more accurate, personalized risk assessment, supporting safer and more effective individualized RT. ConclusionsRT is a fundamental modality in cancer treatment. However, achieving precise tumor ablation while minimizing damage to surrounding healthy tissues remains a significant challenge. AI has demonstrated considerable value across multiple technical stages of RT, enhancing precision, efficiency, and personalization. Nevertheless, challenges such as limited model generalizability, lack of data standardization, and insufficient clinical validation persist. Future work should emphasize the alignment of algorithmic development with clinical demands to facilitate the standardized, reliable, and practical application of AI in RT.
The segmentation of dental models is a crucial step in computer-aided diagnosis and treatment systems for oral healthcare. To address the issues of poor universality and under-segmentation in tooth segmentation techniques, an intelligent tooth segmentation method combining multiple seed region growth and boundary extension is proposed. This method utilized the distribution characteristics of negative curvature meshes in teeth to obtain new seed points and effectively adapted to the structural differences between the top and sides of teeth through differential region growth. Additionally, the boundaries of the initial segmentation were extended based on geometric features, which was effectively compensated for under-segmentation issues in region growth. Ablation experiments and comparative experiments with current state-of-the-art algorithms demonstrated that the proposed method achieved better segmentation of crowded dental models and exhibited strong algorithm universality, thus possessing the capability to meet the practical segmentation needs in oral healthcare.
Objective To propose an innovative self-supervised learning method for vascular segmentation in computed tomography angiography (CTA) images by integrating feature reconstruction with masked autoencoding. Methods A 3D masked autoencoder-based framework was developed, where in 3D histogram of oriented gradients (HOG) was utilized for multi-scale vascular feature extraction. During pre-training, random masking was applied to local patches of CTA images, and the model was trained to jointly reconstruct original voxels and HOG features of masked regions. The pre-trained model was further fine-tuned on two annotated datasets for clinical-level vessel segmentation. Results Evaluated on two independent datasets (30 labeled CTA images each), our method achieved superior segmentation accuracy to the supervised neural network U-Net (nnU-Net) baseline, with Dice similarity coefficients of 91.2% vs. 89.7% (aorta) and 84.8% vs. 83.2% (coronary arteries). Conclusion The proposed self-supervised model significantly reduces manual annotation costs without compromising segmentation precision, showing substantial potential for enhancing clinical workflows in vascular disease management.
Corona virus disease 2019 (COVID-19) is an acute respiratory infectious disease with strong contagiousness, strong variability, and long incubation period. The probability of misdiagnosis and missed diagnosis can be significantly decreased with the use of automatic segmentation of COVID-19 lesions based on computed tomography images, which helps doctors in rapid diagnosis and precise treatment. This paper introduced the level set generalized Dice loss function (LGDL) in conjunction with the level set segmentation method based on COVID-19 lesion segmentation network and proposed a dual-path COVID-19 lesion segmentation network (Dual-SAUNet++) to address the pain points such as the complex symptoms of COVID-19 and the blurred boundaries that are challenging to segment. LGDL is an adaptive weight joint loss obtained by combining the generalized Dice loss of the mask path and the mean square error of the level set path. On the test set, the model achieved Dice similarity coefficient of (87.81 ± 10.86)%, intersection over union of (79.20 ± 14.58)%, sensitivity of (94.18 ± 13.56)%, specificity of (99.83 ± 0.43)% and Hausdorff distance of 18.29 ± 31.48 mm. Studies indicated that Dual-SAUNet++ has a great anti-noise capability and it can segment multi-scale lesions while simultaneously focusing on their area and border information. The method proposed in this paper assists doctors in judging the severity of COVID-19 infection by accurately segmenting the lesion, and provides a reliable basis for subsequent clinical treatment.
Accurate segmentation of pediatric echocardiograms is a challenging task, because significant heart-size changes with age and faster heart rate lead to more blurred boundaries on cardiac ultrasound images compared with adults. To address these problems, a dual decoder network model combining channel attention and scale attention is proposed in this paper. Firstly, an attention-guided decoder with deep supervision strategy is used to obtain attention maps for the ventricular regions. Then, the generated ventricular attention is fed back to multiple layers of the network through skip connections to adjust the feature weights generated by the encoder and highlight the left and right ventricular areas. Finally, a scale attention module and a channel attention module are utilized to enhance the edge features of the left and right ventricles. The experimental results demonstrate that the proposed method in this paper achieves an average Dice coefficient of 90.63% in acquired bilateral ventricular segmentation dataset, which is better than some conventional and state-of-the-art methods in the field of medical image segmentation. More importantly, the method has a more accurate effect in segmenting the edge of the ventricle. The results of this paper can provide a new solution for pediatric echocardiographic bilateral ventricular segmentation and subsequent auxiliary diagnosis of congenital heart disease.
Focus on the inconsistency of the shape, location and size of brain glioma, a dual-channel 3-dimensional (3D) densely connected network is proposed to automatically segment brain glioma tumor on magnetic resonance images. Our method is based on a 3D convolutional neural network frame, and two convolution kernel sizes are adopted in each channel to extract multi-scale features in different scales of receptive fields. Then we construct two densely connected blocks in each pathway for feature learning and transmission. Finally, the concatenation of two pathway features was sent to classification layer to classify central region voxels to segment brain tumor automatically. We train and test our model on open brain tumor segmentation challenge dataset, and we also compared our results with other models. Experimental results show that our algorithm can segment different tumor lesions more accurately. It has important application value in the clinical diagnosis and treatment of brain tumor diseases.
The count and recognition of white blood cells in blood smear images play an important role in the diagnosis of blood diseases including leukemia. Traditional manual test results are easily disturbed by many factors. It is necessary to develop an automatic leukocyte analysis system to provide doctors with auxiliary diagnosis, and blood leukocyte segmentation is the basis of automatic analysis. In this paper, we improved the U-Net model and proposed a segmentation algorithm of leukocyte image based on dual path and atrous spatial pyramid pooling. Firstly, the dual path network was introduced into the feature encoder to extract multi-scale leukocyte features, and the atrous spatial pyramid pooling was used to enhance the feature extraction ability of the network. Then the feature decoder composed of convolution and deconvolution was used to restore the segmented target to the original image size to realize the pixel level segmentation of blood leukocytes. Finally, qualitative and quantitative experiments were carried out on three leukocyte data sets to verify the effectiveness of the algorithm. The results showed that compared with other representative algorithms, the proposed blood leukocyte segmentation algorithm had better segmentation results, and the mIoU value could reach more than 0.97. It is hoped that the method could be conducive to the automatic auxiliary diagnosis of blood diseases in the future.
Objective To develop a neural network architecture based on deep learning to assist knee CT images automatic segmentation, and validate its accuracy. Methods A knee CT scans database was established, and the bony structure was manually annotated. A deep learning neural network architecture was developed independently, and the labeled database was used to train and test the neural network. Metrics of Dice coefficient, average surface distance (ASD), and Hausdorff distance (HD) were calculated to evaluate the accuracy of the neural network. The time of automatic segmentation and manual segmentation was compared. Five orthopedic experts were invited to score the automatic and manual segmentation results using Likert scale and the scores of the two methods were compared. Results The automatic segmentation achieved a high accuracy. The Dice coefficient, ASD, and HD of the femur were 0.953±0.037, (0.076±0.048) mm, and (3.101±0.726) mm, respectively; and those of the tibia were 0.950±0.092, (0.083±0.101) mm, and (2.984±0.740) mm, respectively. The time of automatic segmentation was significantly shorter than that of manual segmentation [(2.46±0.45) minutes vs. (64.73±17.07) minutes; t=36.474, P<0.001). The clinical scores of the femur were 4.3±0.3 in the automatic segmentation group and 4.4±0.2 in the manual segmentation group, and the scores of the tibia were 4.5±0.2 and 4.5±0.3, respectively. There was no significant difference between the two groups (t=1.753, P=0.085; t=0.318, P=0.752). Conclusion The automatic segmentation of knee CT images based on deep learning has high accuracy and can achieve rapid segmentation and three-dimensional reconstruction. This method will promote the development of new technology-assisted techniques in total knee arthroplasty.
Aiming at the problems of low accuracy and large difference of segmentation boundary distance in anterior cruciate ligament (ACL) image segmentation of knee joint, this paper proposes an ACL image segmentation model by fusing dilated convolution and residual hybrid attention U-shaped network (DRH-UNet). The proposed model builds upon the U-shaped network (U-Net) by incorporating dilated convolutions to expand the receptive field, enabling a better understanding of the contextual relationships within the image. Additionally, a residual hybrid attention block is designed in the skip connections to enhance the expression of critical features in key regions and reduce the semantic gap, thereby improving the representation capability for the ACL area. This study constructs an enhanced annotated ACL dataset based on the publicly available Magnetic Resonance Imaging Network (MRNet) dataset. The proposed method is validated on this dataset, and the experimental results demonstrate that the DRH-UNet model achieves a Dice similarity coefficient (DSC) of (88.01±1.57)% and a Hausdorff distance (HD) of 5.16±0.85, outperforming other ACL segmentation methods. The proposed approach further enhances the segmentation accuracy of ACL, providing valuable assistance for subsequent clinical diagnosis by physicians.
Lung diseases such as lung cancer and COVID-19 seriously endanger human health and life safety, so early screening and diagnosis are particularly important. computed tomography (CT) technology is one of the important ways to screen lung diseases, among which lung parenchyma segmentation based on CT images is the key step in screening lung diseases, and high-quality lung parenchyma segmentation can effectively improve the level of early diagnosis and treatment of lung diseases. Automatic, fast and accurate segmentation of lung parenchyma based on CT images can effectively compensate for the shortcomings of low efficiency and strong subjectivity of manual segmentation, and has become one of the research hotspots in this field. In this paper, the research progress in lung parenchyma segmentation is reviewed based on the related literatures published at domestic and abroad in recent years. The traditional machine learning methods and deep learning methods are compared and analyzed, and the research progress of improving the network structure of deep learning model is emphatically introduced. Some unsolved problems in lung parenchyma segmentation were discussed, and the development prospect was prospected, providing reference for researchers in related fields.