Skip to main content
Erschienen in: Journal of Orthopaedic Surgery and Research 1/2024

Open Access 01.12.2024 | Research article

Supraspinatus extraction from MRI based on attention-dense spatial pyramid UNet network

verfasst von: Peng Wang, Yang Liu, Zhong Zhou

Erschienen in: Journal of Orthopaedic Surgery and Research | Ausgabe 1/2024

Abstract

Background

With potential of deep learning in musculoskeletal image interpretation being explored, this paper focuses on the common site of rotator cuff tears, the supraspinatus. It aims to propose and validate a deep learning model to automatically extract the supraspinatus, verifying its superiority through comparison with several classical image segmentation models.

Method

Imaging data were retrospectively collected from 60 patients who underwent inpatient treatment for rotator cuff tears at a hospital between March 2021 and May 2023. A dataset of the supraspinatus from MRI was constructed after collecting, filtering, and manually annotating at the pixel level. This paper proposes a novel A-DAsppUnet network that can automatically extract the supraspinatus after training and optimization. The analysis of model performance is based on three evaluation metrics: precision, intersection over union, and Dice coefficient.

Results

The experimental results demonstrate that the precision, intersection over union, and Dice coefficients of the proposed model are 99.20%, 83.38%, and 90.94%, respectively. Furthermore, the proposed model exhibited significant advantages over the compared models.

Conclusion

The designed model in this paper accurately extracts the supraspinatus from MRI, and the extraction results are complete and continuous with clear boundaries. The feasibility of using deep learning methods for musculoskeletal extraction and assisting in clinical decision-making was verified. This research holds practical significance and application value in the field of utilizing artificial intelligence for assisting medical decision-making.
Hinweise

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Abkürzungen
RCTs
Otator cuff tears
CAD
Computer-aided diagnosis
DL
Deep learning
A-DAsppUnet
Attention-dense atrous spatial pyramid pooling UNet
DASPP
Dense atrous spatial pyramid pooling
DenseNet
Dense convolutional network
PD
Proton density
FCN
Fully convolution network
SegNet
Semantic segmentation network
BCELoss
Binary cross-entropy loss function
Pre
Precision
IoU
Intersection over union
CNN
Convolutional neural network

Background

The rotator cuff is composed of the subscapularis, supraspinatus, infraspinatus, and teres minor tendons. This structure connects the scapula to the humeral head and maintains dynamic stability of the glenohumeral joint through a concave compression mechanism. It is also an essential component in maintaining the equilibrium of couples in the shoulder joint [1]. Rotator cuff tears (RCTs) can cause pain and limitation of shoulder motion, with the supraspinatus tendon being the most commonly affected site. A recent study found that RCTs account for approximately 50%–85% of shoulder disorders treated by clinicians, and the morbidity increases with age [2].
The rotator cuff tendinopathy has been classically summarized as extrinsic, intrinsic, or a combination of both. Intrinsic mechanisms, such as the mechanical properties, age-related degeneration, and vascularity of the rotator cuff, along with extrinsic mechanisms such as internal and external impingement caused by alterations in scapular and glenohumeral kinematics, appear to be significant contributors to RCTs [3]. Tears most commonly occur in and around the critical zone of the supraspinatus tendon, which lies in the region between the bony insertion of the tendon and the nearest musculotendinous junction [4]. This anatomic factor combined with multiple internal and external mechanisms contributes to this result, such as the morphology of supraspinatus [5], subacromial impingement [6], and the presence of “critical zone” [7, 8]. The high incidence of supraspinatus tear gives its segmentation a higher priority and considerable clinical significance for the diagnosis of RCTs.
In clinical practice, shoulder magnetic resonance imaging (MRI) plays a crucial role in diagnosing RCTs, assessing the extent of tears, and formulating surgical plans. MRI offers advantages such as non-invasiveness, non-ionizing, anatomical reproducibility, and excellent tissue contrast, making it a common modality for the clinical diagnosis of RCTs and preoperative preparation [9].
Currently, computer-aided diagnosis (CAD) techniques have been widely applied in medical image analysis, significantly enhancing diagnostic accuracy and efficiency. With the advent of the artificial intelligence (AI) revolution, AI-enabled health care has become a hot research field. Targeting the issues of low efficiency in the interpretation of massive MRI images and subjective differences in human interpretation, this paper proposes deep learning (DL) methods and builds an innovative DL model based on existing research findings. It was developed to automatically segment and extract regions of interest, aiming to alleviate the workload of clinical doctors.
The supraspinatus is the most common site of RCTs, and its tears along with atrophy can reflect the severity of damage. Therefore, the supraspinatus is chosen as the segmentation target, and an improved DL model that can accurately extract the supraspinatus in the coronal plane was proposed in this paper. Compared to the more extensively studied sagittal plane, the coronal plane is a vertical plane perpendicular to the body. It is commonly used to observe the anterior–posterior thickness and morphology of the supraspinatus muscle in the shoulder. The coronal plane is particularly useful in assessing tears or changes in the anterior–posterior thickness of the supraspinatus muscle. Therefore, compared to segmentation based on the sagittal plane, extracting the supraspinatus muscle based on the coronal plane can improve the efficiency of diagnosing and treating RCTs. It has a significant impact on clinical decision-making and the formulation of surgical plans. This innovation holds certain practical significance in the field of intelligent recognition and interpretation of medical images.

Methods

Shoulder MRI has relatively simple semantics and fixed structures. Both high-level and low-level semantic information are equally important, and there is a high demand for timeliness in medical diagnosis. Therefore, image segmentation algorithms are commonly used in research to improve segmentation results and accuracy while reducing manual intervention and segmentation time. Considering these characteristics, this paper chooses the LinkNet [10] model as the base framework. LinkNet has demonstrated good performance in achieving accurate segmentation results. It utilizes a combination of encoder and decoder structure to capture local and global context information, which is crucial for accurately segmenting the supraspinatus muscle and distinguishing it from surrounding tissues. LinkNet strikes a balance between accuracy and efficiency based on a lightweight network structure. It computes efficiently while still maintaining competitive segmentation performance. This is especially valuable in clinical settings for real-time or near real-time MRI image analysis. The LinkNet shows robustness in handling image quality and noise variations. The model’s structure helps mitigate the impact of noise and artifacts, resulting in more reliable segmentation results, showing. This robustness is essential for accurate segmentation of the supraspinatus muscle. And LinkNet can be pre-trained on large datasets, such as medical image repositories, to enhance model generalization and improve segmentation performance. Considering these factors, the LinkNet model is a suitable choice for accurate and efficient supraspinatus muscle segmentation in MRI images. However, the MRI images of the supraspinatus encompass intricate details and local features, accurate segmentation is challenging, and the selection and improvement of any segmentation model depends on the specific requirements of the task. This paper constructs an attention-dense atrous spatial pyramid pooling UNet (A-DAsppUnet) network for the segmentation of the supraspinatus in shoulder MRI. As shown in Fig. 1, the proposed model involves an encoder ResNet34 [11], a channel attention module, and dense atrous spatial pyramid pooling (DenseASPP), which connects the encoder and decoder network. The encoder can extract deep semantic feature information, while the channel attention incorporates skip connections to enhance feature representation during encoding and decoding. Drawing from the structural experience of the D-LinkNet model [12], DenseASPP is beneficial to capture multi-context information, intensive feature extraction, and parameter sharing and improves the accuracy of semantic segmentation. It is widely used in object segmentation [13] and scene semantic recognition [14, 15]. The aforementioned structures were integrated into the model and had been innovatively applied to muscle tissue segmentation in medical MRI images. The proposed model demonstrates the ability to resist noise and image quality interference, enabling efficient and accurate segmentation of the supraspinatus and surrounding tissues.
As shown in the diagram, the selected sequences were downloaded and exported as TIFF files and saved across three RGB image channels, which were adjusted to 8-bit 512 × 512 × 3-pixel portable network graphics (PNG) files using Photoshop to match the standardized network input. Compared with the original grayscale image, which only contains brightness information, RGB image helps to separate the supraspinatus muscle from the surrounding tissue. Due to the inclusion of three information channels, RGB images facilitate the differentiation of certain muscle diseases or conditions that may result in color variations in muscle tissue. Moreover, the richness of contextual information in RGB images enhances their visualization effect, making them more intuitive and suitable for specific algorithms. The ResNet34 model utilizes multiple down-sampling steps to extract the desired target features. At the end of the encoding process, the image dimensions are reduced to a size of 16 × 16 with 512 channels. Subsequently, the feature map is passed into the DenseASPP module. This step is beneficial for expanding the receptive field without compromising the resolution of the feature map. Additionally, it ensures the preservation of abundant spatial information. In the decoding stage, the feature map size is restored through transposed convolution for up-sampling. The model utilizes skip connections and channel attention modules to fuse and complement the feature information, enhancing both the integrity of the feature information and the exchange of channel features. This approach significantly improves the network's capability to extract target regions in complex MRI images, ensuring high accuracy and robustness in feature extraction.

Channel attention module

In medical image segmentation, extracting structural features of target regions is often challenging. Additionally, the performance and stability of medical image segmentation models are often compromised due to the lack of high-quality manually labeled datasets and class imbalance among the samples. Attention mechanisms have demonstrated their effectiveness in enhancing a model's ability to focus on important features [16]. In this paper, a channel attention mechanism is introduced to adjust the importance of each channel in the feature maps after encoding and down-sampling. This mechanism dynamically adjusts the network's attention to different features, thereby effectively enhancing feature extraction and utilization. It helps alleviate class imbalance issues, reduce noise, and eliminate redundant information. As a result, it improves the robustness of the model and enhances segmentation accuracy [17]. Figure 2 illustrates the structure of the channel attention module.

Dense atrous spatial pyramid pooling

In the middle part of the model, the DenseASPP structure based on the dense convolutional network (DenseNet) [19] model is used to connect the encoding and decoding networks. Figure 3 shows the structure of DenseASPP, it utilizes multiple branches of different void convolution kernels to extract multi-scale features from the input data. In medical image segmentation research, accurate segmentation of the target region is a crucial performance metric. The DenseASPP module expands the receptive field and adapts to multi-scale input images by utilizing dilated convolutions and pyramid pooling within a dense block structure. In Fig. 3, d represents the dilation rate of dilated convolutions. This module enhances the semantic expressive power of features and exhibits outstanding performance in image segmentation tasks.

Experiment and analysis

Experimental data and comparative models

Experimental data

This paper was approved by the Ethics Committee of Jiangsu Province Hospital with Integration of Chinese and Western Medicine, and the approval number is 2023-LWKYZ-033. Personal informed consent was waived for this retrospective study. It retrospectively collected data from 60 patients who underwent inpatient treatment for RCTs at the hospital between March 2021 and May 2023. Patients with other shoulder conditions, such as fractures, dislocations, and calcific tendinitis, were excluded from the study.
Examinations were acquired with a 1.5 T MR scanner (General Electric, SIGNA CREATOR). Conventional two-dimensional MR images were obtained from the proton density (PD) fat-suppressed sequence in the oblique coronal plane. The acquisition parameters are as follows: TR = 2278 + ms, TE = 12.6–84.2 ms, FOV = 20 cm, NEX = 2, bandwidth = 31.25 Hz/pixel, slice thickness = 4 mm, and spacing: 0.5 mm.
Due to the physiological differences between individuals and the diverse location of tears, the author selected images capable of displaying the supraspinatus clearly from acquired images, about 3–5 images per sequence. After the selection process, a total of 200 MRI images were chosen for further analysis. The 60 subjects were randomly divided into a training set, a validation set, and a test set, ensuring that images from the same subject in the training dataset were not used in the validation or evaluation processes in this paper. To fully train the model in this paper, the experimental data were expanded threefold using data augmentation techniques, including image rotation, horizontal flipping, and vertical flipping. These techniques aim to enhance the robustness and generalization capability of the model, thereby improving the accuracy of target extraction during model training.
The images were in RGB format with a size of 512 × 512 pixels. The supraspinatus was manually annotated by tracing its contour on the images. The proximal end of the annotation started at the scapular spine, while the distal end ended at the greater tuberosity of the humerus. The superior boundary was defined by the acromion, shoulder joint capsule, and trapezius, while the inferior boundary was determined by the scapular spine, the upper aspect of the humeral head, and the supraglenoid tubercle. These annotations output corresponding labels for the supraspinatus tendon. The data were annotated by three graduate students and physicians specialized in musculoskeletal imaging. The annotations underwent verification by experienced physicians to ensure accuracy and reliability. The original images and extended data are shown in Fig. 4.

Comparative models

To assess the feasibility and high accuracy of the proposed model for segmenting the supraspinatus tendon in MRI images, several classic image segmentation algorithms, including fully convolution network (FCN), UNet, semantic segmentation network (SegNet), and DenseNet, were selected for comparative experiments on the dataset employed in this paper. These models were chosen to evaluate the accuracy and performance of the proposed method against established approaches.

Experimental environment and evaluation metrics

Experimental environment

The experiments were conducted using the Python 3.7 programming language and the PyTorch 1.8.1 deep learning framework. All experiments were performed on a computer equipped with an AMD Ryzen 7 3700X CPU and an NVIDIA GeForce RTX 2700 graphics card with 8 GB of VRAM. For model training, the binary cross-entropy loss function (BCELoss) was used, along with the Adam optimizer to update the network parameters. The batch size was set to 2, and the learning rate was set to 0.0001. The models were trained for 30 epochs.

Evaluation metrics

Image segmentation is evaluated using metrics such as precision (Pre), F1 score, and intersection over union (IoU). However, in medical image segmentation, the Dice coefficient is often used to assess model performance. Therefore, this paper uses Pre, the Dice coefficient, and IoU as the evaluation metrics to measure model performance and the accuracy of supraspinatus tendon segmentation. The formulas for these evaluation metrics are as follows:
$${\text{Precision}} = \frac{{{\text{TP}}}}{{{\text{TP}} + {\text{FP}}}} \times 100\%$$
(1)
$${\text{IoU}} = \frac{{{\text{TP}}}}{{{\text{TP}} + {\text{FP}} + {\text{FN}}}} \times 100\%$$
(2)
$${\text{Dice}} = \frac{{2{\text{TP}}}}{{\left( {{\text{TP}} + {\text{FP}}} \right) + \left( {{\text{TP}} + {\text{FN}}} \right)}} \times 100\%$$
(3)
In the equations, TP represents the number of pixels correctly predicted as supraspinatus tendon, FP represents the number of pixels incorrectly predicted as supraspinatus tendon, TN represents the number of pixels correctly predicted as background, and FN represents the number of pixels incorrectly predicted as background.

Experimental results and analysis

To ensure fair and objective analysis of the experimental results, all experiments in this paper were conducted using the same dataset and experimental environment. Table 1 displays the quantitative statistical results of the five models on the experiment's test set for supraspinatus tendon extraction. The evaluation of model performance is conducted using three assessment metrics: Pre, IoU, and Dice coefficient.
Table 1
Extraction result statistics
Model
Pre (%)
IoU (%)
Dice (%)
FCN
98.28
65.07
78.84
SegNet
98.53
70.59
82.76
UNet
98.74
73.61
84.80
DenseNet
98.79
75.86
86.27
A-DAsppUnet
99.20
83.38
90.94
Bold values indicate the performance of the proposed method over the comparison method
According to the table, the mentioned models are capable of extracting the supraspinatus tendon to some extent, with differences in terms of completeness, continuity, and accuracy of the extraction. The proposed method in this paper achieved Pre, IoU, and Dice coefficients of 99.20%, 83.38%, and 90.94%, respectively. The comparison clearly indicates that this method has a significant advantage and performs in terms of extracting the supraspinatus tendon. Compared to the four comparative models, the proposed method exhibited improvements in the evaluation metrics. The “Pre” metric showed an enhancement of approximately 0.4%–1%, the “IoU” metric witnessed an improvement of 7.5%–18.3%, and the "Dice coefficient" experienced an increase of approximately 4.7%–12.1%. These improvements were significant across all indicators. Among the comparative models, DenseNet performed the best, followed by UNet and SegNet, while FCN had the worst effect and had a large gap with the proposed algorithm in IoU and Dice indicators in this paper.
Based on the data in the table and the equation above, it can be observed that the improvement in the Pre metric is not significant. The reason is that the target pixel constitutes a relatively small proportion of the total number of pixels, and reducing false-positive pixels does not lead to a significant increase in accuracy. According to the IoU indicator and its calculation formula, a higher IoU value signifies a larger proportion of accurately classified supraspinatus pixels relative to the total number of correctly classified pixels, with fewer incorrectly predicted pixels. In the supraspinatus segmentation task, the significant improvement in the IoU measure indicates that the proposed model achieves the highest accuracy in supraspinatus segmentation. In addition, with a Dice coefficient of 90.94%, it can be inferred that the performance of the model proposed in this paper is superior.
To fully validate the above conclusions, this paper conducted a visual analysis of the supraspinatus segmentation results. Figure 5 illustrates a visual comparison between the results obtained using the proposed method and those obtained using the comparative model on the test set images. Four representative images are provided in the figure for comparison. As shown in the figure, the segmentation results obtained using the proposed method exhibit the best performance in terms of completeness and capturing fine details. Specifically, the extracted supraspinatus is delineated clearly from structures such as the humeral tuberosity, scapular spine, and inferior glenohumeral capsule. The segmentation results exhibit well-preserved details, and there is a high level of accuracy in aligning the upper and lower boundaries of the segmented region with the ground truth labels. In contrast, in the results obtained using the comparative model, the boundaries of the supraspinatus extraction are blurred near the side of the scapular spine in images (1) and (3). Image (1) shows poor overall segmentation performance, with a significant portion of supraspinatus pixels left unsegmented. Image (3) exhibits numerous erroneous segmentations. On the other hand, in images (2) and (4), the proposed method accurately delineated the edges of the supraspinatus, particularly at the tendon junction with the humeral head and the superior border of the deltoid muscle.
In the comparison model, DenseNet performs well in extracting the target and achieves reasonably accurate segmentation. However, it fails to capture fine details, especially in capturing the blurry boundaries with the deltoid muscle, resulting in insufficient accuracy. The UNet model lacks completeness in target extraction. For example, in image (1), there is information missing in the proximal part of the supraspinatus, and the extracted region is smaller than the actual boundaries defined by the labels. For the SegNet and FCN models, their segmentation results exhibit more false positives and false negatives, as shown in images (1) and (3).

Robustness analysis through ablation experiments

To thoroughly validate the effectiveness of the proposed innovative model, it conducted ablation experiments to investigate the impact of the deep encoding network, channel attention, and dense spatial pyramid pooling modules on the performance of the proposed model. The experimental setup is described as follows:
LinkNet18 was selected as the baseline model in this paper. Scheme 1: ResNet34 is selected as the model encoder network. Scheme 2: Add the channel attention mechanism at the jump junction of the Scheme 1 model. Scheme 3: DenseASPP in the middle of the Scheme 1 model to connect the encoder and decoder networks. The models of the ablation experiment are shown in Table 2.
Table 2
Ablation experimental model
Model
ResNet34
Channel attention
DenseASPP
Baseline
   
Scheme 1
  
Scheme 2
 
Scheme 3
 
A-DAsppUnet
The ablation experiments were carried out in the same environment as the experimental dataset. The extraction results of the supraspinatus from the images of the test set by each ablation model are shown in Table 3.
Table 3
Extraction results of ablation experiments
Model
Pre (%)
IoU (%)
Dice (%)
Baseline
89.03
73.66
84.83
Scheme 1
91.71
75.91
86.3
Scheme 2
89.01
76.85
86.91
Scheme 3
90.24
78.11
87.71
A-DAsppUnet
92.77
83.38
90.94
Bold values indicate the performance of the proposed method over the comparison method
Figure 6 shows the extraction results of the supraspinatus in the test set images of each ablation model. According to the comprehensive table and Fig. 6, the Pre, IoU, and Dice indexes of Scheme 1 increased by 2.68%, 2.25%, and 1.47%, respectively, and the extraction integrity of the supraspinatus was improved by the model. For example, the extraction results of images (2) and (3) are complete, continuous, and clear. In Scheme 2, the channel attention mechanism is added on the basis of Scheme 1, and the IoU and Dice coefficients are increased by 0.94% and 0.61%, respectively. The channel attention mechanism enhances the fusion of important features in the jump connection, thereby improving the accuracy of supraspinatus edge extraction.
In Scheme 3, the DenseASPP module is used as the middle part of connection coding–decoding, and the model performance indexes Pre, IoU, and Dice are increased by 1.21%, 4.45%, and 2.88%, respectively. The DenseASPP module extends the receptive field of the down-sampled feature maps obtained from the encoder without reducing their resolution. It preserves rich feature information and effectively helps the model recognize and extract target regions after up-sampling in the decoder. This module achieved the best results in terms of the integrity, accuracy, and clarity of supraspinatus edge extraction.

Discussion

With the advancement of medical imaging technology, the quantity and complexity of medical imaging data are continuously increasing. In most cases, even with access to shoulder MRI, nonorthopedic surgeons may find it challenging to identify and diagnose RCTs. In this case, the application of CAD techniques provides support for ensuring high efficiency and accuracy in clinical diagnosis.
Ledley et al. [20] pioneered the field of CAD by building a mathematical model for lung cancer diagnosis. With the emergence of artificial intelligence, CAD has evolved into a DL approach, which has shown great potential and widespread application in image processing and computer vision. DL models have revolutionized the field of medical image analysis by leveraging their ability to extract complex patterns and features from images. Through training on extensive datasets, these models can learn to identify subtle abnormalities, assist in disease diagnosis, and provide valuable insights for clinical work.
DL methods have made significant contributions to medical imaging research, with representative works including brain tumor segmentation [21], lung nodule detection [22], and case image segmentation [23]. In the field of musculoskeletal imaging, accurate imaging diagnosis is crucial, which has spurred the vigorous development of DL techniques. Research in this domain encompasses various areas, such as knee cartilage injury [24], meniscus and ligament tears [25, 26], spinal canal stenosis [27], bone age detection, and osteoporosis diagnosis [28], all of which have achieved fruitful results. In this context, the focus is on shoulder MRI, where the mature technologies primarily concentrate on the segmentation of bony tissues. However, the extraction of imaging features related to musculoskeletal tissue, as well as research on their role in assisting diagnosis, is still under development. Research on robust and accurate algorithms for the segmentation and analysis of these soft tissues in shoulder MRI holds great potential for improving diagnostic accuracy and facilitating treatment planning in orthopedics.
Indeed, DL research based on shoulder MRI has made significant progress. Kim et al. [29] developed a FCN model for the segmentation of the supraspinatus and supraspinatus fossa in the sagittal plane of MRI, which visualizes the degree of supraspinatus atrophy and fatty infiltration. Medina et al. [30] utilized an improved UNet convolutional neural network (CNN) architecture to accurately segment the supraspinatus, infraspinatus, and subscapularis in sagittal plane MRI. Ro et al. [31] employed a CNN-based approach to segment the supraspinatus and supraspinatus fossa. They analyzed the occupation rate of the supraspinatus and utilized an improved Otsu thresholding technique to quantify the extent of fatty infiltration in the supraspinatus. These studies, focusing on sagittal plane of shoulder MRI, enable physicians to accurately assess the degree of supraspinatus atrophy and fatty infiltration and predict the effectiveness of rotator cuff repair surgery.
However, RCTs are primarily categorized as tendinopathy. Only knowing the atrophy and fatty infiltration of the supraspinatus has limited clinical significance. Therefore, the current trend is to study the tendons themselves. Yao et al. [32] employed a three-stage pipeline consisting of ResNet, UNet, and CNN to perform screening, segmentation, and binary classification (tear or no tear) of supraspinatus images. Hess et al. [33] utilized nnUNet to segment both the bony structures (humerus and scapula) and the rotator cuff on a shoulder MR T1-weighted sequence. Lin et al. [34] used four parallel 3D ResNet50 convolutional neural network architectures to detect and classify RCTs based on tear types.
This paper focuses on the supraspinatus and constructs an A-DAsppUnet model, attempting to segment the supraspinatus in the same MRI sequence. Compared with the results of other segmentation models, the proposed model has better segmentation accuracy and performance. It validated the feasibility of using DL methods for segmenting the rotator cuff, and the results provide a reference for clinical treatment and surgical planning in this paper.
However, it is important to acknowledge the limitations of the study. Although the data volume was increased through data augmentation, the experimental data in this study are still not abundant, the prediction results may have minor errors in displaying subtle tears. The training set images only outline the contour of the supraspinatus, so the model prediction results cannot reflect internal injuries and tendon quality. Full-thickness tears mean continuous interruptions of the rotator cuff, leading to significant errors in the segmentation of the tendon stump. Additionally, the boundaries between the supraspinatus and adjacent muscles, such as the trapezius, appear unsatisfactory due to the similarity in pixel grayscale values on MRI. Furthermore, it excluded cases affected by other shoulder diseases, limiting the clinical utility of this model. Addressing the aforementioned issue and expanding the dataset to encompass a broader range of cases would enhance the model's generalization capability.

Conclusion

In this study, it aimed to investigate the effectiveness of DL models for the extraction of the supraspinatus from shoulder MRI. An improved DL network model was designed, and extensive experiments were carried out on self-constructed supraspinatus dataset.
The experimental results demonstrated that the proposed improved DL model has excellent performance in extracting the supraspinatus the coronal plane of shoulder MRI. The model achieved a high segmentation accuracy with Dice coefficient, precision, and IoU of 0.91, 0.99, and 0.83, respectively. These results indicate that the DL method is capable of accurately segmenting the supraspinatus in shoulder MRI.
Furthermore, the analysis revealed several advantages of the model. The proposed model demonstrates robustness to variations in the position and shape of the supraspinatus. It exhibits resistance to noise interference and achieves high-quality and complete extraction. Compared to traditional image processing techniques, the model outperforms them and shows greater potential in clinical research and applications.
DL-based image segmentation has several advantages compared to the detection and classification of RCTs. Image segmentation offers more detailed information and supports quantitative analysis. It accurately delineates structures or abnormalities at the pixel level, enabling precise localization and providing rich anatomical and pathological details. Therefore, DL-based image segmentation is better suited for handling complex scenarios and personalized medical interventions.
In summary, the research demonstrates the effectiveness of DL models in extracting the supraspinatus from the coronal plane of shoulder MRI. This validates the experimental value and practical significance of DL methods in assisting medical decision-making. Future studies can make breakthroughs by continuously exploring attention mechanisms and multi-scale structures, such as dilated convolutions, and utilizing high-quality data from multiple centers, fully harnessing the potential of DL methods in musculoskeletal imaging.

Acknowledgements

Thanks to the physicians and masters who contributed to the development of our dataset.

Declarations

This paper was approved by the Ethics Committee of Jiangsu Province Hospital with Integration of Chinese and Western Medicine. The approval number is 2023-LWKYZ-033. Personal informed consent was waived for this retrospective study.
Written informed consent for publication was obtained from all participants.

Competing interests

The authors declare that they have no conflict of interest.
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://​creativecommons.​org/​licenses/​by/​4.​0/​. The Creative Commons Public Domain Dedication waiver (http://​creativecommons.​org/​publicdomain/​zero/​1.​0/​) applies to the data made available in this article, unless otherwise stated in a credit line to the data.

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Literatur
1.
Zurück zum Zitat Huegel J, Williams AA, Soslowsky LJ. Rotator cuff biology and biomechanics: a review of normal and pathological conditions. Curr Rheumatol Rep. 2015;17:1–9.CrossRef Huegel J, Williams AA, Soslowsky LJ. Rotator cuff biology and biomechanics: a review of normal and pathological conditions. Curr Rheumatol Rep. 2015;17:1–9.CrossRef
2.
Zurück zum Zitat Doiron-Cadrin P, Lafrance S, Saulnier M, Cournoyer É, Roy J-S, Dyer J-O, et al. Shoulder rotator cuff disorders: a systematic review of clinical practice guidelines and semantic analyses of recommendations. Arch Phys Med Rehabil. 2020;101:1233–42.CrossRefPubMed Doiron-Cadrin P, Lafrance S, Saulnier M, Cournoyer É, Roy J-S, Dyer J-O, et al. Shoulder rotator cuff disorders: a systematic review of clinical practice guidelines and semantic analyses of recommendations. Arch Phys Med Rehabil. 2020;101:1233–42.CrossRefPubMed
3.
Zurück zum Zitat Seitz AL, McClure PW, Finucane S, Boardman ND III, Michener LA. Mechanisms of rotator cuff tendinopathy: intrinsic, extrinsic, or both? Clin Biomech. 2011;26:1–12.CrossRef Seitz AL, McClure PW, Finucane S, Boardman ND III, Michener LA. Mechanisms of rotator cuff tendinopathy: intrinsic, extrinsic, or both? Clin Biomech. 2011;26:1–12.CrossRef
4.
Zurück zum Zitat Mehta S, Gimbel JA, Soslowsky LJ. Etiologic and pathogenetic factors for rotator cuff tendinopathy. Clin Sports Med. 2003;22:791–812.CrossRefPubMed Mehta S, Gimbel JA, Soslowsky LJ. Etiologic and pathogenetic factors for rotator cuff tendinopathy. Clin Sports Med. 2003;22:791–812.CrossRefPubMed
5.
Zurück zum Zitat Clark JM, Harryman DT 2nd. Tendons, ligaments, and capsule of the rotator cuff. Gross and microscopic anatomy. JBJS. 1992;74:713–25.CrossRef Clark JM, Harryman DT 2nd. Tendons, ligaments, and capsule of the rotator cuff. Gross and microscopic anatomy. JBJS. 1992;74:713–25.CrossRef
6.
Zurück zum Zitat Charles S, Neer I. Anterior acromioplasty for the chronic impingement syndrome in the shoulder: a preliminary report. JBJS. 1972;54:41–50.CrossRef Charles S, Neer I. Anterior acromioplasty for the chronic impingement syndrome in the shoulder: a preliminary report. JBJS. 1972;54:41–50.CrossRef
8.
Zurück zum Zitat Adler RS, Fealy S, Rudzki JR, Kadrmas W, Verma NN, Pearle A, et al. Rotator cuff in asymptomatic volunteers: contrast-enhanced US depiction of intratendinous and peritendinous vascularity. Radiology. 2008;248:954–61.CrossRefPubMed Adler RS, Fealy S, Rudzki JR, Kadrmas W, Verma NN, Pearle A, et al. Rotator cuff in asymptomatic volunteers: contrast-enhanced US depiction of intratendinous and peritendinous vascularity. Radiology. 2008;248:954–61.CrossRefPubMed
9.
Zurück zum Zitat Zhu Q. Normal anatomy and related pathological changes of shoulder on MRI. Chin J Surg. 2000;38(4):259–62.PubMed Zhu Q. Normal anatomy and related pathological changes of shoulder on MRI. Chin J Surg. 2000;38(4):259–62.PubMed
10.
Zurück zum Zitat Chaurasia A, Culurciello E. LinkNet: exploiting encoder representations for efficient semantic segmentation. 2017 IEEE Computer Vision and Pattern Recognition (CVPR) [Internet]. 2017 [cited 2022 Dec 4]. p. 1–4. Available from: http://arxiv.org/abs/1707.03718. Chaurasia A, Culurciello E. LinkNet: exploiting encoder representations for efficient semantic segmentation. 2017 IEEE Computer Vision and Pattern Recognition (CVPR) [Internet]. 2017 [cited 2022 Dec 4]. p. 1–4. Available from: http://​arxiv.​org/​abs/​1707.​03718.​
12.
Zurück zum Zitat Zhou L, Zhang C, Wu M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. 2018 IEEE/CVF conference on computer vision and pattern recognition workshops (CVPRW) [Internet]. Salt Lake City, UT, USA: IEEE; 2018 [cited 2022 Oct 26]. p. 192–1924. Available from: https://ieeexplore.ieee.org/document/8575492/. Zhou L, Zhang C, Wu M. D-LinkNet: LinkNet with pretrained encoder and dilated convolution for high resolution satellite imagery road extraction. 2018 IEEE/CVF conference on computer vision and pattern recognition workshops (CVPRW) [Internet]. Salt Lake City, UT, USA: IEEE; 2018 [cited 2022 Oct 26]. p. 192–1924. Available from: https://​ieeexplore.​ieee.​org/​document/​8575492/​.​
13.
Zurück zum Zitat Kim M, Ilyas N, Kim K. AMSASeg: an attention-based multi-scale atrous convolutional neural network for real-time object segmentation from 3D point cloud. IEEE Access. 2021;9:70789–96.CrossRef Kim M, Ilyas N, Kim K. AMSASeg: an attention-based multi-scale atrous convolutional neural network for real-time object segmentation from 3D point cloud. IEEE Access. 2021;9:70789–96.CrossRef
14.
Zurück zum Zitat Yang M, Yu K, Zhang C, Li Z, Yang K. DenseASPP for semantic segmentation in street scenes. In: 2018 IEEE/CVF conference on computer vision and pattern recognition. 2018. p. 3684–92. Yang M, Yu K, Zhang C, Li Z, Yang K. DenseASPP for semantic segmentation in street scenes. In: 2018 IEEE/CVF conference on computer vision and pattern recognition. 2018. p. 3684–92.
15.
Zurück zum Zitat Dong G, Yan Y, Shen C, Wang H. Real-time high-performance semantic image segmentation of urban street scenes. IEEE. 2021. Dong G, Yan Y, Shen C, Wang H. Real-time high-performance semantic image segmentation of urban street scenes. IEEE. 2021.
16.
Zurück zum Zitat Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2018. p. 7132–41. Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 2018. p. 7132–41.
17.
Zurück zum Zitat Woo S, Park J, Lee J-Y, Kweon IS. Cbam: convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV). 2018. p. 3–19. Woo S, Park J, Lee J-Y, Kweon IS. Cbam: convolutional block attention module. In: Proceedings of the European conference on computer vision (ECCV). 2018. p. 3–19.
20.
Zurück zum Zitat Ledley RS, Lusted LB. Reasoning foundations of medical diagnosis: symbolic logic, probability, and value theory aid our understanding of how physicians reason. Science. 1959;130:9–21.CrossRefPubMed Ledley RS, Lusted LB. Reasoning foundations of medical diagnosis: symbolic logic, probability, and value theory aid our understanding of how physicians reason. Science. 1959;130:9–21.CrossRefPubMed
21.
Zurück zum Zitat Pereira S, Pinto A, Alves V, Silva CA. Brain tumor segmentation using convolutional neural networks in MRI images. Social Science Electronic Publishing; 2016.CrossRef Pereira S, Pinto A, Alves V, Silva CA. Brain tumor segmentation using convolutional neural networks in MRI images. Social Science Electronic Publishing; 2016.CrossRef
22.
Zurück zum Zitat Roth HR, Lu L, Liu J, Yao J, Seff A, Cherry K, et al. Improving computer-aided detection using convolutional neural networks and random view aggregation. IEEE Trans Med Imaging. 2016;35:1170–81.CrossRefPubMed Roth HR, Lu L, Liu J, Yao J, Seff A, Cherry K, et al. Improving computer-aided detection using convolutional neural networks and random view aggregation. IEEE Trans Med Imaging. 2016;35:1170–81.CrossRefPubMed
24.
Zurück zum Zitat Norman B, Pedoia V, Majumdar S. Use of 2D U-net convolutional neural networks for automated cartilage and meniscus segmentation of knee MR imaging data to determine relaxometry and morphometry. Radiology. 2018;288:177–85.CrossRefPubMed Norman B, Pedoia V, Majumdar S. Use of 2D U-net convolutional neural networks for automated cartilage and meniscus segmentation of knee MR imaging data to determine relaxometry and morphometry. Radiology. 2018;288:177–85.CrossRefPubMed
25.
Zurück zum Zitat Liu F, Zhou Z, Alexey S, Donna B, Will L, Andrew K, et al. Deep learning approach for evaluating knee MR images: achieving high diagnostic performance for cartilage lesion detection. Radiology. 2018;289:172986.CrossRef Liu F, Zhou Z, Alexey S, Donna B, Will L, Andrew K, et al. Deep learning approach for evaluating knee MR images: achieving high diagnostic performance for cartilage lesion detection. Radiology. 2018;289:172986.CrossRef
26.
Zurück zum Zitat Bien N, Rajpurkar P, Ball RL, Irvin J, Lungren MP. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of MRNet. PLoS Med. 2018;15:e1002699.CrossRefPubMedPubMedCentral Bien N, Rajpurkar P, Ball RL, Irvin J, Lungren MP. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of MRNet. PLoS Med. 2018;15:e1002699.CrossRefPubMedPubMedCentral
27.
Zurück zum Zitat Jamaludin A, Lootus M, Kadir T, Zisserman A, Urban J, Battié MC, et al. Issls prize in bioengineering science 2017: automation of reading of radiological features from magnetic resonance images (MRIs) of the lumbar spine without human intervention is comparable with an expert radiologist. Eur Spine J. 2017;26:1374–83.CrossRefPubMed Jamaludin A, Lootus M, Kadir T, Zisserman A, Urban J, Battié MC, et al. Issls prize in bioengineering science 2017: automation of reading of radiological features from magnetic resonance images (MRIs) of the lumbar spine without human intervention is comparable with an expert radiologist. Eur Spine J. 2017;26:1374–83.CrossRefPubMed
28.
Zurück zum Zitat Larson DB, Chen MC, Lungren MP, Halabi SS, Stence NV, Langlotz CP. Performance of a deep-learning neural network model in assessing skeletal maturity on pediatric hand radiographs. Radiology. 2017;287:313.CrossRefPubMed Larson DB, Chen MC, Lungren MP, Halabi SS, Stence NV, Langlotz CP. Performance of a deep-learning neural network model in assessing skeletal maturity on pediatric hand radiographs. Radiology. 2017;287:313.CrossRefPubMed
29.
Zurück zum Zitat Shim E, Kim JY, Yoon JP, Ki SY, Lho T, Kim Y, et al. Author Correction: Automated rotator cuff tear classification using 3D convolutional neural network. Sci Rep. 2021;11:15996.CrossRefPubMedPubMedCentral Shim E, Kim JY, Yoon JP, Ki SY, Lho T, Kim Y, et al. Author Correction: Automated rotator cuff tear classification using 3D convolutional neural network. Sci Rep. 2021;11:15996.CrossRefPubMedPubMedCentral
30.
Zurück zum Zitat Medina G, Buckless CG, Thomasson E, Oh LS, Torriani M. Deep learning method for segmentation of rotator cuff muscles on MR images. Skeletal Radiol. 2020;50:1–10. Medina G, Buckless CG, Thomasson E, Oh LS, Torriani M. Deep learning method for segmentation of rotator cuff muscles on MR images. Skeletal Radiol. 2020;50:1–10.
31.
Zurück zum Zitat Ro K, Kim JY, Park H, Cho BH, Yoo JC. Deep-learning framework and computer assisted fatty infiltration analysis for the supraspinatus muscle in MRI. Scientific Reports; 2021.CrossRef Ro K, Kim JY, Park H, Cho BH, Yoo JC. Deep-learning framework and computer assisted fatty infiltration analysis for the supraspinatus muscle in MRI. Scientific Reports; 2021.CrossRef
32.
Zurück zum Zitat Yao J, Chepelev L, Nisha Y, Sathiadoss P, Rybicki FJ, Sheikh AM. Evaluation of a deep learning method for the automated detection of supraspinatus tears on MRI. Skeletal Radiol. 2022;51:1765–75.CrossRefPubMed Yao J, Chepelev L, Nisha Y, Sathiadoss P, Rybicki FJ, Sheikh AM. Evaluation of a deep learning method for the automated detection of supraspinatus tears on MRI. Skeletal Radiol. 2022;51:1765–75.CrossRefPubMed
33.
Zurück zum Zitat Hess H, Ruckli AC, Bürki F, Gerber N, Menzemer J, Burger J, et al. Deep-learning-based segmentation of the shoulder from MRI with inference accuracy prediction. Diagnostics. 2023;13:1668.CrossRefPubMedPubMedCentral Hess H, Ruckli AC, Bürki F, Gerber N, Menzemer J, Burger J, et al. Deep-learning-based segmentation of the shoulder from MRI with inference accuracy prediction. Diagnostics. 2023;13:1668.CrossRefPubMedPubMedCentral
34.
Zurück zum Zitat Lin D, Schwier M, Geiger B, Raithel E, Von Busch H, Fritz J, et al. Deep learning diagnosis and classification of rotator cuff tears on shoulder MRI. Investigative radiology. 2023;Publish Ahead of Print. Lin D, Schwier M, Geiger B, Raithel E, Von Busch H, Fritz J, et al. Deep learning diagnosis and classification of rotator cuff tears on shoulder MRI. Investigative radiology. 2023;Publish Ahead of Print.
Metadaten
Titel
Supraspinatus extraction from MRI based on attention-dense spatial pyramid UNet network
verfasst von
Peng Wang
Yang Liu
Zhong Zhou
Publikationsdatum
01.12.2024
Verlag
BioMed Central
Erschienen in
Journal of Orthopaedic Surgery and Research / Ausgabe 1/2024
Elektronische ISSN: 1749-799X
DOI
https://doi.org/10.1186/s13018-023-04509-7

Weitere Artikel der Ausgabe 1/2024

Journal of Orthopaedic Surgery and Research 1/2024 Zur Ausgabe

Arthropedia

Grundlagenwissen der Arthroskopie und Gelenkchirurgie. Erweitert durch Fallbeispiele, Videos und Abbildungen. 
» Jetzt entdecken

Knie-TEP: Kein Vorteil durch antibiotikahaltigen Knochenzement

29.05.2024 Periprothetische Infektionen Nachrichten

Zur Zementierung einer Knie-TEP wird in Deutschland zu über 98% Knochenzement verwendet, der mit einem Antibiotikum beladen ist. Ob er wirklich besser ist als Zement ohne Antibiotikum, kann laut Registerdaten bezweifelt werden.

Häusliche Gewalt in der orthopädischen Notaufnahme oft nicht erkannt

28.05.2024 Häusliche Gewalt Nachrichten

In der Notaufnahme wird die Chance, Opfer von häuslicher Gewalt zu identifizieren, von Orthopäden und Orthopädinnen offenbar zu wenig genutzt. Darauf deuten die Ergebnisse einer Fragebogenstudie an der Sahlgrenska-Universität in Schweden hin.

Fehlerkultur in der Medizin – Offenheit zählt!

28.05.2024 Fehlerkultur Podcast

Darüber reden und aus Fehlern lernen, sollte das Motto in der Medizin lauten. Und zwar nicht nur im Sinne der Patientensicherheit. Eine negative Fehlerkultur kann auch die Behandelnden ernsthaft krank machen, warnt Prof. Dr. Reinhard Strametz. Ein Plädoyer und ein Leitfaden für den offenen Umgang mit kritischen Ereignissen in Medizin und Pflege.

Mehr Frauen im OP – weniger postoperative Komplikationen

21.05.2024 Allgemeine Chirurgie Nachrichten

Ein Frauenanteil von mindestens einem Drittel im ärztlichen Op.-Team war in einer großen retrospektiven Studie aus Kanada mit einer signifikanten Reduktion der postoperativen Morbidität assoziiert.

Update Orthopädie und Unfallchirurgie

Bestellen Sie unseren Fach-Newsletter und bleiben Sie gut informiert.