Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 52
Filtrar
1.
Comput Med Imaging Graph ; 116: 102410, 2024 Sep.
Artículo en Inglés | MEDLINE | ID: mdl-38905961

RESUMEN

Trabecular bone analysis plays a crucial role in understanding bone health and disease, with applications like osteoporosis diagnosis. This paper presents a comprehensive study on 3D trabecular computed tomography (CT) image restoration, addressing significant challenges in this domain. The research introduces a backbone model, Cascade-SwinUNETR, for single-view 3D CT image restoration. This model leverages deep layer aggregation with supervision and capabilities of Swin-Transformer to excel in feature extraction. Additionally, this study also brings DVSR3D, a dual-view restoration model, achieving good performance through deep feature fusion with attention mechanisms and Autoencoders. Furthermore, an Unsupervised Domain Adaptation (UDA) method is introduced, allowing models to adapt to input data distributions without additional labels, holding significant potential for real-world medical applications, and eliminating the need for invasive data collection procedures. The study also includes the curation of a new dual-view dataset for CT image restoration, addressing the scarcity of real human bone data in Micro-CT. Finally, the dual-view approach is validated through downstream medical bone microstructure measurements. Our contributions open several paths for trabecular bone analysis, promising improved clinical outcomes in bone health assessment and diagnosis.


Asunto(s)
Hueso Esponjoso , Aprendizaje Profundo , Imagenología Tridimensional , Tomografía Computarizada por Rayos X , Humanos , Tomografía Computarizada por Rayos X/métodos , Imagenología Tridimensional/métodos , Hueso Esponjoso/diagnóstico por imagen
2.
Sensors (Basel) ; 24(9)2024 Apr 24.
Artículo en Inglés | MEDLINE | ID: mdl-38732827

RESUMEN

Arterial blood pressure (ABP) serves as a pivotal clinical metric in cardiovascular health assessments, with the precise forecasting of continuous blood pressure assuming a critical role in both preventing and treating cardiovascular diseases. This study proposes a novel continuous non-invasive blood pressure prediction model, DSRUnet, based on deep sparse residual U-net combined with improved SE skip connections, which aim to enhance the accuracy of using photoplethysmography (PPG) signals for continuous blood pressure prediction. The model first introduces a sparse residual connection approach for path contraction and expansion, facilitating richer information fusion and feature expansion to better capture subtle variations in the original PPG signals, thereby enhancing the network's representational capacity and predictive performance and mitigating potential degradation in the network performance. Furthermore, an enhanced SE-GRU module was embedded in the skip connections to model and weight global information using an attention mechanism, capturing the temporal features of the PPG pulse signals through GRU layers to improve the quality of the transferred feature information and reduce redundant feature learning. Finally, a deep supervision mechanism was incorporated into the decoder module to guide the lower-level network to learn effective feature representations, alleviating the problem of gradient vanishing and facilitating effective training of the network. The proposed DSRUnet model was trained and tested on the publicly available UCI-BP dataset, with the average absolute errors for predicting systolic blood pressure (SBP), diastolic blood pressure (DBP), and mean blood pressure (MBP) being 3.36 ± 6.61 mmHg, 2.35 ± 4.54 mmHg, and 2.21 ± 4.36 mmHg, respectively, meeting the standards set by the Association for the Advancement of Medical Instrumentation (AAMI), and achieving Grade A according to the British Hypertension Society (BHS) Standard for SBP and DBP predictions. Through ablation experiments and comparisons with other state-of-the-art methods, the effectiveness of DSRUnet in blood pressure prediction tasks, particularly for SBP, which generally yields poor prediction results, was significantly higher. The experimental results demonstrate that the DSRUnet model can accurately utilize PPG signals for real-time continuous blood pressure prediction and obtain high-quality and high-precision blood pressure prediction waveforms. Due to its non-invasiveness, continuity, and clinical relevance, the model may have significant implications for clinical applications in hospitals and research on wearable devices in daily life.


Asunto(s)
Presión Sanguínea , Fotopletismografía , Humanos , Fotopletismografía/métodos , Presión Sanguínea/fisiología , Algoritmos , Procesamiento de Señales Asistido por Computador , Redes Neurales de la Computación , Determinación de la Presión Sanguínea/métodos
3.
Comput Biol Med ; 172: 108250, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38493603

RESUMEN

Lung tumor segmentation in medical imaging is a critical step in the diagnosis and treatment planning for lung cancer. Accurate segmentation, however, is challenging due to the variability in tumor size, shape, and contrast against surrounding tissues. In this work, we present MSMV-Net, a novel deep learning architecture that integrates multi-scale multi-view (MSMV) learning modules and multi-scale uncertainty-based deep supervision (MUDS) for enhanced segmentation of lung tumors in computed tomography images. MSMV-Net capitalizes on the strengths of multi-view analysis and multi-scale feature extraction to address the limitations posed by small 3D lung tumors. The results indicate that MSMV-Net achieves state-of-the-art performance in lung tumor segmentation, recording a global Dice score of 55.60% on the LUNA dataset and 59.94% on the MSD dataset. Ablation studies conducted on the MSD dataset further validate that our method enhances segmentation accuracy.


Asunto(s)
Neoplasias Pulmonares , Humanos , Neoplasias Pulmonares/diagnóstico por imagen , Tórax , Tomografía Computarizada por Rayos X , Incertidumbre , Procesamiento de Imagen Asistido por Computador
4.
Biomedicines ; 12(3)2024 Mar 05.
Artículo en Inglés | MEDLINE | ID: mdl-38540193

RESUMEN

Differentiating between a salvageable Ischemic Penumbra (IP) and an irreversibly damaged Infarct Core (IC) is important for therapy decision making for acute ischemic stroke (AIS) patients. Existing methods rely on Computed Tomography Perfusion (CTP) or Diffusion-Weighted Imaging-Fluid Attenuated Inversion Recovery (DWI-FLAIR). We designed a novel Convolutional Neural Network named I2PC-Net, which relies solely on Non-Contrast Computed Tomography (NCCT) for the automatic and simultaneous segmentation of the IP and IC. In the encoder, Multi-Scale Convolution (MSC) blocks were proposed to capture effective features of ischemic lesions, and in the deep levels of the encoder, Symmetry Enhancement (SE) blocks were also designed to enhance anatomical symmetries. In the attention-based decoder, hierarchical deep supervision was introduced to address the challenge of differentiating between the IP and IC. We collected 197 NCCT scans from AIS patients to evaluate the proposed method. On the test set, I2PC-Net achieved Dice Similarity Scores of 42.76 ± 21.84%, 33.54 ± 24.13% and 65.67 ± 12.30% and lesion volume correlation coefficients of 0.95 (p < 0.001), 0.61 (p < 0.001) and 0.93 (p < 0.001) for the IP, IC and IP + IC, respectively. The results indicated that NCCT could potentially be used as a surrogate technique of CTP for the quantitative evaluation of the IP and IC.

5.
Artif Intell Med ; 150: 102800, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38553146

RESUMEN

Image segmentation is one of the vital steps in medical image analysis. A large number of methods based on convolutional neural networks have emerged, which can extract abstract features from multiple-modality medical images, learn valuable information that is difficult to recognize by humans, and obtain more reliable results than traditional image segmentation approaches. U-Net, due to its simple structure and excellent performance, is widely used in medical image segmentation. In this paper, to further improve the performance of U-Net, we propose a channel and space compound attention (CSCA) convolutional neural network, CSCA U-Net in abbreviation, which increases the network depth and employs a double squeeze-and-excitation (DSE) block in the bottleneck layer to enhance feature extraction and obtain more high-level semantic features. Moreover, the characteristics of the proposed method are three-fold: (1) channel and space compound attention (CSCA) block, (2) cross-layer feature fusion (CLFF), and (3) deep supervision (DS). Extensive experiments on several available medical image datasets, including Kvasir-SEG, CVC-ClinicDB, CVC-ColonDB, ETIS, CVC-T, 2018 Data Science Bowl (2018 DSB), ISIC 2018, and JSUAH-Cerebellum, show that CSCA U-Net achieves competitive results and significantly improves generalization performance. The codes and trained models are available at https://github.com/xiaolanshu/CSCA-U-Net.


Asunto(s)
Ciencia de los Datos , Aprendizaje , Humanos , Redes Neurales de la Computación , Semántica , Procesamiento de Imagen Asistido por Computador
6.
J Xray Sci Technol ; 32(3): 707-723, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-38552134

RESUMEN

Highlights: • Introduce a data augmentation strategy to expand the required different morphological data during the training and learning phase, and improve the algorithm's feature learning ability for complex and diverse tumor morphology CT images.• Design attention mechanisms for encoding and decoding paths to extract fine pixel level features, improve feature extraction capabilities, and achieve efficient spatial channel feature fusion.• The deep supervision layer is used to correct and decode the final image data to provide high accuracy of results.• The effectiveness of this method has been affirmed through validation on the LITS, 3DIRCADb, and SLIVER datasets. BACKGROUND: Accurately extracting liver and liver tumors from medical images is an important step in lesion localization and diagnosis, surgical planning, and postoperative monitoring. However, the limited number of radiation therapists and a great number of images make this work time-consuming. OBJECTIVE: This study designs a spatial attention deep supervised network (SADSNet) for simultaneous automatic segmentation of liver and tumors. METHOD: Firstly, self-designed spatial attention modules are introduced at each layer of the encoder and decoder to extract image features at different scales and resolutions, helping the model better capture liver tumors and fine structures. The designed spatial attention module is implemented through two gate signals related to liver and tumors, as well as changing the size of convolutional kernels; Secondly, deep supervision is added behind the three layers of the decoder to assist the backbone network in feature learning and improve gradient propagation, enhancing robustness. RESULTS: The method was testing on LITS, 3DIRCADb, and SLIVER datasets. For the liver, it obtained dice similarity coefficients of 97.03%, 96.11%, and 97.40%, surface dice of 81.98%, 82.53%, and 86.29%, 95% hausdorff distances of 8.96 mm, 8.26 mm, and 3.79 mm, and average surface distances of 1.54 mm, 1.19 mm, and 0.81 mm. Additionally, it also achieved precise tumor segmentation, which with dice scores of 87.81% and 87.50%, surface dice of 89.63% and 84.26%, 95% hausdorff distance of 12.96 mm and 16.55 mm, and average surface distances of 1.11 mm and 3.04 mm on LITS and 3DIRCADb, respectively. CONCLUSION: The experimental results show that the proposed method is effective and superior to some other methods. Therefore, this method can provide technical support for liver and liver tumor segmentation in clinical practice.


Asunto(s)
Algoritmos , Neoplasias Hepáticas , Hígado , Tomografía Computarizada por Rayos X , Neoplasias Hepáticas/diagnóstico por imagen , Humanos , Hígado/diagnóstico por imagen , Tomografía Computarizada por Rayos X/métodos , Imagenología Tridimensional/métodos , Redes Neurales de la Computación , Aprendizaje Profundo
7.
Phys Med Biol ; 69(5)2024 Feb 23.
Artículo en Inglés | MEDLINE | ID: mdl-38306960

RESUMEN

Automatically delineating colorectal cancers with fuzzy boundaries from 3D images is a challenging task, but the problem of fuzzy boundary delineation in existing deep learning-based methods have not been investigated in depth. Here, an encoder-decoder-based U-shaped network (U-Net) based on top-down deep supervision (TdDS) was designed to accurately and automatically delineate the fuzzy boundaries of colorectal cancer. TdDS refines the semantic targets of the upper and lower stages by mapping ground truths that are more consistent with the stage properties than upsampling deep supervision. This stage-specific approach can guide the model to learn a coarse-to-fine delineation process and improve the delineation accuracy of fuzzy boundaries by gradually shrinking the boundaries. Experimental results showed that TdDS is more customizable and plays a role similar to the attentional mechanism, and it can further improve the capability of the model to delineate colorectal cancer contours. A total of 103, 12, and 29 3D pelvic magnetic resonance imaging volumes were used for training, validation, and testing, respectively. The comparative results indicate that the proposed method exhibits the best comprehensive performance, with a dice similarity coefficient (DSC) of 0.805 ± 0.053 and a hausdorff distance (HD) of 9.28 ± 5.14 voxels. In the delineation performance analysis section also showed that 44.49% of the delineation results are satisfactory and do not require revisions. This study can provide new technical support for the delineation of 3D colorectal cancer. Our method is open source, and the code is available athttps://github.com/odindis/TdDS/tree/main.


Asunto(s)
Neoplasias Colorrectales , Pelvis , Humanos , Semántica , Neoplasias Colorrectales/diagnóstico por imagen
8.
Med Biol Eng Comput ; 62(2): 465-478, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-37914958

RESUMEN

This work presents a deep network architecture to improve nuclei detection performance and achieve the high localization accuracy of nuclei in breast cancer histopathology images. The proposed model consists of two parts, generating nuclear candidate module and refining nuclear localization module. We first design a novel patch learning method to obtain high-quality nuclear candidates, where in addition to categories, location representations are also added to the patch information to implement the multi-task learning process of nuclear classification and localization; meanwhile, the deep supervision mechanism is introduced to obtain the coherent contributions from each scale layer. In order to refine nuclear localization, we propose an iterative correction strategy to make the prediction progressively closer to the ground truth, which significantly improves the accuracy of nuclear localization and facilitates neighbor size selection in the nonmaximum suppression step. Experimental results demonstrate the superior performance of our method for nuclei detection on the H&E stained histopathological image dataset as compared to previous state-of-the-art methods, especially in multiple cluttered nuclei detection, can achieve better results than existing techniques.


Asunto(s)
Algoritmos , Neoplasias de la Mama , Humanos , Femenino , Mama/diagnóstico por imagen , Mama/patología , Neoplasias de la Mama/diagnóstico por imagen , Neoplasias de la Mama/patología , Núcleo Celular/patología , Procesamiento de Imagen Asistido por Computador/métodos
9.
Comput Biol Med ; 169: 107858, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38113680

RESUMEN

The U-shaped and Transformer architectures have achieved exceptional performance in medical image segmentation and natural language processing, respectively. Their combination has also led to remarkable results but still suffers from enormous loss of image features during downsampling and the difficulty of recovering spatial information during upsampling. In this paper, we propose a novel encoder-decoder architecture for medical image segmentation, which has a flexibly adjustable hybrid encoder and two expanding paths decoder. The hybrid encoder incorporates the feature double reuse (FDR) block and the encoder of Vision Transformer (ViT), which can extract local and global pixel localization information, and alleviate image feature loss effectively. Meanwhile, we retain the original class-token sequence in the Vision Transformer and develop an additional corresponding expanding path. The class-token sequence and abstract image features are leveraged by two independent expanding paths with the deep-supervision strategy, which can better recover the image spatial information and accelerate model convergence. To further mitigate the feature loss and improve spatial information recovery, we introduce successive residual connections throughout the entire network. We evaluated our model on the COVID-19 lung segmentation and the infection area segmentation tasks. The mIoU index increased by 1.5 points and 3.9 points compared to other models which demonstrates a performance improvement.


Asunto(s)
COVID-19 , Humanos , Procesamiento de Lenguaje Natural , Procesamiento de Imagen Asistido por Computador
10.
Comput Biol Med ; 169: 107879, 2024 Feb.
Artículo en Inglés | MEDLINE | ID: mdl-38142549

RESUMEN

The liver is one of the organs with the highest incidence rate in the human body, and late-stage liver cancer is basically incurable. Therefore, early diagnosis and lesion location of liver cancer are of important clinical value. This study proposes an enhanced network architecture ELTS-Net based on the 3D U-Net model, to address the limitations of conventional image segmentation methods and the underutilization of image spatial features by the 2D U-Net network structure. ELTS-Net expands upon the original network by incorporating dilated convolutions to increase the receptive field of the convolutional kernel. Additionally, an attention residual module, comprising an attention mechanism and residual connections, replaces the original convolutional module, serving as the primary components of the encoder and decoder. This design enables the network to capture contextual information globally in both channel and spatial dimensions. Furthermore, deep supervision modules are integrated between different levels of the decoder network, providing additional feedback from deeper intermediate layers. This constrains the network weights to the target regions and optimizing segmentation results. Evaluation on the LiTS2017 dataset shows improvements in evaluation metrics for liver and tumor segmentation tasks compared to the baseline 3D U-Net model, achieving 95.2% liver segmentation accuracy and 71.9% tumor segmentation accuracy, with accuracy improvements of 0.9% and 3.1% respectively. The experimental results validate the superior segmentation performance of ELTS-Net compared to other comparison models, offering valuable guidance for clinical diagnosis and treatment.


Asunto(s)
Neoplasias Hepáticas , Humanos , Algoritmos , Benchmarking , Procesamiento de Imagen Asistido por Computador
11.
Front Neurosci ; 17: 1249331, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-38075284

RESUMEN

Introduction: The accurate segmentation of retinal vessels is of utmost importance in the diagnosis of retinal diseases. However, the complex vessel structure often leads to poor segmentation performance, particularly in the case of microvessels. Methods: To address this issue, we propose a vessel segmentation method composed of preprocessing and a multi-scale feature attention network (MFA-UNet). The preprocessing stage involves the application of gamma correction and contrast-limited adaptive histogram equalization to enhance image intensity and vessel contrast. The MFA-UNet incorporates the Multi-scale Fusion Self-Attention Module(MSAM) that adjusts multi-scale features and establishes global dependencies, enabling the network to better preserve microvascular structures. Furthermore, the multi-branch decoding module based on deep supervision (MBDM) replaces the original output layer to achieve targeted segmentation of macrovessels and microvessels. Additionally, a parallel attention mechanism is embedded into the decoder to better exploit multi-scale features in skip paths. Results: The proposed MFA-UNet yields competitive performance, with dice scores of 82.79/83.51/84.17/78.60/81.75/84.04 and accuracies of 95.71/96.4/96.71/96.81/96.32/97.10 on the DRIVE, STARE, CHASEDB1, HRF, IOSTAR and FIVES datasets, respectively. Discussion: It is expected to provide reliable segmentation results in clinical diagnosis.

12.
Bioengineering (Basel) ; 10(12)2023 Dec 15.
Artículo en Inglés | MEDLINE | ID: mdl-38136019

RESUMEN

Non-contact remote photoplethysmography can be used in a variety of medical and healthcare fields by measuring vital signs continuously and unobtrusively. Recently, end-to-end deep learning methods have been proposed to replace the existing handcrafted features. However, since the existing deep learning methods are known as black box models, the problem of interpretability has been raised, and the same problem exists in the remote photoplethysmography (rPPG) network. In this study, we propose a method to visualize temporal and spectral representations for hidden layers, deeply supervise the spectral representation of intermediate layers through the depth of networks and optimize it for a lightweight model. The optimized network improves performance and enables fast training and inference times. The proposed spectral deep supervision helps to achieve not only high performance but also fast convergence speed through the regularization of the intermediate layers. The effect of the proposed methods was confirmed through a thorough ablation study on public datasets. As a result, similar or outperforming results were obtained in comparison to state-of-the-art models. In particular, our model achieved an RMSE of 1 bpm on the PURE dataset, demonstrating its high accuracy. Moreover, it excelled on the V4V dataset with an impressive RMSE of 6.65 bpm, outperforming other methods. We observe that our model began converging from the very first epoch, a significant improvement over other models in terms of learning efficiency. Our approach is expected to be generally applicable to models that learn spectral domain information as well as to the applications of regression that require the representations of periodicity.

13.
Curr Med Imaging ; 2023 Nov 06.
Artículo en Inglés | MEDLINE | ID: mdl-37936446

RESUMEN

BACKGROUND: Lung cancer is a pervasive and persistent issue worldwide, with the highest morbidity and mortality among all cancers for many years. In the medical field, computer tomography (CT) images of the lungs are currently recognized as the best way to help doctors detect lung nodules and thus diagnose lung cancer. U-Net is a deep learning network with an encoder-decoder structure, which is extensively employed for medical image segmentation and has derived many improved versions. However, these advancements do not utilize various feature information from all scales, and there is still room for future enhancement. METHODS: In this study, we proposed a new model called Blend U-Net, which incorporates nested structures, redesigned long and short skip connections, and deep supervisions. The nested structures and the long and short skip connections combined characteristic information of different levels from feature maps in all scales, while the deep supervision learning hierarchical representations from all-scale concatenated feature maps. Additionally, we employed a mixed loss function to obtain more accurate results. RESULTS: We evaluated the performance of the Blend U-Net against other architectures on the publicly available Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) dataset. Moreover, the accuracy of the segmentation was verified by using the dice coefficient. Blend U-Net with a boost of 0.83 points produced the best outcome in a number of baselines. CONCLUSION: Based on the results, our method achieves superior performance in terms of dice coefficient compared to other methods and demonstrates greater proficiency in segmenting lung nodules of varying sizes.

14.
Comput Biol Med ; 167: 107578, 2023 12.
Artículo en Inglés | MEDLINE | ID: mdl-37918260

RESUMEN

Pixel differences between classes with low contrast in medical image semantic segmentation tasks often lead to confusion in category classification, posing a typical challenge for recognition of small targets. To address this challenge, we propose a Contrastive Adaptive Augmented Semantic Segmentation Network with a differentiable pooling function. Firstly, an Adaptive Contrast Augmentation module is constructed to automatically extract local high-frequency information, thereby enhancing image details and accentuating the differences between classes. Subsequently, the Frequency-Efficient Channel Attention mechanism is designed to select useful features in the encoding phase, where multifrequency information is employed to extract channel features. One-dimensional convolutional cross-channel interactions are adopted to reduce model complexity. Finally, a differentiable approximation of max pooling is introduced in order to replace standard max pooling, strengthening the connectivity between neurons and reducing information loss caused by downsampling. We evaluated the effectiveness of our proposed method through several ablation experiments and comparison experiments under homogeneous conditions. The experimental results demonstrate that our method competes favorably with other state-of-the-art networks on five medical image datasets, including four public medical image datasets and one clinical image dataset. It can be effectively applied to medical image segmentation.


Asunto(s)
Web Semántica , Semántica , Procesamiento de Imagen Asistido por Computador
15.
Artif Intell Med ; 145: 102679, 2023 11.
Artículo en Inglés | MEDLINE | ID: mdl-37925209

RESUMEN

Facial wrinkles are important indicators of human aging. Recently, a method using deep learning and a semi-automatic labeling was proposed to segment facial wrinkles, which showed much better performance than conventional image-processing-based methods. However, the difficulty of wrinkle segmentation remains challenging due to the thinness of wrinkles and their small proportion in the entire image. Therefore, performance improvement in wrinkle segmentation is still necessary. To address this issue, we propose a novel loss function that takes into account the thickness of wrinkles based on the semi-automatic labeling approach. First, considering the different spatial dimensions of the decoder in the U-Net architecture, we generated weighted wrinkle maps from ground truth. These weighted wrinkle maps were used to calculate the training losses more accurately than the existing deep supervision approach. This new loss computation approach is defined as weighted deep supervision in our study. The proposed method was evaluated using an image dataset obtained from a professional skin analysis device and labeled using semi-automatic labeling. In our experiment, the proposed weighted deep supervision showed higher Jaccard Similarity Index (JSI) performance for wrinkle segmentation compared to conventional deep supervision and traditional image processing methods. Additionally, we conducted experiments on the labeling using a semi-automatic labeling approach, which had not been explored in previous research, and compared it with human labeling. The semi-automatic labeling technology showed more consistent wrinkle labels than human-made labels. Furthermore, to assess the scalability of the proposed method to other domains, we applied it to retinal vessel segmentation. The results demonstrated superior performance of the proposed method compared to existing retinal vessel segmentation approaches. In conclusion, the proposed method offers high performance and can be easily applied to various biomedical domains and U-Net-based architectures. Therefore, the proposed approach will be beneficial for various biomedical imaging approaches. To facilitate this, we have made the source code of the proposed method publicly available at: https://github.com/resemin/WeightedDeepSupervision.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Vasos Retinianos , Humanos , Procesamiento de Imagen Asistido por Computador/métodos
16.
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi ; 40(5): 928-937, 2023 Oct 25.
Artículo en Chino | MEDLINE | ID: mdl-37879922

RESUMEN

Accurate segmentation of pediatric echocardiograms is a challenging task, because significant heart-size changes with age and faster heart rate lead to more blurred boundaries on cardiac ultrasound images compared with adults. To address these problems, a dual decoder network model combining channel attention and scale attention is proposed in this paper. Firstly, an attention-guided decoder with deep supervision strategy is used to obtain attention maps for the ventricular regions. Then, the generated ventricular attention is fed back to multiple layers of the network through skip connections to adjust the feature weights generated by the encoder and highlight the left and right ventricular areas. Finally, a scale attention module and a channel attention module are utilized to enhance the edge features of the left and right ventricles. The experimental results demonstrate that the proposed method in this paper achieves an average Dice coefficient of 90.63% in acquired bilateral ventricular segmentation dataset, which is better than some conventional and state-of-the-art methods in the field of medical image segmentation. More importantly, the method has a more accurate effect in segmenting the edge of the ventricle. The results of this paper can provide a new solution for pediatric echocardiographic bilateral ventricular segmentation and subsequent auxiliary diagnosis of congenital heart disease.


Asunto(s)
Ecocardiografía , Ventrículos Cardíacos , Adulto , Humanos , Niño , Ventrículos Cardíacos/diagnóstico por imagen , Procesamiento de Imagen Asistido por Computador
17.
Comput Med Imaging Graph ; 109: 102301, 2023 10.
Artículo en Inglés | MEDLINE | ID: mdl-37738774

RESUMEN

Accurate segmentation of the renal cancer structure, including the kidney, renal tumors, veins, and arteries, has great clinical significance, which can assist clinicians in diagnosing and treating renal cancer. For accurate segmentation of the renal cancer structure in contrast-enhanced computed tomography (CT) images, we proposed a novel encoder-decoder structure segmentation network named MDM-U-Net comprising a multi-scale anisotropic convolution block, dual activation attention block, and multi-scale deep supervision mechanism. The multi-scale anisotropic convolution block was used to improve the feature extraction ability of the network, the dual activation attention block as a channel-wise mechanism was used to guide the network to exploit important information, and the multi-scale deep supervision mechanism was used to supervise the layers of the decoder part for improving segmentation performance. In this study, we developed a feasible and generalizable MDM-U-Net model for renal cancer structure segmentation, trained the model from the public KiPA22 dataset, and tested it on the KiPA22 dataset and an in-house dataset. For the KiPA22 dataset, our method ranked first in renal cancer structure segmentation, achieving state-of-the-art (SOTA) performance in terms of 6 of 12 evaluation metrics (3 metrics per structure). For the in-house dataset, our method achieves SOTA performance in terms of 9 of 12 evaluation metrics (3 metrics per structure), demonstrating its superiority and generalization ability over the compared networks in renal structure segmentation from contrast-enhanced CT scans.


Asunto(s)
Neoplasias Renales , Humanos , Neoplasias Renales/diagnóstico por imagen , Riñón , Arterias , Benchmarking , Relevancia Clínica , Procesamiento de Imagen Asistido por Computador
18.
Comput Biol Med ; 165: 107374, 2023 10.
Artículo en Inglés | MEDLINE | ID: mdl-37611428

RESUMEN

BACKGROUND AND OBJECTIVE: Image-guided clinical diagnosis can be achieved by automatically and accurately segmenting prostate and prostatic cancer in male pelvic magnetic resonance imaging (MRI) images. For accurate tumor removal, the location, number, and size of prostate cancer are crucial, especially in surgical patients. The morphological differences between the prostate and tumor regions are small, the size of the tumor is uncertain, the boundary between the tumor and surrounding tissue is blurred, and the classification that separates the normal region from the tumor is uneven. Therefore, segmenting prostate and tumor on MRI images is challenging. METHODS: This study offers a new prostate and prostatic cancer segmentation network based on double branch attention driven multi-scale learning for MRI. To begin, the dual branch structure provides two input images with different scales for feature coding, as well as a multi-scale attention module that collects details from different scales. The features of the double branch structure are then entered into the built feature fusion module to get more complete context information. Finally, to give a more precise learning representation, each stage is built using a deep supervision mechanism. RESULTS: The results of our proposed network's prostate and tumor segmentation on a variety of male pelvic MRI data sets show that it outperforms existing techniques. For prostate and prostatic cancer MRI segmentation, the dice similarity coefficient (DSC) values were 91.65% and 84.39%, respectively. CONCLUSIONS: Our method maintains high correlation and consistency between automatic segmentation results and expert manual segmentation results. Accurate automatic segmentation of prostate and prostate cancer has important clinical significance.


Asunto(s)
Aprendizaje Profundo , Neoplasias de la Próstata , Humanos , Masculino , Próstata/diagnóstico por imagen , Imagen por Resonancia Magnética/métodos , Neoplasias de la Próstata/diagnóstico por imagen , Pelvis , Procesamiento de Imagen Asistido por Computador/métodos
19.
Comput Biol Med ; 163: 107218, 2023 09.
Artículo en Inglés | MEDLINE | ID: mdl-37393784

RESUMEN

Accurate gland segmentation is critical in determining adenocarcinoma. Automatic gland segmentation methods currently suffer from challenges such as less accurate edge segmentation, easy mis-segmentation, and incomplete segmentation. To solve these problems, this paper proposes a novel gland segmentation network Dual-branch Attention-guided Refinement and Multi-scale Features Fusion U-Net (DARMF-UNet), which fuses multi-scale features using deep supervision. At the first three layers of feature concatenation, a Coordinate Parallel Attention (CPA) is proposed to guide the network to focus on the key regions. A Dense Atrous Convolution (DAC) block is used in the fourth layer of feature concatenation to perform multi-scale features extraction and obtain global information. A hybrid loss function is adopted to calculate the loss of each segmentation result of the network to achieve deep supervision and improve the accuracy of segmentation. Finally, the segmentation results at different scales in each part of the network are fused to obtain the final gland segmentation result. The experimental results on the gland datasets Warwick-QU and Crag show that the network improves in terms of the evaluation metrics of F1 Score, Object Dice, Object Hausdorff, and the segmentation effect is better than the state-of-the-art network models.


Asunto(s)
Adenocarcinoma , Aprendizaje Profundo , Adenocarcinoma/diagnóstico por imagen , Adenocarcinoma/patología , Humanos
20.
Med Biol Eng Comput ; 61(8): 1929-1946, 2023 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-37243853

RESUMEN

Accurate segmentation of lung nodules is the key to diagnosing the lesion type of lung nodule. The complex boundaries of lung nodules and the visual similarity to surrounding tissues make precise segmentation of lung nodules challenging. Traditional CNN based lung nodule segmentation models focus on extracting local features from neighboring pixels and ignore global contextual information, which is prone to incomplete segmentation of lung nodule boundaries. In the U-shaped encoder-decoder structure, variations of image resolution caused by up-sampling and down-sampling result in the loss of feature information, which reduces the reliability of output features. This paper proposes transformer pooling module and dual-attention feature reorganization module to effectively improve the above two defects. Transformer pooling module innovatively fuses the self-attention layer and pooling layer in the transformer, which compensates for the limitation of convolution operation, reduces the loss of feature information in the pooling process, and decreases the computational complexity of the Transformer significantly. Dual-attention feature reorganization module innovatively employs the dual-attention mechanism of channel and spatial to improve the sub-pixel convolution, minimizing the loss of feature information during up-sampling. In addition, two convolutional modules are proposed in this paper, which together with transformer pooling module form an encoder that can adequately extract local features and global dependencies. We use the fusion loss function and deep supervision strategy in the decoder to train the model. The proposed model has been extensively experimented and evaluated on the LIDC-IDRI dataset, the highest Dice Similarity Coefficient is 91.84 and the highest sensitivity is 92.66, indicating the model's comprehensive capability has surpassed state-of-the-art UTNet. The model proposed in this paper has superior segmentation performance for lung nodules and can provide a more in-depth assessment of lung nodules' shape, size, and other characteristics, which is of important clinical significance and application value to assist physicians in the early diagnosis of lung nodules.


Asunto(s)
Relevancia Clínica , Médicos , Humanos , Reproducibilidad de los Resultados , Suministros de Energía Eléctrica , Pulmón/diagnóstico por imagen , Procesamiento de Imagen Asistido por Computador
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA