Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 71
Filtrar
1.
Med Phys ; 2024 Sep 20.
Artículo en Inglés | MEDLINE | ID: mdl-39302604

RESUMEN

BACKGROUND: Deformable image registration is an essential technique of medical image analysis, which plays important roles in several clinical applications. Existing deep learning-based registration methods have already achieved promising performance for the registrations with small deformations, while it is still challenging to deal with the large deformation registration due to the limits of the image intensity-similarity-based objective function. PURPOSE: To achieve the image registration with large-scale deformations, we proposed a multilevel network architecture FCNet to gradually refine the registration results based on semantic feature consistency constraint and flow normalization (FN) strategy. METHODS: At each level of FCNet, the architecture is mainly composed to a FeaExtractor, a FN module, and a spatial transformation module. FeaExtractor consists of three parallel streams which are used to extract the individual features of fixed and moving images, as well as their joint features, respectively. Using these features, the initial deformation field is estimated, which passes through a FN module to refine the deformation field based on the difference map of deformation filed between two adjacent levels. This allows the FCNet to progressively improve the registration performance. Finally, a spatial transformation module is used to get the warped image based on the deformation field. Moreover, in addition to the image intensity-similarity-based objective function, a semantic-feature consistency constraint is also introduced, which can further promote the alignments by imposing the similarity between the fixed and warped image features. To validate the effectiveness of the proposed method, we compared our method with the state-of-the-art methods on three different datasets. In EMPIRE10 dataset, 20, 3, and 7 fixed and moving 3D computer tomography (CT) image pairs were used for training, validation, and testing respectively; in IXI dataset, atlas to individual image registration task was performed, with 3D MR images of 408, 58, and 115 individuals were used for training, validation, and testing respectively; in the in-house dataset, patient to atlas registration task was implemented, with the 3D MR images of 94, 3, and 15 individuals being training, validation, and testing sets, respectively. RESULTS: The qualitative and quantitative comparison results demonstrated that the proposed method is beneficial for handling large deformation image registration problems, with the DSC and ASSD improved by at least 1.0% and 25.9% on EMPIRE10 dataset. The ablation experiments also verified the effectiveness of the proposed feature combination strategy, feature consistency constraint, and FN module. CONCLUSIONS: Our proposed FCNet enables multiscale registration from coarse to fine, surpassing existing SOTA registration methods and effectively handling long-range spatial relationships.

2.
Med Image Anal ; 96: 103212, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-38830326

RESUMEN

Deformable image registration is an essential component of medical image analysis and plays an irreplaceable role in clinical practice. In recent years, deep learning-based registration methods have demonstrated significant improvements in convenience, robustness and execution time compared to traditional algorithms. However, registering images with large displacements, such as those of the liver organ, remains underexplored and challenging. In this study, we present a novel convolutional neural network (CNN)-based unsupervised learning registration method, Cascaded Multi-scale Spatial-Channel Attention-guided Network (CMAN), which addresses the challenge of large deformation fields using a double coarse-to-fine registration approach. The main contributions of CMAN include: (i) local coarse-to-fine registration in the base network, which generates the displacement field for each resolution and progressively propagates these local deformations as auxiliary information for the final deformation field; (ii) global coarse-to-fine registration, which stacks multiple base networks for sequential warping, thereby incorporating richer multi-layer contextual details into the final deformation field; (iii) integration of the spatial-channel attention module in the decoder stage, which better highlights important features and improves the quality of feature maps. The proposed network was trained using two public datasets and evaluated on another public dataset as well as a private dataset across several experimental scenarios. We compared CMAN with four state-of-the-art CNN-based registration methods and two well-known traditional algorithms. The results show that the proposed double coarse-to-fine registration strategy outperforms other methods in most registration evaluation metrics. In conclusion, CMAN can effectively handle the large-deformation registration problem and show potential for application in clinical practice. The source code is made publicly available at https://github.com/LocPham263/CMAN.git.


Asunto(s)
Imagenología Tridimensional , Hígado , Redes Neurales de la Computación , Tomografía Computarizada por Rayos X , Humanos , Tomografía Computarizada por Rayos X/métodos , Hígado/diagnóstico por imagen , Imagenología Tridimensional/métodos , Algoritmos , Aprendizaje Profundo , Interpretación de Imagen Radiográfica Asistida por Computador/métodos
3.
Comput Biol Med ; 177: 108613, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38781644

RESUMEN

Deep learning-based image segmentation and detection models have largely improved the efficiency of analyzing retinal landmarks such as optic disc (OD), optic cup (OC), and fovea. However, factors including ophthalmic disease-related lesions and low image quality issues may severely complicate automatic OD/OC segmentation and fovea detection. Most existing works treat the identification of each landmark as a single task, and take into account no prior information. To address these issues, we propose a prior guided multi-task transformer framework for joint OD/OC segmentation and fovea detection, named JOINEDTrans. JOINEDTrans effectively combines various spatial features of the fundus images, relieving the structural distortions induced by lesions and other imaging issues. It contains a segmentation branch and a detection branch. To be noted, we employ an encoder with prior-learning in a vessel segmentation task to effectively exploit the positional relationship among vessel, OD/OC, and fovea, successfully incorporating spatial prior into the proposed JOINEDTrans framework. There are a coarse stage and a fine stage in JOINEDTrans. In the coarse stage, OD/OC coarse segmentation and fovea heatmap localization are obtained through a joint segmentation and detection module. In the fine stage, we crop regions of interest for subsequent refinement and use predictions obtained in the coarse stage to provide additional information for better performance and faster convergence. Experimental results demonstrate that JOINEDTrans outperforms existing state-of-the-art methods on the publicly available GAMMA, REFUGE, and PALM fundus image datasets. We make our code available at https://github.com/HuaqingHe/JOINEDTrans.


Asunto(s)
Aprendizaje Profundo , Fóvea Central , Disco Óptico , Humanos , Disco Óptico/diagnóstico por imagen , Fóvea Central/diagnóstico por imagen , Interpretación de Imagen Asistida por Computador/métodos , Algoritmos
4.
Sensors (Basel) ; 24(7)2024 Mar 29.
Artículo en Inglés | MEDLINE | ID: mdl-38610414

RESUMEN

In recent years, semantic segmentation has made significant progress in visual place recognition (VPR) by using semantic information that is relatively invariant to appearance and viewpoint, demonstrating great potential. However, in some extreme scenarios, there may be semantic occlusion and semantic sparsity, which can lead to confusion when relying solely on semantic information for localization. Therefore, this paper proposes a novel VPR framework that employs a coarse-to-fine image matching strategy, combining semantic and appearance information to improve algorithm performance. First, we construct SemLook global descriptors using semantic contours, which can preliminarily screen images to enhance the accuracy and real-time performance of the algorithm. Based on this, we introduce SemLook local descriptors for fine screening, combining robust appearance information extracted by deep learning with semantic information. These local descriptors can address issues such as semantic overlap and sparsity in urban environments, further improving the accuracy of the algorithm. Through this refined screening process, we can effectively handle the challenges of complex image matching in urban environments and obtain more accurate results. The performance of SemLook descriptors is evaluated on three public datasets (Extended-CMU Season, Robot-Car Seasons v2, and SYNTHIA) and compared with six state-of-the-art VPR algorithms (HOG, CoHOG, AlexNet_VPR, Region VLAD, Patch-NetVLAD, Forest). In the experimental comparison, considering both real-time performance and evaluation metrics, the SemLook descriptors are found to outperform the other six algorithms. Evaluation metrics include the area under the curve (AUC) based on the precision-recall curve, Recall@100%Precision, and Precision@100%Recall. On the Extended-CMU Season dataset, SemLook descriptors achieve a 100% AUC value, and on the SYNTHIA dataset, they achieve a 99% AUC value, demonstrating outstanding performance. The experimental results indicate that introducing global descriptors for initial screening and utilizing local descriptors combining both semantic and appearance information for precise matching can effectively address the issue of location recognition in scenarios with semantic ambiguity or sparsity. This algorithm enhances descriptor performance, making it more accurate and robust in scenes with variations in appearance and viewpoint.

5.
Int J Comput Assist Radiol Surg ; 19(1): 97-108, 2024 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-37322299

RESUMEN

PURPOSE: Pelvic bone segmentation and landmark definition from computed tomography (CT) images are prerequisite steps for the preoperative planning of total hip arthroplasty. In clinical applications, the diseased pelvic anatomy usually degrades the accuracies of bone segmentation and landmark detection, leading to improper surgery planning and potential operative complications. METHODS: This work proposes a two-stage multi-task algorithm to improve the accuracy of pelvic bone segmentation and landmark detection, especially for the diseased cases. The two-stage framework uses a coarse-to-fine strategy which first conducts global-scale bone segmentation and landmark detection and then focuses on the important local region to further refine the accuracy. For the global stage, a dual-task network is designed to share the common features between the segmentation and detection tasks, so that the two tasks mutually reinforce each other's performance. For the local-scale segmentation, an edge-enhanced dual-task network is designed for simultaneous bone segmentation and edge detection, leading to the more accurate delineation of the acetabulum boundary. RESULTS: This method was evaluated via threefold cross-validation based on 81 CT images (including 31 diseased and 50 healthy cases). The first stage achieved DSC scores of 0.94, 0.97, and 0.97 for the sacrum, left and right hips, respectively, and an average distance error of 3.24 mm for the bone landmarks. The second stage further improved the DSC of the acetabulum by 5.42%, and this accuracy outperforms the state-of-the-arts (SOTA) methods by 0.63%. Our method also accurately segmented the diseased acetabulum boundaries. The entire workflow took ~ 10 s, which was only half of the U-Net run time. CONCLUSION: Using the multi-task networks and the coarse-to-fine strategy, this method achieved more accurate bone segmentation and landmark detection than the SOTA method, especially for diseased hip images. Our work contributes to accurate and rapid design of acetabular cup prostheses.


Asunto(s)
Aprendizaje Profundo , Humanos , Tomografía Computarizada por Rayos X/métodos , Cadera , Pelvis/diagnóstico por imagen , Acetábulo , Procesamiento de Imagen Asistido por Computador/métodos
6.
J Comput Chem ; 45(8): 487-497, 2024 Mar 30.
Artículo en Inglés | MEDLINE | ID: mdl-37966714

RESUMEN

Transition state (TS) on the potential energy surface (PES) plays a key role in determining the kinetics and thermodynamics of chemical reactions. Inspired by the fact that the dynamics of complex systems are always driven by rare but significant transition events, we herein propose a TS search method in accordance with the Q-learning algorithm. Appropriate reward functions are set for a given PES to optimize the reaction pathway through continuous trial and error, and then the TS can be obtained from the optimized reaction pathway. The validity of this Q-learning method with reasonable settings of Q-value table including actions, states, learning rate, greedy rate, discount rate, and so on, is exemplified in 2 two-dimensional potential functions. In the applications of the Q-learning method to two chemical reactions, it is demonstrated that the Q-learning method can predict consistent TS and reaction pathway with those by ab initio calculations. Notably, the PES must be well prepared before using the Q-learning method, and a coarse-to-fine PES scanning scheme is thus introduced to save the computational time while maintaining the accuracy of the Q-learning prediction. This work offers a simple and reliable Q-learning method to search for all possible TS and reaction pathway of a chemical reaction, which may be a new option for effectively exploring the PES in an extensive search manner.

7.
Med Image Anal ; 91: 103038, 2024 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-38000258

RESUMEN

Deformable image registration, the estimation of the spatial transformation between different images, is an important task in medical imaging. Deep learning techniques have been shown to perform 3D image registration efficiently. However, current registration strategies often only focus on the deformation smoothness, which leads to the ignorance of complicated motion patterns (e.g., separate or sliding motions), especially for the intersection of organs. Thus, the performance when dealing with the discontinuous motions of multiple nearby objects is limited, causing undesired predictive outcomes in clinical usage, such as misidentification and mislocalization of lesions or other abnormalities. Consequently, we proposed a novel registration method to address this issue: a new Motion Separable backbone is exploited to capture the separate motion, with a theoretical analysis of the upper bound of the motions' discontinuity provided. In addition, a novel Residual Aligner module was used to disentangle and refine the predicted motions across the multiple neighboring objects/organs. We evaluate our method, Residual Aligner-based Network (RAN), on abdominal Computed Tomography (CT) scans and it has shown to achieve one of the most accurate unsupervised inter-subject registration for the 9 organs, with the highest-ranked registration of the veins (Dice Similarity Coefficient (%)/Average surface distance (mm): 62%/4.9mm for the vena cava and 34%/7.9mm for the portal and splenic vein), with a smaller model structure and less computation compared to state-of-the-art methods. Furthermore, when applied to lung CT, the RAN achieves comparable results to the best-ranked networks (94%/3.0mm), also with fewer parameters and less computation.


Asunto(s)
Algoritmos , Tomografía Computarizada por Rayos X , Humanos , Tomografía Computarizada por Rayos X/métodos , Movimiento (Física) , Pulmón/diagnóstico por imagen , Imagenología Tridimensional , Procesamiento de Imagen Asistido por Computador/métodos
8.
Sensors (Basel) ; 23(24)2023 Dec 15.
Artículo en Inglés | MEDLINE | ID: mdl-38139683

RESUMEN

Point cloud registration is widely used in autonomous driving, SLAM, and 3D reconstruction, and it aims to align point clouds from different viewpoints or poses under the same coordinate system. However, point cloud registration is challenging in complex situations, such as a large initial pose difference, high noise, or incomplete overlap, which will cause point cloud registration failure or mismatching. To address the shortcomings of the existing registration algorithms, this paper designed a new coarse-to-fine registration two-stage point cloud registration network, CCRNet, which utilizes an end-to-end form to perform the registration task for point clouds. The multi-scale feature extraction module, coarse registration prediction module, and fine registration prediction module designed in this paper can robustly and accurately register two point clouds without iterations. CCRNet can link the feature information between two point clouds and solve the problems of high noise and incomplete overlap by using a soft correspondence matrix. In the standard dataset ModelNet40, in cases of large initial pose difference, high noise, and incomplete overlap, the accuracy of our method, compared with the second-best popular registration algorithm, was improved by 7.0%, 7.8%, and 22.7% on the MAE, respectively. Experiments showed that our CCRNet method has advantages in registration results in a variety of complex conditions.

9.
Sensors (Basel) ; 23(18)2023 Sep 18.
Artículo en Inglés | MEDLINE | ID: mdl-37766021

RESUMEN

Object-level simultaneous localization and mapping (SLAM) has gained popularity in recent years since it can provide a means for intelligent robot-to-environment interactions. However, most of these methods assume that the distribution of the errors is Gaussian. This assumption is not valid under many circumstances. Further, these methods use a delayed initialization of the objects in the map. During this delayed period, the solution relies on the motion model provided by an inertial measurement unit (IMU). Unfortunately, the errors tend to accumulate quickly due to the dead-reckoning nature of these motion models. Finally, the current solutions depend on a set of salient features on the object's surface and not the object's shape. This research proposes an accurate object-level solution to the SLAM problem with a 4.1 to 13.1 cm error in the position (0.005 to 0.021 of the total path). The developed solution is based on Rao-Blackwellized Particle Filtering (RBPF) that does not assume any predefined error distribution for the parameters. Further, the solution relies on the shape and thus can be used for objects that lack texture on their surface. Finally, the developed tightly coupled IMU/camera solution is based on an undelayed initialization of the objects in the map.

10.
Neural Netw ; 165: 774-785, 2023 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-37418860

RESUMEN

Image registration is a fundamental problem in computer vision and robotics. Recently, learning-based image registration methods have made great progress. However, these methods are sensitive to abnormal transformation and have insufficient robustness, which leads to more mismatched points in the actual environment. In this paper, we propose a new registration framework based on ensemble learning and dynamic adaptive kernel. Specifically, we first use a dynamic adaptive kernel to extract deep features at the coarse level to guide fine-level registration. Then we added an adaptive feature pyramid network based on the integrated learning principle to realize the fine-level feature extraction. Through different scale, receptive fields, not only the local geometric information of each point is considered, but also its low texture information at the pixel level is considered. According to the actual registration environment, fine features are adaptively obtained to reduce the sensitivity of the model to abnormal transformation. We use the global receptive field provided in the transformer to obtain feature descriptors based on these two levels. In addition, we use the cosine loss directly defined on the corresponding relationship to train the network and balance the samples, to achieve feature point registration based on the corresponding relationship. Extensive experiments on object-level and scene-level datasets show that the proposed method outperforms existing state-of-the-art techniques by a large margin. More critically, it has the best generalization ability in unknown scenes with different sensor modes.


Asunto(s)
Aprendizaje , Robótica , Generalización Psicológica , Procesamiento de Imagen Asistido por Computador
11.
Psychon Bull Rev ; 30(6): 2230-2239, 2023 Dec.
Artículo en Inglés | MEDLINE | ID: mdl-37221279

RESUMEN

Action representation of man-made tools consists of two subtypes: structural action representation concerning how to grasp an object, and functional action representation concerning the skilled use of an object. Compared to structural action representation, functional action representation plays the dominant role in fine-grained (i.e., basic level) object recognition. However, it remains unclear whether the two types of action representation are involved differently in the coarse semantic processing in which the object is recognized at a superordinate level (i.e., living/non-living). Here we conducted three experiments using the priming paradigm, in which video clips displaying structural and functional action hand gestures were used as prime stimuli and grayscale photos of man-made tools were used as target stimuli. Participants recognized the target objects at the basic level in Experiment 1 (i.e., naming task) and at the superordinate level in Experiments 2 and 3 (i.e., categorization task). We observed a significant priming effect for functional action prime-target pairs only in the naming task. In contrast, no priming effect was found in either the naming or the categorization task for the structural action prime-target pairs (Experiment 2), even when the categorization task was preceded by a preliminary action imitation of the prime gestures (Experiment 3). Our results suggest that only functional action information is retrieved during fine-grained object processing. In contrast, coarse semantic processing does not require the integration of either structural or functional action information.


Asunto(s)
Semántica , Percepción Visual , Humanos , Mapeo Encefálico , Gestos , Tiempo de Reacción
12.
Neuroimage ; 274: 120139, 2023 07 01.
Artículo en Inglés | MEDLINE | ID: mdl-37137434

RESUMEN

Natural images exhibit luminance variations aligned across a broad spectrum of spatial frequencies (SFs). It has been proposed that, at early stages of processing, the coarse signals carried by the low SF (LSF) of the visual input are sent rapidly from primary visual cortex (V1) to ventral, dorsal and frontal regions to form a coarse representation of the input, which is later sent back to V1 to guide the processing of fine-grained high SFs (i.e., HSF). We used functional resonance imaging (fMRI) to investigate the role of human V1 in the coarse-to-fine integration of visual input. We disrupted the processing of the coarse and fine content of full-spectrum human face stimuli via backward masking of selective SF ranges (LSFs: <1.75cpd and HSFs: >1.75cpd) at specific times (50, 83, 100 or 150 ms). In line with coarse-to-fine proposals, we found that (1) the selective masking of stimulus LSF disrupted V1 activity in the earliest time window, and progressively decreased in influence, while (2) an opposite trend was observed for the masking of stimulus' HSF. This pattern of activity was found in V1, as well as in ventral (i.e. the Fusiform Face area, FFA), dorsal and orbitofrontal regions. We additionally presented subjects with contrast negated stimuli. While contrast negation significantly reduced response amplitudes in the FFA, as well as coupling between FFA and V1, coarse-to-fine dynamics were not affected by this manipulation. The fact that V1 response dynamics to strictly identical stimulus sets differed depending on the masked scale adds to growing evidence that V1 role goes beyond the early and quasi-passive transmission of visual information to the rest of the brain. It instead indicates that V1 may yield a 'spatially registered common forum' or 'blackboard' that integrates top-down inferences with incoming visual signals through its recurrent interaction with high-level regions located in the inferotemporal, dorsal and frontal regions.


Asunto(s)
Corteza Prefrontal , Visión Ocular , Humanos , Imagen por Resonancia Magnética/métodos , Estimulación Luminosa/métodos , Análisis de Varianza
13.
Comput Biol Med ; 154: 106543, 2023 03.
Artículo en Inglés | MEDLINE | ID: mdl-36682179

RESUMEN

To facilitate both the detection and the interpretation of findings in chest X-rays, comparison with a previous image of the same patient is very valuable to radiologists. Today, the most common approach for deep learning methods to automatically inspect chest X-rays disregards the patient history and classifies only single images as normal or abnormal. Nevertheless, several methods for assisting in the task of comparison through image registration have been proposed in the past. However, as we illustrate, they tend to miss specific types of pathological changes like cardiomegaly and effusion. Due to assumptions on fixed anatomical structures or their measurements of registration quality, they produce unnaturally deformed warp fields impacting visualization of differences between moving and fixed images. We aim to overcome these limitations, through a new paradigm based on individual rib pair segmentation for anatomy penalized registration. Our method proves to be a natural way to limit the folding percentage of the warp field to 1/6 of the state of the art while increasing the overlap of ribs by more than 25%, implying difference images showing pathological changes overlooked by other methods. We develop an anatomically penalized convolutional multi-stage solution on the National Institutes of Health (NIH) data set, starting from less than 25 fully and 50 partly labeled training images, employing sequential instance memory segmentation with hole dropout, weak labeling, coarse-to-fine refinement and Gaussian mixture model histogram matching. We statistically evaluate the benefits of our method and highlight the limits of currently used metrics for registration of chest X-rays.


Asunto(s)
Interpretación de Imagen Radiográfica Asistida por Computador , Tomografía Computarizada por Rayos X , Humanos , Rayos X , Radiografía , Tomografía Computarizada por Rayos X/métodos , Interpretación de Imagen Radiográfica Asistida por Computador/métodos , Costillas/diagnóstico por imagen , Procesamiento de Imagen Asistido por Computador/métodos
14.
Vision Res ; 204: 108165, 2023 03.
Artículo en Inglés | MEDLINE | ID: mdl-36584582

RESUMEN

Rapid analysis of low spatial frequencies (LSFs) in the brain conveys the global shape of the object and allows for rapid expectations about the visual input. Evidence has suggested that LSF processing differs as a function of the semantic category to identify. The present study sought to specify the neural dynamics of the LSF contribution to the rapid object representation of living versus non-living objects. In this EEG experiment, participants had to categorize an object displayed at different spatial frequencies (LSF or non-filtered). Behavioral results showed an advantage for living versus non-living objects and a decrease in performance with LSF pictures of pieces of furniture only. Moreover, despite a difference in classification performance between LSF and non-filtered pictures for living items, the behavioral performance was maintained, which suggests that classification under our specific condition can be based on LSF information, in particular for living items.


Asunto(s)
Encéfalo , Reconocimiento Visual de Modelos , Humanos
15.
Phys Med Biol ; 68(2)2023 01 05.
Artículo en Inglés | MEDLINE | ID: mdl-36535028

RESUMEN

Delineation of brain metastases (BMs) is a paramount step in stereotactic radiosurgery treatment. Clinical practice has specific expectation on BM auto-delineation that the method is supposed to avoid missing of small lesions and yield accurate contours for large lesions. In this study, we propose a novel coarse-to-fine framework, named detector-based segmentation (DeSeg), to incorporate object-level detection into pixel-wise segmentation so as to meet the clinical demand. DeSeg consists of three components: a center-point-guided single-shot detector to localize the potential lesion regions, a multi-head U-Net segmentation model to refine contours, and a data cascade unit to connect both tasks smoothly. Performance on tiny lesions is measured by the object-based sensitivity and positive predictive value (PPV), while that on large lesions is quantified by dice similarity coefficient (DSC), average symmetric surface distance (ASSD) and 95% Hausdorff distance (HD95). Besides, computational complexity is also considered to study the potential of method in real-time processing. This study retrospectively collected 240 BM patients with Gadolinium injected contrast-enhanced T1-weighted magnetic resonance imaging (T1c-MRI), which were randomly split into training, validating and testing datasets (192, 24 and 24 scans, respectively). The lesions in the testing dataset were further divided into two groups based on the volume size (smallS: ≤1.5 cc,N= 88; largeL: > 1.5 cc,N= 15). On average, DeSeg yielded a sensitivity of 0.91 and a PPV of 0.77 on S group, and a DSC of 0.86, an ASSD 0f 0.76 mm and a HD95 of 2.31 mm onLgroup. The results indicated that DeSeg achieved leading sensitivity and PPV for tiny lesions as well as segmentation metrics for large ones. After our clinical validation, DeSeg showed competitive segmentation performance while kept faster processing speed comparing with existing 3D models.


Asunto(s)
Neoplasias Encefálicas , Radiocirugia , Humanos , Estudios Retrospectivos , Neoplasias Encefálicas/diagnóstico por imagen , Neoplasias Encefálicas/secundario , Imagen por Resonancia Magnética/métodos , Radiocirugia/métodos
16.
Comput Biol Med ; 152: 106410, 2023 01.
Artículo en Inglés | MEDLINE | ID: mdl-36516578

RESUMEN

Accurate and automatic pancreas segmentation from abdominal computed tomography (CT) scans is crucial for the diagnosis and prognosis of pancreatic diseases. However, the pancreas accounts for a relatively small portion of the scan and presents high anatomical variability and low contrast, making traditional automated segmentation methods fail to generate satisfactory results. In this paper, we propose an extension-contraction transformation network (ECTN) and deploy it into a cascaded two-stage segmentation framework for accurate pancreas segmenting. This model can enhance the perception of 3D context by distinguishing and exploiting the extension and contraction transformation of the pancreas between slices. It consists of an encoder, a segmentation decoder, and an extension-contraction (EC) decoder. The EC decoder is responsible for predicting the inter-slice extension and contraction transformation of the pancreas by feeding the extension and contraction information generated by the segmentation decoder; meanwhile, its output is combined with the output of the segmentation decoder to reconstruct and refine the segmentation results. Quantitative evaluation is performed on NIH Pancreas Segmentation (Pancreas-CT) dataset using 4-fold cross-validation. We obtained average Precision of 86.59±6.14% , Recall of 85.11±5.96%, Dice similarity coefficient (DSC) of 85.58±3.98%. and Jaccard Index (JI) of 74.99±5.86%. The performance of our method outperforms several baseline and state-of-the-art methods.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Páncreas , Procesamiento de Imagen Asistido por Computador/métodos , Páncreas/diagnóstico por imagen , Abdomen , Tomografía Computarizada por Rayos X/métodos
17.
Cells ; 11(24)2022 12 17.
Artículo en Inglés | MEDLINE | ID: mdl-36552872

RESUMEN

3D point clouds are gradually becoming more widely used in the medical field, however, they are rarely used for 3D representation of intracranial vessels and aneurysms due to the time-consuming data reconstruction. In this paper, we simulate the incomplete intracranial vessels (including aneurysms) in the actual collection from different angles, then propose Multi-Scope Feature Extraction Network (MSENet) for Intracranial Aneurysm 3D Point Cloud Completion. MSENet adopts a multi-scope feature extraction encoder to extract the global features from the incomplete point cloud. This encoder utilizes different scopes to fuse the neighborhood information for each point fully. Then a folding-based decoder is applied to obtain the complete 3D shape. To enable the decoder to intuitively match the original geometric structure, we engage the original points coordinates input to perform residual linking. Finally, we merge and sample the complete but coarse point cloud from the decoder to obtain the final refined complete 3D point cloud shape. We conduct extensive experiments on both 3D intracranial aneurysm datasets and general 3D vision PCN datasets. The results demonstrate the effectiveness of the proposed method on three evaluation metrics compared to baseline: our model increases the F-score to 0.379 (+21.1%)/0.320 (+7.7%), reduces Chamfer Distance score to 0.998 (-33.8%)/0.974 (-6.4%), and reduces the Earth Mover's Distance to 2.750 (17.8%)/2.858 (-0.8%).


Asunto(s)
Aneurisma Intracraneal , Humanos
18.
Sensors (Basel) ; 22(18)2022 Sep 08.
Artículo en Inglés | MEDLINE | ID: mdl-36146148

RESUMEN

Low-light image enhancement can effectively assist high-level vision tasks that often fail in poor illumination conditions. Most previous data-driven methods, however, implemented enhancement directly from severely degraded low-light images that may provide undesirable enhancement results, including blurred detail, intensive noise, and distorted color. In this paper, inspired by a coarse-to-fine strategy, we propose an end-to-end image-level alignment with pixel-wise perceptual information enhancement pipeline for low-light image enhancement. A coarse adaptive global photometric alignment sub-network is constructed to reduce style differences, which facilitates improving illumination and revealing under-exposure area information. After the learned aligned image, a hierarchy pyramid enhancement sub-network is used to optimize image quality, which helps to remove amplified noise and enhance the local detail of low-light images. We also propose a multi-residual cascade attention block (MRCAB) that involves channel split and concatenation strategy, polarized self-attention mechanism, which leads to high-resolution reconstruction images in perceptual quality. Extensive experiments have demonstrated the effectiveness of our method on various datasets and significantly outperformed other state-of-the-art methods in detail and color reproduction.

19.
J Neurosci ; 42(37): 7047-7059, 2022 09 14.
Artículo en Inglés | MEDLINE | ID: mdl-35927035

RESUMEN

The spatial-frequency (SF) tuning of neurons in the early visual cortex is adjusted for stimulus contrast. As the contrast increases, SF tuning is modulated so that the transmission of fine features is facilitated. A variety of mechanisms are involved in shaping SF tunings, but those responsible for the contrast-dependent modulations are unclear. To address this, we measured the time course of SF tunings of area 17 neurons in male cats under different contrasts with a reverse correlation. After response onset, the optimal SF continuously shifted to a higher SF over time, with a larger shift for higher contrast. At high contrast, whereas neurons with a large shift of optimal SF exhibited a large bandwidth decrease, those with a negligible shift increased the bandwidth over time. Between these two extremes, the degree of SF shift and bandwidth change continuously varied. At low contrast, bandwidth generally decreased over time. These dynamic effects enhanced the processing of high-frequency range under a high-contrast condition and allowed time-average SF tuning curves to show contrast-dependent modulation, like that of steady-state SF tuning curves reported previously. Combinations of two mechanisms, one that decreases bandwidth and shifts optimal SF, and another that increases bandwidth without shifting optimal SF, would explain the full range of SF tuning dynamics. Our results indicate that one of the essential roles of tuning dynamics of area 17 neurons, which have been observed for various visual features, is to adjust tunings depending on contrast.SIGNIFICANCE STATEMENT The spatial scales of features transmitted by cortical neurons are adjusted depending on stimulus contrast. However, the underlying mechanism is not fully understood. We measured the time course of spatial frequency tunings of cat area 17 neurons under different contrast conditions and observed a variety of dynamic effects that contributed to spatial-scale adjustment, allowing neurons to adjust their spatial frequency tuning range depending on contrast. Our results suggest that one of the essential roles of tuning dynamics of area 17 neurons, which have been observed for various visual features, is to adjust tunings depending on contrast.


Asunto(s)
Corteza Visual , Animales , Masculino , Neuronas/fisiología , Estimulación Luminosa/métodos , Corteza Visual/fisiología
20.
Diagnostics (Basel) ; 12(8)2022 Aug 07.
Artículo en Inglés | MEDLINE | ID: mdl-36010263

RESUMEN

In intensive care units (ICUs), after endotracheal intubation, the position of the endotracheal tube (ETT) should be checked to avoid complications. The malposition can be detected by the distance between the ETT tip and the Carina (ETT-Carina distance). However, it struggles with a limited performance for two major problems, i.e., occlusion by external machine, and the posture and machine of taking chest radiographs. While previous studies addressed these problems, they always suffered from the requirements of manual intervention. Therefore, the purpose of this paper is to locate the ETT tip and the Carina more accurately for detecting the malposition without manual intervention. The proposed architecture is composed of FCOS: Fully Convolutional One-Stage Object Detection, an attention mechanism named Coarse-to-Fine Attention (CTFA), and a segmentation branch. Moreover, a post-process algorithm is adopted to select the final location of the ETT tip and the Carina. Three metrics were used to evaluate the performance of the proposed method. With the dataset provided by National Cheng Kung University Hospital, the accuracy of the malposition detected by the proposed method achieves 88.82% and the ETT-Carina distance errors are less than 5.333±6.240 mm.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA