Deep learning
Imputing spatial transcriptomics through gene network constructed from protein language model
Commun Biol. 2024 Oct 5;7(1):1271. doi: 10.1038/s42003-024-06964-2.
ABSTRACT
Image-based spatial transcriptomic sequencing technologies have enabled the measurement of gene expression at single-cell resolution, but with a limited number of genes. Current computational approaches attempt to overcome these limitations by imputing missing genes, but face challenges regarding prediction accuracy and identification of cell populations due to the neglect of gene-gene relationships. In this context, we present stImpute, a method to impute spatial transcriptomics according to reference scRNA-seq data based on the gene network constructed from the protein language model ESM-2. Specifically, stImpute employs an autoencoder to create gene expression embeddings for both spatial transcriptomics and scRNA-seq data, which are used to identify the nearest neighboring cells between scRNA-seq and spatial transcriptomics datasets. According to the neighbored cells, the gene expressions of spatial transcriptomics cells are imputed through a graph neural network, where nodes are genes, and edges are based on cosine similarity between the ESM-2 embeddings of the gene-encoding proteins. The gene prediction uncertainty is further measured through a deep learning model. stImpute was shown to consistently outperform state-of-the-art methods across multiple datasets concerning imputation and clustering. stImpute also demonstrates robustness in producing consistent results that are insensitive to model parameters.
PMID:39369061 | DOI:10.1038/s42003-024-06964-2
Significantly improving zero-shot X-ray pathology classification via fine-tuning pre-trained image-text encoders
Sci Rep. 2024 Oct 5;14(1):23199. doi: 10.1038/s41598-024-73695-z.
ABSTRACT
Deep neural networks are increasingly used in medical imaging for tasks such as pathological classification, but they face challenges due to the scarcity of high-quality, expert-labeled training data. Recent efforts have utilized pre-trained contrastive image-text models like CLIP, adapting them for medical use by fine-tuning the model with chest X-ray images and corresponding reports for zero-shot pathology classification, thus eliminating the need for pathology-specific annotations. However, most studies continue to use the same contrastive learning objectives as in the general domain, overlooking the multi-labeled nature of medical image-report pairs. In this paper, we propose a new fine-tuning strategy that includes positive-pair loss relaxation and random sentence sampling. We aim to improve the performance of zero-shot pathology classification without relying on external knowledge. Our method can be applied to any pre-trained contrastive image-text encoder and easily transferred to out-of-domain datasets without further training, as it does not use external data. Our approach consistently improves overall zero-shot pathology classification across four chest X-ray datasets and three pre-trained models, with an average macro AUROC increase of 4.3%. Additionally, our method outperforms the state-of-the-art and marginally surpasses board-certified radiologists in zero-shot classification for the five competition pathologies in the CheXpert dataset.
PMID:39369048 | DOI:10.1038/s41598-024-73695-z
A lighter hybrid feature fusion framework for polyp segmentation
Sci Rep. 2024 Oct 5;14(1):23179. doi: 10.1038/s41598-024-72763-8.
ABSTRACT
Colonoscopy is widely recognized as the most effective method for the detection of colon polyps, which is crucial for early screening of colorectal cancer. Polyp identification and segmentation in colonoscopy images require specialized medical knowledge and are often labor-intensive and expensive. Deep learning provides an intelligent and efficient approach for polyp segmentation. However, the variability in polyp size and the heterogeneity of polyp boundaries and interiors pose challenges for accurate segmentation. Currently, Transformer-based methods have become a mainstream trend for polyp segmentation. However, these methods tend to overlook local details due to the inherent characteristics of Transformer, leading to inferior results. Moreover, the computational burden brought by self-attention mechanisms hinders the practical application of these models. To address these issues, we propose a novel CNN-Transformer hybrid model for polyp segmentation (CTHP). CTHP combines the strengths of CNN, which excels at modeling local information, and Transformer, which excels at modeling global semantics, to enhance segmentation accuracy. We transform the self-attention computation over the entire feature map into the width and height directions, significantly improving computational efficiency. Additionally, we design a new information propagation module and introduce additional positional bias coefficients during the attention computation process, which reduces the dispersal of information introduced by deep and mixed feature fusion in the Transformer. Extensive experimental results demonstrate that our proposed model achieves state-of-the-art performance on multiple benchmark datasets for polyp segmentation. Furthermore, cross-domain generalization experiments show that our model exhibits excellent generalization performance.
PMID:39369043 | DOI:10.1038/s41598-024-72763-8
Unsupervised few shot learning architecture for diagnosis of periodontal disease in dental panoramic radiographs
Sci Rep. 2024 Oct 5;14(1):23237. doi: 10.1038/s41598-024-73665-5.
ABSTRACT
In the domain of medical imaging, the advent of deep learning has marked a significant progression, particularly in the nuanced area of periodontal disease diagnosis. This study specifically targets the prevalent issue of scarce labeled data in medical imaging. We introduce a novel unsupervised few-shot learning algorithm, meticulously crafted for classifying periodontal diseases using a limited collection of dental panoramic radiographs. Our method leverages UNet architecture for generating regions of interest (RoI) from radiographs, which are then processed through a Convolutional Variational Autoencoder (CVAE). This approach is pivotal in extracting critical latent features, subsequently clustered using an advanced algorithm. This clustering is key in our methodology, enabling the assignment of labels to images indicative of periodontal diseases, thus circumventing the challenges posed by limited datasets. Our validation process, involving a comparative analysis with traditional supervised learning and standard autoencoder-based clustering, demonstrates a marked improvement in both diagnostic accuracy and efficiency. For three real-world validation datasets, our UNet-CVAE architecture achieved up to average 14% higher accuracy compared to state-of-the-art supervised models including the vision transformer model when trained with 100 labeled images. This study not only highlights the capability of unsupervised learning in overcoming data limitations but also sets a new benchmark for diagnostic methodologies in medical AI, potentially transforming practices in data-constrained scenarios.
PMID:39369017 | DOI:10.1038/s41598-024-73665-5
Differentiable modeling and optimization of non-aqueous Li-based battery electrolyte solutions using geometric deep learning
Nat Commun. 2024 Oct 5;15(1):8649. doi: 10.1038/s41467-024-51653-7.
ABSTRACT
Electrolytes play a critical role in designing next-generation battery systems, by allowing efficient ion transfer, preventing charge transfer, and stabilizing electrode-electrolyte interfaces. In this work, we develop a differentiable geometric deep learning (GDL) model for chemical mixtures, DiffMix, which is applied in guiding robotic experimentation and optimization towards fast-charging battery electrolytes. In particular, we extend mixture thermodynamic and transport laws by creating GDL-learnable physical coefficients. We evaluate our model with mixture thermodynamics and ion transport properties, where we show improved prediction accuracy and model robustness of DiffMix than its purely data-driven variants. Furthermore, with a robotic experimentation setup, Clio, we improve ionic conductivity of electrolytes by over 18.8% within 10 experimental steps, via differentiable optimization built on DiffMix gradients. By combining GDL, mixture physics laws, and robotic experimentation, DiffMix expands the predictive modeling methods for chemical mixtures and enables efficient optimization in large chemical spaces.
PMID:39369004 | DOI:10.1038/s41467-024-51653-7
LungHist700: A dataset of histological images for deep learning in pulmonary pathology
Sci Data. 2024 Oct 5;11(1):1088. doi: 10.1038/s41597-024-03944-3.
ABSTRACT
Accurate detection and classification of lung malignancies are crucial for early diagnosis, treatment planning, and patient prognosis. Conventional histopathological analysis is time-consuming, limiting its clinical applicability. To address this, we present a dataset of 691 high-resolution (1200 × 1600 pixels) histopathological lung images, covering adenocarcinomas, squamous cell carcinomas, and normal tissues from 45 patients. These images are subdivided into three differentiation levels for both pathological types: well, moderately, and poorly differentiated, resulting in seven classes for classification. The dataset includes images at 20x and 40x magnification, reflecting real clinical diversity. We evaluated image classification using deep neural network and multiple instance learning approaches. Each method was used to classify images at 20x and 40x magnification into three superclasses. We achieved accuracies between 81% and 92%, depending on the method and resolution, demonstrating the dataset's utility.
PMID:39368979 | DOI:10.1038/s41597-024-03944-3
Enhanced image quality and lesion detection in FLAIR MRI of white matter hyperintensity through deep learning-based reconstruction
Asian J Surg. 2024 Oct 4:S1015-9584(24)02201-2. doi: 10.1016/j.asjsur.2024.09.156. Online ahead of print.
ABSTRACT
OBJECTIVE: To delve deeper into the study of degenerative diseases, it becomes imperative to investigate whether deep-learning reconstruction (DLR) can improve the evaluation of white matter hyperintensity (WMH) on 3.0T scanners, and compare its lesion detection capabilities with conventional reconstruction (CR).
METHODS: A total of 131 participants (mean age, 46 years ±17; 46 men) were included in the study. The images of these participants were evaluated by readers blinded to clinical data. Two readers independently assessed subjective image indicators on a 4-point scale. The severity of WMH was assessed by four raters using the Fazekas scale. To evaluate the relative detection capabilities of each method, we employed the Wilcoxon signed rank test to compare scores between the DLR and the CR group. Additionally, we assessed interrater reliability using weighted k statistics and intraclass correlation coefficient to test consistency among the raters.
RESULTS: In terms of subjective image scoring, the DLR group exhibited significantly better scores compared to the CR group (P < 0.001). Regarding the severity of WMH, the DL group demonstrated superior performance in detecting lesions. Majority readers agreed that the DL group provided clearer visualization of the lesions compared to the conventional group.
CONCLUSION: DLR exhibits notable advantages over CR, including subjective image quality, lesion detection sensitivity, and inter reader reliability.
PMID:39368951 | DOI:10.1016/j.asjsur.2024.09.156
Deep learning-based characterization of pathological subtypes in lung invasive adenocarcinoma utilizing (18)F-deoxyglucose positron emission tomography imaging
BMC Cancer. 2024 Oct 5;24(1):1229. doi: 10.1186/s12885-024-13018-7.
ABSTRACT
OBJECTIVE: To evaluate the diagnostic efficacy of a deep learning (DL) model based on PET/CT images for distinguishing and predicting various pathological subtypes of invasive lung adenocarcinoma.
METHODS: A total of 250 patients diagnosed with invasive lung cancer were included in this retrospective study. The pathological subtypes of the cancer were recorded. PET/CT images were analyzed, including measurements and recordings of the short and long diameters on the maximum cross-sectional plane of the CT image, the density of the lesion, and the associated imaging signs. The SUVmax, SUVmean, and the lesion's long and short diameters on the PET image were also measured. A manual diagnostic model was constructed to analyze its diagnostic performance across different pathological subtypes. The acquired images were first denoised, followed by data augmentation to expand the dataset. The U-Net network architecture was then employed for feature extraction and network segmentation. The classification network was based on the ResNet residual network to address the issue of gradient vanishing in deep networks. Batch normalization was applied to ensure the feature matrix followed a distribution with a mean of 0 and a variance of 1. The images were divided into training, validation, and test sets in a ratio of 6:2:2 to train the model. The deep learning model was then constructed to analyze its diagnostic performance across different pathological subtypes.
RESULTS: Statistically significant differences (P < 0.05) were observed among the four different subtypes in PET/CT imaging performance. The AUC and diagnostic accuracy of the manual diagnostic model for different pathological subtypes were as follows: APA: 0.647, 0.664; SPA: 0.737, 0.772; PPA: 0.698, 0.780; LPA: 0.849, 0.904. Chi-square tests indicated significant statistical differences among these subtypes (P < 0.05). The AUC and diagnostic accuracy of the deep learning model for the different pathological subtypes were as follows: APA: 0.854, 0.864; SPA: 0.930, 0.936; PPA: 0.878, 0.888; LPA: 0.900, 0.920. Chi-square tests also indicated significant statistical differences among these subtypes (P < 0.05). The Delong test showed that the diagnostic performance of the deep learning model was superior to that of the manual diagnostic model (P < 0.05).
CONCLUSIONS: The deep learning model based on PET/CT images exhibits high diagnostic efficacy in distinguishing and diagnosing various pathological subtypes of invasive lung adenocarcinoma, demonstrating the significant potential of deep learning techniques in accurately identifying and predicting disease subgroups.
PMID:39369213 | DOI:10.1186/s12885-024-13018-7
Artificial intelligence for detection and characterization of focal hepatic lesions: a review
Abdom Radiol (NY). 2024 Oct 5. doi: 10.1007/s00261-024-04597-x. Online ahead of print.
ABSTRACT
Focal liver lesions (FLL) are common incidental findings in abdominal imaging. While the majority of FLLs are benign and asymptomatic, some can be malignant or pre-malignant, and need accurate detection and classification. Current imaging techniques, such as computed tomography (CT) and magnetic resonance imaging (MRI), play a crucial role in assessing these lesions. Artificial intelligence (AI), particularly deep learning (DL), offers potential solutions by analyzing large data to identify patterns and extract clinical features that aid in the early detection and classification of FLLs. This manuscript reviews the diagnostic capacity of AI-based algorithms in processing CT and MRIs to detect benign and malignant FLLs, with an emphasis in the characterization and classification of these lesions and focusing on differentiating benign from pre-malignant and potentially malignant lesions. A comprehensive literature search from January 2010 to April 2024 identified 45 relevant studies. The majority of AI systems employed convolutional neural networks (CNNs), with expert radiologists providing reference standards through manual lesion delineation, and histology as the gold standard. The studies reviewed indicate that AI-based algorithms demonstrate high accuracy, sensitivity, specificity, and AUCs in detecting and characterizing FLLs. These algorithms excel in differentiating between benign and malignant lesions, optimizing diagnostic protocols, and reducing the needs of invasive procedures. Future research should concentrate on the expansion of data sets, the improvement of model explainability, and the validation of AI tools across a range of clinical setting to ensure the applicability and reliability of such tools.
PMID:39369107 | DOI:10.1007/s00261-024-04597-x
A novel multi-scale network intrusion detection model with transformer
Sci Rep. 2024 Oct 5;14(1):23239. doi: 10.1038/s41598-024-74214-w.
ABSTRACT
Network is an essential tool today, and the Intrusion Detection System (IDS) can ensure the safe operation. However, with the explosive growth of data, current methods are increasingly struggling as they often detect based on a single scale, leading to the oversight of potential features in the extensive traffic data, which may result in degraded performance. In this work, we propose a novel detection model utilizing multi-scale transformer namely IDS-MTran. In essence, the collaboration of multi-scale traffic features broads the pattern coverage of intrusion detection. Firstly, we employ convolution operators with various kernels to generate multi-scale features. Secondly, to enhance the representation of features and the interaction between branches, we propose Patching with Pooling (PwP) to serve as a bridge. Next, we design multi-scale transformer-based backbone to model the features at diverse scales, extracting potential intrusion trails. Finally, to fully capitalize these multi-scale branches, we propose the Cross Feature Enrichment (CFE) to integrate and enrich features, and then output the results. Sufficient experiments show that compared with other models, the proposed method can distinguish different attack types more effectively. Specifically, the accuracy on three common datasets NSL-KDD, CIC-DDoS 2019 and UNSW-NB15 has all exceeded 99%, which is more accurate and stable.
PMID:39369065 | DOI:10.1038/s41598-024-74214-w
An initial game-theoretic assessment of enhanced tissue preparation and imaging protocols for improved deep learning inference of spatial transcriptomics from tissue morphology
Brief Bioinform. 2024 Sep 23;25(6):bbae476. doi: 10.1093/bib/bbae476.
ABSTRACT
The application of deep learning to spatial transcriptomics (ST) can reveal relationships between gene expression and tissue architecture. Prior work has demonstrated that inferring gene expression from tissue histomorphology can discern these spatial molecular markers to enable population scale studies, reducing the fiscal barriers associated with large-scale spatial profiling. However, while most improvements in algorithmic performance have focused on improving model architectures, little is known about how the quality of tissue preparation and imaging can affect deep learning model training for spatial inference from morphology and its potential for widespread clinical adoption. Prior studies for ST inference from histology typically utilize manually stained frozen sections with imaging on non-clinical grade scanners. Training such models on ST cohorts is also costly. We hypothesize that adopting tissue processing and imaging practices that mirror standards for clinical implementation (permanent sections, automated tissue staining, and clinical grade scanning) can significantly improve model performance. An enhanced specimen processing and imaging protocol was developed for deep learning-based ST inference from morphology. This protocol featured the Visium CytAssist assay to permit automated hematoxylin and eosin staining (e.g. Leica Bond), 40×-resolution imaging, and joining of multiple patients' tissue sections per capture area prior to ST profiling. Using a cohort of 13 pathologic T Stage-III stage colorectal cancer patients, we compared the performance of models trained on slide prepared using enhanced versus traditional (i.e. manual staining and low-resolution imaging) protocols. Leveraging Inceptionv3 neural networks, we predicted gene expression across serial, histologically-matched tissue sections using whole slide images (WSI) from both protocols. The data Shapley was used to quantify and compare marginal performance gains on a patient-by-patient basis attributed to using the enhanced protocol versus the actual costs of spatial profiling. Findings indicate that training and validating on WSI acquired through the enhanced protocol as opposed to the traditional method resulted in improved performance at lower fiscal cost. In the realm of ST, the enhancement of deep learning architectures frequently captures the spotlight; however, the significance of specimen processing and imaging is often understated. This research, informed through a game-theoretic lens, underscores the substantial impact that specimen preparation/imaging can have on spatial transcriptomic inference from morphology. It is essential to integrate such optimized processing protocols to facilitate the identification of prognostic markers at a larger scale.
PMID:39367648 | DOI:10.1093/bib/bbae476
Skin Tone Analysis Through Skin Tone Map Generation With Optical Approach and Deep Learning
Skin Res Technol. 2024 Oct;30(10):e70088. doi: 10.1111/srt.70088.
ABSTRACT
BACKGROUND: Skin tone assessment is critical in both cosmetic and medical fields, yet traditional methods like the individual typology angle (ITA) have limitations, such as sensitivity to illuminants and insensitivity to skin redness.
METHODS: This study introduces an automated image-based method for skin tone mapping by applying optical approaches and deep learning. The method generates skin tone maps by leveraging the illuminant spectrum, segments the skin region from face images, and identifies the corresponding skin tone on the map. The method was evaluated by generating skin tone maps under three standard illuminants (D45, D65, and D85) and comparing the results with those obtained using ITA on skin tone simulation images.
RESULTS: The results showed that skin tone maps generated under the same lighting conditions as the image acquisition (D65) provided the highest accuracy, with a color difference of around 6, which is more than twice as small as those observed under other illuminants. The mapping positions also demonstrated a clear correlation with pigment levels. Compared to ITA, the proposed approach was particularly effective in distinguishing skin tones related to redness.
CONCLUSION: Despite the need to measure the illuminant spectrum and for further physiological validation, the proposed approach shows potential for enhancing skin tone assessment. Its ability to mitigate the effects of illuminants and distinguish between the two dominant pigments offers promising applications in both cosmetic and medical diagnostics.
PMID:39366914 | PMC:PMC11452249 | DOI:10.1111/srt.70088
Coronary artery disease detection using deep learning and ultrahigh-resolution photon-counting coronary CT angiography
Diagn Interv Imaging. 2024 Oct 3:S2211-5684(24)00209-2. doi: 10.1016/j.diii.2024.09.012. Online ahead of print.
ABSTRACT
PURPOSE: The purpose of this study was to evaluate the diagnostic performance of automated deep learning in the detection of coronary artery disease (CAD) on photon-counting coronary CT angiography (PC-CCTA).
MATERIALS AND METHODS: Consecutive patients with suspected CAD who underwent PC-CCTA between January 2022 and December 2023 were included in this retrospective, single-center study. Non-ultra-high resolution (UHR) PC-CCTA images were analyzed by artificial intelligence using two deep learning models (CorEx, Spimed-AI), and compared to human expert reader assessment using UHR PC-CCTA images. Diagnostic performance for global CAD assessment (at least one significant stenosis ≥ 50 %) was estimated at patient and vessel levels.
RESULTS: A total of 140 patients (96 men, 44 women) with a median age of 60 years (first quartile, 51; third quartile, 68) were evaluated. Significant CAD on UHR PC-CCTA was present in 36/140 patients (25.7 %). The sensitivity, specificity, accuracy, positive predictive value), and negative predictive value of deep learning-based CAD were 97.2 %, 81.7 %, 85.7 %, 64.8 %, and 98.9 %, respectively, at the patient level and 96.6 %, 86.7 %, 88.1 %, 53.8 %, and 99.4 %, respectively, at the vessel level. The area under the receiver operating characteristic curve was 0.90 (95 % CI: 0.83-0.94) at the patient level and 0.92 (95 % CI: 0.89-0.94) at the vessel level.
CONCLUSION: Automated deep learning shows remarkable performance for the diagnosis of significant CAD on non-UHR PC-CCTA images. AI pre-reading may be of supportive value to the human reader in daily clinical practice to target and validate coronary artery stenosis using UHR PC-CCTA.
PMID:39366836 | DOI:10.1016/j.diii.2024.09.012
DeepNeoAG: Neoantigen epitope prediction from melanoma antigens using a synergistic deep learning model combining protein language models and multi-window scanning convolutional neural networks
Int J Biol Macromol. 2024 Oct 2:136252. doi: 10.1016/j.ijbiomac.2024.136252. Online ahead of print.
ABSTRACT
Neoantigens, derived from tumor-specific mutations, play a crucial role in eliciting anti-tumor immune responses and have emerged as promising targets for personalized cancer immunotherapy. Accurately identifying neoantigens from a vast pool of potential candidates is crucial for developing effective therapeutic strategies. This study presents a novel deep learning model that leverages the power of protein language models (PLMs) and multi-window scanning convolutional neural networks (CNNs) to predict neoantigens from normal tumor antigens with high accuracy. In this study, we present DeepNeoAG, a novel framework combines the global sequence-level information captured by a pre-trained PLM with the local sequence-based information features extracted by a multi-window scanning CNN, enabling a comprehensive representation of the protein's mutational landscape. We demonstrate the superior performance of DeepNeoAG compared to existing methods and highlight its potential to accelerate the development of personalized cancer immunotherapies.
PMID:39366619 | DOI:10.1016/j.ijbiomac.2024.136252
Spatial heterogeneity of infiltrating immune cells in the tumor microenvironment of non-small cell lung cancer
Transl Oncol. 2024 Oct 3;50:102143. doi: 10.1016/j.tranon.2024.102143. Online ahead of print.
ABSTRACT
Tumor-infiltrating lymphocytes (TILs) are essential components of the tumor microenvironment (TME) of non-small cell lung cancer (NSCLC). Still, it is difficult to describe due to their heterogeneity. In this study, five cell markers from NSCLC patients were analyzed. We segmented tumor cells (TCs) and TILs using Efficientnet-B3 and explored their quantitative information and spatial distribution. After that, we simulated multiplex immunohistochemistry (mIHC) by overlapping continuous single chromogenic IHCs slices. As a result, the proportion and the density of programmed cell death-ligand 1 (PD-L1)-positive TCs were the highest in the core. CD8+ T cells were the closest to the tumor (median distance: 41.71 μm), while PD-1+T cells were the most distant (median distance: 62.2μm), and our study found that most lymphocytes clustered together within the peritumoral range of 10-30 μm where cross-talk with TCs could be achieved. We also found that the classification of TME could be achieved using CD8+ T-cell density, which is correlated with the prognosis of patients. In addition, we achieved single chromogenic IHC slices overlap based on CD4-stained IHC slices. We explored the number and spatial distribution of cells in heterogeneous TME of NSCLC patients and achieved TME classification. We also found a way to show the co-expression of multiple molecules economically.
PMID:39366301 | DOI:10.1016/j.tranon.2024.102143
Image quality improvement in single plane-wave imaging using deep learning
Ultrasonics. 2024 Sep 30;145:107479. doi: 10.1016/j.ultras.2024.107479. Online ahead of print.
ABSTRACT
In ultrasound image diagnosis, single plane-wave imaging (SPWI), which can acquire ultrasound images at more than 1000 fps, has been used to observe detailed tissue and evaluate blood flow. SPWI achieves high temporal resolution by sacrificing the spatial resolution and contrast of ultrasound images. To improve spatial resolution and contrast in SPWI, coherent plane-wave compounding (CPWC) is used to obtain high-quality ultrasound images, i.e., compound images, by coherent addition of radio frequency (RF) signals acquired by transmitting plane waves in different directions. Although CPWC produces high-quality ultrasound images, their temporal resolution is lower than that of SPWI. To address this problem, some methods have been proposed to reconstruct a ultrasound image comparable to a compound image from RF signals obtained by transmitting a small number of plane waves in different directions. These methods do not fully consider the properties of RF signals, resulting in lower image quality compared to a compound image. In this paper, we propose methods to reconstruct high-quality ultrasound images in SPWI by considering the characteristics of RF signal of a single plane wave to obtain ultrasound images with image quality comparable to CPWC. The proposed methods employ encoder-decoder models of 1D U-Net, 2D U-Net, and their combination to generate the high-quality ultrasound images by minimizing the loss that considers the point spread effect of plane waves and frequency spectrum of RF signals in training. We also create a public large-scale SPWI/CPWC dataset for developing and evaluating deep-learning methods. Through a set of experiments using the public dataset and our dataset, we demonstrate that the proposed methods can reconstruct higher-quality ultrasound images from RF signals in SPWI than conventional method.
PMID:39366205 | DOI:10.1016/j.ultras.2024.107479
Shuffled ECA-Net for stress detection from multimodal wearable sensor data
Comput Biol Med. 2024 Oct 3;183:109217. doi: 10.1016/j.compbiomed.2024.109217. Online ahead of print.
ABSTRACT
BACKGROUND: Recently, stress has been recognized as a key factor in the emergence of individual and social issues. Numerous attempts have been made to develop sensor-augmented psychological stress detection techniques, although existing methods are often impractical or overly subjective. To overcome these limitations, we acquired a dataset utilizing both wireless wearable multimodal sensors and salivary cortisol tests for supervised learning. We also developed a novel deep neural network (DNN) model that maximizes the benefits of sensor fusion.
METHOD: We devised a DNN involving a shuffled efficient channel attention (ECA) module called a shuffled ECA-Net, which achieves advanced feature-level sensor fusion by considering inter-modality relationships. Through an experiment involving salivary cortisol tests on 26 participants, we acquired multiple bio-signals including electrocardiograms, respiratory waveforms, and electrogastrograms in both relaxed and stressed mental states. A training dataset was generated from the obtained data. Using the dataset, our proposed model was optimized and evaluated ten times through five-fold cross-validation, while varying a random seed.
RESULTS: Our proposed model achieved acceptable performance in stress detection, showing 0.916 accuracy, 0.917 sensitivity, 0.916 specificity, 0.914 F1-score, and 0.964 area under the receiver operating characteristic curve (AUROC). Furthermore, we demonstrated that combining multiple bio-signals with a shuffled ECA module can more accurately detect psychological stress.
CONCLUSIONS: We believe that our proposed model, coupled with the evidence for the viability of multimodal sensor fusion and a shuffled ECA-Net, would significantly contribute to the resolution of stress-related issues.
PMID:39366142 | DOI:10.1016/j.compbiomed.2024.109217
Two-stage deep learning framework for occlusal crown depth image generation
Comput Biol Med. 2024 Oct 3;183:109220. doi: 10.1016/j.compbiomed.2024.109220. Online ahead of print.
ABSTRACT
The generation of depth images of occlusal dental crowns is complicated by the need for customization in each case. To decrease the workload of skilled dental technicians, various computer vision models have been used to generate realistic occlusal crown depth images with definite crown surface structures that can ultimately be reconstructed to three-dimensional crowns and directly used in patient treatment. However, it has remained difficult to generate images of the structure of dental crowns in a fluid position using computer vision models. In this paper, we propose a two-stage model for generating depth images of occlusal crowns in diverse positions. The model is divided into two parts: segmentation and inpainting to obtain both shape and surface structure accuracy. The segmentation network focuses on the position and size of the crowns, which allows the model to adapt to diverse targets. The inpainting network based on a GAN generates curved structures of the crown surfaces based on the target jaw image and a binary mask made by the segmentation network. The performance of the model is evaluated via quantitative metrics for the area detection and pixel-value metrics. Compared to the baseline model, the proposed method reduced the MSE score from 0.007001 to 0.002618 and increased DICE score from 0.9333 to 0.9648. It indicates that the model showed better performance in terms of the binary mask from the addition of the segmentation network and the internal structure through the use of inpainting networks. Also, the results demonstrated an improved ability of the proposed model to restore realistic details compared to other models.
PMID:39366141 | DOI:10.1016/j.compbiomed.2024.109220
A multicenter dataset for lymph node clinical target volume delineation of nasopharyngeal carcinoma
Sci Data. 2024 Oct 4;11(1):1085. doi: 10.1038/s41597-024-03890-0.
ABSTRACT
The deep learning (DL)-based prediction of accurate lymph node (LN) clinical target volumes (CTVs) for nasopharyngeal carcinoma (NPC) radiotherapy (RT) remains challenging. One of the main reasons is the variability of contours despite standardization processes by expert guidelines in combination with scarce data sharing in the community. Therefore, we retrospectively generated a 262-subjects dataset from four centers to develop the DL models for LN CTVs delineation. This dataset included 440 computed tomography images from different scanning phases, disease stages and treatment strategies. Three clinical expert boards, each comprising two experts (totalling six experts), manually delineated six basic LN CTVs on separate cohorts as the ground truth according to LN involvement and clinical requirements. Several state-of-the-art segmentation algorithms were evaluated on this benchmark, showing promising results for LN CTV segmentation. In conclusion, this work built a multicenter LN CTV segmentation dataset, which may be the first dataset for automatic LN CTV delineation development and evaluation, serving as a benchmark for future research.
PMID:39366975 | PMC:PMC11452638 | DOI:10.1038/s41597-024-03890-0
Behavioral observation and assessment protocol for language and social-emotional development study in children aged 0-6: the Chinese baby connectome project
BMC Psychol. 2024 Oct 4;12(1):533. doi: 10.1186/s40359-024-02031-x.
ABSTRACT
BACKGROUND: The global rise in developmental delays underscores the critical need for a thorough understanding and timely interventions during early childhood. Addressing this issue, the Chinese Baby Connectome Project (CBCP)'s behavior branch is dedicated to examining language acquisition, social-emotional development, and environmental factors affecting Chinese children. The research framework is built around three primary objectives: developing a 0-6 Child Development Assessment Toolkit, implementing an Intelligent Coding System, and investigating environmental influence.
METHODS: Utilizing an accelerated longitudinal design, the CBCP aims to enlist a minimum of 1000 typically developing Chinese children aged 0-6. The data collected in this branch constitutes parental questionnaires, behavioral assessments, and observational experiments to capture their developmental milestones and environmental influences holistically. The parental questionnaires will gauge children's developmental levels in language and social-emotional domains, alongside parental mental well-being, life events, parenting stress, parenting styles, and family relationships. Behavioral assessments will involve neurofunctional developmental evaluations using tools such as the Griffiths Development Scales and Wechsler Preschool and Primary Scale of Intelligence. Additionally, the assessments will encompass measuring children's executive functions (e.g., Head-Toe-Knee-Shoulder), social cognitive abilities (e.g., theory of mind), and language development (e.g., Early Chinese Vocabulary Test). A series of behavior observation. experiments will be conducted targeting children of different age groups, focusing primarily on aspects such as behavioral inhibition, compliance, self-control, and social-emotional regulation. To achieve the objectives, established international questionnaires will be adapted to suit local contexts and devise customized metrics for evaluating children's language and social-emotional development; deep learning algorithms will be developed in the observational experiments to enable automated behavioral analysis; and statistical models will be built to factor in various environmental variables to comprehensively outline developmental trajectories and relationships.
DISCUSSION: This study's integration of diverse assessments and AI technology will offer a detailed analysis of early childhood development in China, particularly in the realms of language acquisition and social-emotional skills. The development of a comprehensive assessment toolkit and coding system will enhance our ability to understand and support the development of Chinese children, contributing significantly to the field of early childhood development research.
TRIAL REGISTRATION: This study was registered with clinicaltrials.gov NCT05040542 on September 10, 2021.
PMID:39367488 | DOI:10.1186/s40359-024-02031-x