Deep learning

Knowledge distillation approach for skin cancer classification on lightweight deep learning model

Thu, 2025-01-16 06:00

Healthc Technol Lett. 2025 Jan 15;12(1):e12120. doi: 10.1049/htl2.12120. eCollection 2025 Jan-Dec.

ABSTRACT

Over the past decade, there has been a global increase in the incidence of skin cancers. Skin cancer has serious consequences if left untreated, potentially leading to more advanced cancer stages. In recent years, deep learning based convolutional neural network have emerged as powerful tools for skin cancer detection. Generally, deep learning approaches are computationally expensive and require large storage space. Therefore, deploying such a large complex model on resource-constrained devices is challenging. An ultra-light and accurate deep learning model is highly desirable for better inference time and memory in low-power-consuming devices. Knowledge distillation is an approach for transferring knowledge from a large network to a small network. This small network is easily compatible with resource-constrained embedded devices while maintaining accuracy. The main aim of this study is to develop a deep learning-based lightweight network based on knowledge distillation that identifies the presence of skin cancer. Here, different training strategies are implemented for the modified benchmark (Phase 1) and custom-made model (Phase 2) and demonstrated various distillation configurations on two datasets: HAM10000 and ISIC2019. In Phase 1, the student model using knowledge distillation achieved accuracies ranging from 88.69% to 93.24% for HAM10000 and from 82.14% to 84.13% on ISIC2019. In Phase 2, the accuracies ranged from 88.63% to 88.89% on HAM10000 and from 81.39% to 83.42% on ISIC2019. These results highlight the effectiveness of knowledge distillation in improving the classification performance across diverse datasets and enabling the student model to approach the performance of the teacher model. In addition, the distilled student model can be easily deployed on resource-constrained devices for automated skin cancer detection due to its lower computational complexity.

PMID:39816700 | PMC:PMC11733311 | DOI:10.1049/htl2.12120

Categories: Literature Watch

Deep cascaded registration and weakly-supervised segmentation of fetal brain MRI

Thu, 2025-01-16 06:00

Heliyon. 2024 Nov 19;11(1):e40148. doi: 10.1016/j.heliyon.2024.e40148. eCollection 2025 Jan 15.

ABSTRACT

Deformable image registration is a cornerstone of many medical image analysis applications, particularly in the context of fetal brain magnetic resonance imaging (MRI), where precise registration is essential for studying the rapidly evolving fetal brain during pregnancy and potentially identifying neurodevelopmental abnormalities. While deep learning has become the leading approach for medical image registration, traditional convolutional neural networks (CNNs) often fall short in capturing fine image details due to their bias toward low spatial frequencies. To address this challenge, we introduce a deep learning registration framework comprising multiple cascaded convolutional networks. These networks predict a series of incremental deformation fields that transform the moving image at various spatial frequency levels, ensuring accurate alignment with the fixed image. This multi-resolution approach allows for a more accurate and detailed registration process, capturing both coarse and fine image structures. Our method outperforms existing state-of-the-art techniques, including other multi-resolution strategies, by a substantial margin. Furthermore, we integrate our registration method into a multi-atlas segmentation pipeline and showcase its competitive performance compared to nnU-Net, achieved using only a small subset of annotated images as atlases. This approach is particularly valuable in the context of fetal brain MRI, where annotated datasets are limited. Our pipeline for registration and multi-atlas segmentation is publicly available at https://github.com/ValBcn/CasReg.

PMID:39816514 | PMC:PMC11732682 | DOI:10.1016/j.heliyon.2024.e40148

Categories: Literature Watch

STDCformer: Spatial-temporal dual-path cross-attention model for fMRI-based autism spectrum disorder identification

Thu, 2025-01-16 06:00

Heliyon. 2024 Jul 10;10(14):e34245. doi: 10.1016/j.heliyon.2024.e34245. eCollection 2024 Jul 30.

ABSTRACT

Resting-state functional magnetic resonance imaging (rs-fMRI) is a non-invasive neuroimaging technique widely utilized in the research of Autism Spectrum Disorder (ASD), providing preliminary insights into the potential biological mechanisms underlying ASD. Deep learning techniques have demonstrated significant potential in the analysis of rs-fMRI. However, accurately distinguishing between healthy control group and ASD has been a longstanding challenge. In this regard, this work proposes a model featuring a dual-path cross-attention framework for spatial and temporal patterns, named STDCformer, aiming to enhance the accuracy of ASD identification. STDCformer can preserve both temporal-specific patterns and spatial-specific patterns while explicitly interacting spatiotemporal information in depth. The embedding layer of the STDCformer embeds temporal and spatial patterns in dual paths. For the temporal path, we introduce a perturbation positional encoding to improve the issue of signal misalignment caused by individual differences. For the spatial path, we propose a correlation metric based on Gramian angular field similarity to establish a more specific whole-brain functional network. Subsequently, we interleave the query and key vectors of dual paths to interact spatial and temporal information. We further propose integrating the dual-path attention into a tensor that retains spatiotemporal dimensions and utilizing 2D convolution for feed-forward processing. Our attention layer allows the model to represent spatiotemporal correlations of signals at multiple scales to alleviate issues of information distortion and loss. Our STDCformer demonstrates competitive results compared to state-of-the-art methods on the ABIDE dataset. Additionally, we conducted interpretative analyses of the model to preliminarily discuss the potential physiological mechanisms of ASD. This work once again demonstrates the potential of deep learning technology in identifying ASD and developing neuroimaging biomarkers for ASD.

PMID:39816341 | PMC:PMC11734066 | DOI:10.1016/j.heliyon.2024.e34245

Categories: Literature Watch

3T dilated inception network for enhanced autism spectrum disorder diagnosis using resting-state fMRI data

Thu, 2025-01-16 06:00

Cogn Neurodyn. 2025 Dec;19(1):22. doi: 10.1007/s11571-024-10202-0. Epub 2025 Jan 13.

ABSTRACT

Autism spectrum disorder (ASD) is one of the complicated neurodevelopmental disorders that impacts the daily functioning and social interactions of individuals. It includes diverse symptoms and severity levels, making it challenging to diagnose and treat efficiently. Various deep learning (DL) based methods have been developed for diagnosing ASD, which rely heavily on behavioral assessment. However, existing techniques have suffered from poor diagnostic outcomes, higher computational complexity, and overfitting issues. To address these challenges, this research work introduces an innovative framework called 3T Dilated Inception Network (3T-DINet) for effective ASD diagnosis using resting-state functional Magnetic Resonance Imaging (rs-fMRI) images. The proposed 3T-DINet technique designs a 3T dilated inception module that incorporates dilated convolutions along with the inception module, allowing it to extract multi-scale features from brain connectivity patterns. The 3T dilated inception module uses three distinct dilation rates (low, medium, and high) in parallel to determine local, mid-level, and global features from the brain. In addition, the proposed approach implements Residual networks (ResNet) to avoid the vanishing gradient problem and enhance the feature extraction ability. The model is further optimized using a Crossover-based Black Widow Optimization (CBWO) algorithm that fine-tunes the hyperparameters thereby enhancing the overall performance of the model. Further, the performance of the 3T-DINet model is evaluated using the five ASD datasets with distinct evaluation parameters. The proposed 3T-DINet technique achieved superior diagnosis results compared to recent previous works. From this simulation validation, it's clear that the 3T-DINet provides an excellent contribution to early ASD diagnosis and enhances patient treatment outcomes.

PMID:39816217 | PMC:PMC11729590 | DOI:10.1007/s11571-024-10202-0

Categories: Literature Watch

MulitDeepsurv: survival analysis of gastric cancer based on deep learning multimodal fusion models

Thu, 2025-01-16 06:00

Biomed Opt Express. 2024 Dec 11;16(1):126-141. doi: 10.1364/BOE.541570. eCollection 2025 Jan 1.

ABSTRACT

Gastric cancer is a leading cause of cancer-related deaths globally. As mortality rates continue to rise, predicting cancer survival using multimodal data-including histopathological images, genomic data, and clinical information-has become increasingly crucial. However, extracting effective predictive features from this complex data has posed challenges for survival analysis due to the high dimensionality and heterogeneity of histopathology images and genomic data. Furthermore, existing methods often lack sufficient interaction between intra- and inter-modal features, significantly impacting model performance. To address these challenges, we developed a deep learning-based multimodal feature fusion model, MultiDeepsurv, designed to predict the survival of gastric cancer patients by integrating histopathological images, clinical data, and gene expression data. Our approach includes a two-branch hybrid network, GLFUnet, which leverages the attention mechanism for enhanced pathology image representation learning. Additionally, we employ a graph convolutional neural network (GCN) to extract features from gene expression data and clinical information. To capture the correlations between different modalities, we utilize the SFusion fusion strategy that employs a self-attention mechanism to learn potential correlations across modalities. Finally, these deeply processed features are fed into Cox regression models for an end-to-end survival analysis. Comprehensive experiments and analyses conducted on a gastric cancer cohort from The Cancer Genome Atlas (TCGA) demonstrate that our proposed MultiDeepsurv model outperforms other methods in terms of prognostic accuracy, with a C-index of 0.806 and an AUC of 0.842.

PMID:39816158 | PMC:PMC11729289 | DOI:10.1364/BOE.541570

Categories: Literature Watch

MXene-based SERS spectroscopic analysis of exosomes for lung cancer differential diagnosis with deep learning

Thu, 2025-01-16 06:00

Biomed Opt Express. 2024 Dec 23;16(1):303-319. doi: 10.1364/BOE.547176. eCollection 2025 Jan 1.

ABSTRACT

Lung cancer with heterogeneity has a high mortality rate due to its late-stage detection and chemotherapy resistance. Liquid biopsy that discriminates tumor-related biomarkers in body fluids has emerged as an attractive technique for early-stage and accurate diagnosis. Exosomes, carrying membrane and cytosolic information from original tumor cells, impart themselves endogeneity and heterogeneity, which offer extensive and unique advantages in the field of liquid biopsy for cancer differential diagnosis. Herein, we demonstrate a Gramian angular summation field and MobileNet V2 (GASF-MobileNet)-assisted surface-enhanced Raman spectroscopy (SERS) technique for analyzing exosomes, aimed at precise diagnosis of lung cancer. Specifically, a composite substrate was synthesized for SERS detection of exosomes based on Ti3C2Tx Mxene and the array of gold-silver core-shell nanocubes (MGS), that combines sensitivity and signal stability. The employment of MXene facilitates the non-selective capture and enrichment of exosomes. To overcome the issue of potentially overlooking spatial features in spectral data analysis, 1-D spectra were first transformed into 2-D images through GASF. By using transformed images as the input data, a deep learning model based on the MobileNet V2 framework extracted spectral features from higher dimensions, which identified different non-small cell lung cancer (NSCLC) cell lines with an overall accuracy of 95.23%. Moreover, the area under the curve (AUC) for each category exceeded 0.95, demonstrating the great potential of integrating label-free SERS with deep learning for precise lung cancer differential diagnosis. This approach allows routine cancer management, and meanwhile, its non-specific analysis of SERS signatures is anticipated to be expanded to other cancers.

PMID:39816152 | PMC:PMC11729284 | DOI:10.1364/BOE.547176

Categories: Literature Watch

Biochemical components of corneal stroma: a study on myopia classification based on Raman spectroscopy and deep learning methods

Thu, 2025-01-16 06:00

Biomed Opt Express. 2024 Dec 3;16(1):28-41. doi: 10.1364/BOE.539721. eCollection 2025 Jan 1.

ABSTRACT

The study aimed to identify differences in the biochemical composition of corneal stroma lenses across varying degrees of myopia using Raman spectrum characteristics. Corneal stroma lens samples from 38 patients who underwent small incision lens extraction (SMILE) surgery, were categorized into low (n = 9, spherical power ≧ -3.00D), moderate (n = 23, spherical power < -3.00D and > -6.00D), and high myopia (n = 6, spherical power ≦-6.00D) groups. A custom-built microscopic confocal Raman system (MCRS) was used to collect Raman spectra, which were processed by smoothing, denoising, and baseline calibrating to refine raw data. Independent sample t-tests were used to analyze spectral feature peaks among sample types. Significant differences (P < 0.001) were found in multiple Raman spectral characteristic peaks (854 cm-1, 937 cm-1, 1002 cm-1, 1243 cm-1, 1448 cm-1, and 2940 cm-1) between low and high myopia samples, particularly at 2940 cm-1. Differences were also found between low and moderate, and moderate and high myopia samples, although fewer than between low and high myopia samples. The three-classification model, particularly with PLS-KNN training, exhibited superior discriminative performance with accuracy rates of 95%. Similarly, the two-classification model for low and high myopia achieved high accuracy with PLS-KNN (94.4%) compared to PCA-KNN (93.3%). PLS dimensionality reduction slightly outperformed PCA, enhancing classification accuracy. In addition, in both reduction methods, the KNN algorithm demonstrated the highest accuracy and performance. The optimal PLS-KNN classification model showed AUC values of 0.99, 0.98, and 1.00 for ROC curves corresponding to low, moderate, and high myopia, respectively. Classification accuracy rates were 89.7% and 96.9%, and 100% for low and high myopia, respectively. For the two-classification model, accuracy reached 94.4% with an AUC of 0.98, indicating strong performance in distinguishing between high and low myopic corneal stroma. We found significant biochemical differences such as collagen, lipids, and nucleic acids in corneal stroma lenses across varying degrees of myopia, suggesting that Raman spectroscopy holds substantial potential in elucidating the pathogenesis of myopia.

PMID:39816144 | PMC:PMC11729285 | DOI:10.1364/BOE.539721

Categories: Literature Watch

Psoriasis severity assessment: Optimizing diagnostic models with deep learning

Thu, 2025-01-16 06:00

Narra J. 2024 Dec;4(3):e1512. doi: 10.52225/narra.v4i3.1512. Epub 2024 Dec 19.

ABSTRACT

Psoriasis is a chronic skin condition with challenges in the accurate assessment of its severity due to subtle differences between severity levels. The aim of this study was to evaluate deep learning models for automated classification of psoriasis severity. A dataset containing 1,546 clinical images was subjected to pre-processing techniques, including cropping and applying noise reduction through median filtering. The dataset was categorized into four severity classes: none, mild, moderate, and severe, based on the Psoriasis Area and Severity Index (PASI). It was split into 1,082 images for training (70%) and 463 images for validation and testing (30%). Five modified deep convolutional neural networks (DCNN) were evaluated, including ResNet50, VGGNet19, MobileNetV3, MnasNet, and EfficientNetB0. The data were validated based on accuracy, precision, sensitivity, specificity, and F1-score, which were weighted to reflect class representation; Pairwise McNemar's test, Cochran's Q test, Cohen's Kappa, and Post-hoc test were performed on the model performance, where overall accuracy and balanced accuracy were determined. Findings revealed that among the five deep learning models, ResNet50 emerged as the optimum model with an accuracy of 92.50% (95%CI: 91.2-93.8%). The precision, sensitivity, specificity, and F1-score of this model were found to be 93.10%, 92.50%, 97.37%, and 92.68%, respectively. In conclusion, ResNet50 has the potential to provide consistent and objective assessments of psoriasis severity, which could aid dermatologists in timely diagnoses and treatment planning. Further clinical validation and model refinement remain required.

PMID:39816098 | PMC:PMC11731931 | DOI:10.52225/narra.v4i3.1512

Categories: Literature Watch

MultiChem: predicting chemical properties using multi-view graph attention network

Wed, 2025-01-15 06:00

BioData Min. 2025 Jan 16;18(1):4. doi: 10.1186/s13040-024-00419-4.

ABSTRACT

BACKGROUND: Understanding the molecular properties of chemical compounds is essential for identifying potential candidates or ensuring safety in drug discovery. However, exploring the vast chemical space is time-consuming and costly, necessitating the development of time-efficient and cost-effective computational methods. Recent advances in deep learning approaches have offered deeper insights into molecular structures. Leveraging this progress, we developed a novel multi-view learning model.

RESULTS: We introduce a graph-integrated model that captures both local and global structural features of chemical compounds. In our model, graph attention layers are employed to effectively capture essential local structures by jointly considering atom and bond features, while multi-head attention layers extract important global features. We evaluated our model on nine MoleculeNet datasets, encompassing both classification and regression tasks, and compared its performance with state-of-the-art methods. Our model achieved an average area under the receiver operating characteristic (AUROC) of 0.822 and a root mean squared error (RMSE) of 1.133, representing a 3% improvement in AUROC and a 7% improvement in RMSE over state-of-the-art models in extensive seed testing.

CONCLUSION: MultiChem highlights the importance of integrating both local and global structural information in predicting molecular properties, while also assessing the stability of the models across multiple datasets using various random seed values.

IMPLEMENTATION: The codes are available at https://github.com/DMnBI/MultiChem .

PMID:39815309 | DOI:10.1186/s13040-024-00419-4

Categories: Literature Watch

Signatures of H3K4me3 modification predict cancer immunotherapy response and identify a new immune checkpoint-SLAMF9

Wed, 2025-01-15 06:00

Respir Res. 2025 Jan 15;26(1):17. doi: 10.1186/s12931-024-03093-6.

ABSTRACT

H3 lysine 4 trimethylation (H3K4me3) modification and related regulators extensively regulate various crucial transcriptional courses in health and disease. However, the regulatory relationship between H3K4me3 modification and anti-tumor immunity has not been fully elucidated. We identified 72 independent prognostic genes of lung adenocarcinoma (LUAD) whose transcriptional expression were closely correlated with known 27 H3K4me3 regulators. We constructed three H3K4me3 modification patterns utilizing the expression profiles of the 72 genes, and patients classified in each pattern exhibited unique tumor immune infiltration characteristics. Using the principal component analysis (PCA) of H3K4me3-related patterns, we constructed a H3K4me3 risk score (H3K4me3-RS) system. The deep learning analysis using 12,159 cancer samples from 26 cancer types and 725 cancer samples from 5 immunotherapy cohorts revealed that H3K4me3-RS was significantly correlated with cancer immune tolerance and sensitivity. Importantly, this risk-score system showed satisfactory predictive performance for the ICB therapy responses of patients suffering from several cancer types, and we identified that SLAMF9 was one of the immunosuppressive phenotype and immunotherapy resistance-determined genes of H3K4me3-RS. The mice melanoma model showed Slamf9 knockdown remarkably restrained cancer progression and enhanced the efficacy of anti-CTLA-4 and anti-PD-L1 therapies by elevating CD8 + T cell infiltration. This study provided a new H3K4me3-associated biomarker system to predict tumor immunotherapy response and suggested the preclinical rationale for investigating the roles of SLAMF9 in cancer immunity regulation and treatment.

PMID:39815269 | DOI:10.1186/s12931-024-03093-6

Categories: Literature Watch

A radiomics and deep learning nomogram developed and validated for predicting no-collapse survival in patients with osteonecrosis after multiple drilling

Wed, 2025-01-15 06:00

BMC Med Inform Decis Mak. 2025 Jan 15;25(1):26. doi: 10.1186/s12911-025-02859-2.

ABSTRACT

PURPOSE: Identifying patients who may benefit from multiple drilling are crucial. Hence, the purpose of the study is to utilize radiomics and deep learning for predicting no-collapse survival in patients with femoral head osteonecrosis.

METHODS: Patients who underwent multiple drilling were enrolled. Radiomics and deep learning features were extracted from pelvic radiographs and selected by LASSO-COX regression, radiomics and DL signature were then built. The clinical variables were selected through univariate and multivariate Cox regression analysis, and the clinical, radiomics, DL and DLRC model were constructed. Model performance was evaluated using the concordance index (C-index), area under the receiver operating characteristic curve (AUC), net reclassification index (NRI), integrated discrimination improvement (IDI), calibration curves, and Decision Curve Analysis (DCA).

RESULTS: A total of 144 patients (212 hips) were included in the study. ARCO classification, bone marrow edema, and combined necrotic angle were identified as independent risk factors for collapse. The DLRC model exhibited superior discrimination ability with higher C-index of 0.78 (95%CI: 0.73-0.84) and AUC values (0.83 and 0.87) than other models. The DLRC model demonstrated superior predictive performance with a higher C-index of 0.78 (95% CI: 0.73-0.84) and area under the curve (AUC) values of 0.83 for 3-year survival and 0.87 for 5-year survival, outperforming other models. The DLRC model also exhibited favorable calibration and clinical utility, with Kaplan-Meier survival curves revealing significant differences in survival rates between high-risk and low-risk cohorts.

CONCLUSION: This study introduces a novel approach that integrates radiomics and deep learning techniques and demonstrates superior predictive performance for no-collapse survival after multiple drilling. It offers enhanced discrimination ability, favorable calibration, and strong clinical utility, making it a valuable tool for stratifying patients into high-risk and low-risk groups. The model has the potential to provide personalized risk assessment, guiding treatment decisions and improving outcomes in patients with osteonecrosis of the femoral head.

PMID:39815247 | DOI:10.1186/s12911-025-02859-2

Categories: Literature Watch

TopoQual polishes circular consensus sequencing data and accurately predicts quality scores

Wed, 2025-01-15 06:00

BMC Bioinformatics. 2025 Jan 16;26(1):17. doi: 10.1186/s12859-024-06020-0.

ABSTRACT

BACKGROUND: Pacific Biosciences (PacBio) circular consensus sequencing (CCS), also known as high fidelity (HiFi) technology, has revolutionized modern genomics by producing long (10 + kb) and highly accurate reads. This is achieved by sequencing circularized DNA molecules multiple times and combining them into a consensus sequence. Currently, the accuracy and quality value estimation provided by HiFi technology are more than sufficient for applications such as genome assembly and germline variant calling. However, there are limitations in the accuracy of the estimated quality scores when it comes to somatic variant calling on single reads.

RESULTS: To address the challenge of inaccurate quality scores for somatic variant calling, we introduce TopoQual, a novel tool designed to enhance the accuracy of base quality predictions. TopoQual leverages techniques including partial order alignments (POA), topologically parallel bases, and deep learning algorithms to polish consensus sequences. Our results demonstrate that TopoQual corrects approximately 31.9% of errors in PacBio consensus sequences. Additionally, it validates base qualities up to q59, which corresponds to one error in 0.9 million bases. These improvements will significantly enhance the reliability of somatic variant calling using HiFi data.

CONCLUSION: TopoQual represents a significant advancement in genomics by improving the accuracy of base quality predictions for PacBio HiFi sequencing data. By correcting a substantial proportion of errors and achieving high base quality validation, TopoQual enables confident and accurate somatic variant calling. This tool not only addresses a critical limitation of current HiFi technology but also opens new possibilities for precise genomic analysis in various research and clinical applications.

PMID:39815230 | DOI:10.1186/s12859-024-06020-0

Categories: Literature Watch

Efficient evidence selection for systematic reviews in traditional Chinese medicine

Wed, 2025-01-15 06:00

BMC Med Res Methodol. 2025 Jan 15;25(1):10. doi: 10.1186/s12874-024-02430-z.

ABSTRACT

PURPOSE: The process of searching for and selecting clinical evidence for systematic reviews (SRs) or clinical guidelines is essential for researchers in Traditional Chinese medicine (TCM). However, this process is often time-consuming and resource-intensive. In this study, we introduce a novel precision-preferred comprehensive information extraction and selection procedure to enhance both the efficiency and accuracy of evidence selection for TCM practitioners.

METHODS: We integrated an established deep learning model (Evi-BERT combined rule-based method) with Boolean logic algorithms and an expanded retrieval strategy to automatically and accurately select potential evidence with minimal human intervention. The selection process is recorded in real-time, allowing researchers to backtrack and verify its accuracy. This innovative approach was tested on ten high-quality, randomly selected systematic reviews of TCM-related topics written in Chinese. To evaluate its effectiveness, we compared the screening time and accuracy of this approach with traditional evidence selection methods.

RESULTS: Our finding demonstrated that the new method accurately selected potential literature based on consistent criteria while significantly reducing the time required for the process. Additionally, in some cases, this approach identified a broader range of relevant evidence and enabled the tracking of selection progress for future reference. The study also revealed that traditional screening methods are often subjective and prone to errors, frequently resulting in the inclusion of literature that does not meet established standards. In contrast, our method offers a more accurate and efficient way to select clinical evidence for TCM practitioners, outperforming traditional manual approaches.

CONCLUSION: We proposed an innovative approach for selecting clinical evidence for TCM reviews and guidelines, aiming to reduce the workload for researchers. While this method showed promise in improving the efficiency and accuracy of evidence-based selection, its full potential required further validation. Additionally, it may serve as a useful tool for editors to assess manuscript quality in the future.

PMID:39815209 | DOI:10.1186/s12874-024-02430-z

Categories: Literature Watch

Deep Radiogenomics Sequencing for Breast Tumor Gene-Phenotype Decoding Using Dynamic Contrast Magnetic Resonance Imaging

Wed, 2025-01-15 06:00

Mol Imaging Biol. 2025 Jan 15. doi: 10.1007/s11307-025-01981-x. Online ahead of print.

ABSTRACT

PURPOSE: We aim to perform radiogenomic profiling of breast cancer tumors using dynamic contrast magnetic resonance imaging (MRI) for the estrogen receptor (ER), progesterone receptor (PR), and human epidermal growth factor receptor 2 (HER2) genes.

METHODS: The dataset used in the current study consists of imaging data of 922 biopsy-confirmed invasive breast cancer patients with ER, PR, and HER2 gene mutation status. Breast MR images, including a T1-weighted pre-contrast sequence and three post-contrast sequences, were enrolled for analysis. All images were corrected using N4 bias correction algorithms. Based on all images and tumor masks, a bounding box of 128 × 128 × 68 was chosen to include all tumor regions. All networks were implemented in 3D fashion with input sizes of 128 × 128 × 68, and four images were input to each network for multi-channel analysis. Data were randomly split into train/validation (80%) and test set (20%) with stratification in class (patient-wise), and all metrics were reported in 20% of the untouched test dataset.

RESULTS: For ER prediction, SEResNet50 achieved an AUC mean of 0.695 (CI95%: 0.610-0.775), a sensitivity of 0.564, and a specificity of 0.787. For PR prediction, ResNet34 achieved an AUC mean of 0.658 (95% CI: 0.573-0.741), a sensitivity of 0.593, and a specificity of 0.734. For HER2 prediction, SEResNext101 achieved an AUC mean of 0.698 (95% CI: 0.560-0.822), a sensitivity of 0.750, and a specificity of 0.625.

CONCLUSION: The current study demonstrated the feasibility of imaging gene-phenotype decoding in breast tumors using MR images and deep learning algorithms with moderate performance.

PMID:39815134 | DOI:10.1007/s11307-025-01981-x

Categories: Literature Watch

Multistage deep learning for classification of Helicobacter pylori infection status using endoscopic images

Wed, 2025-01-15 06:00

J Gastroenterol. 2025 Jan 15. doi: 10.1007/s00535-024-02209-5. Online ahead of print.

ABSTRACT

BACKGROUND: The automated classification of Helicobacter pylori infection status is gaining attention, distinguishing among uninfected (no history of H. pylori infection), current infection, and post-eradication. However, this classification has relatively low performance, primarily due to the intricate nature of the task. This study aims to develop a new multistage deep learning method for automatically classifying H. pylori infection status.

METHODS: The proposed multistage deep learning method was developed using a training set of 538 subjects, then tested on a validation set of 146 subjects. The classification performance of this new method was compared with the findings of four physicians.

RESULTS: The accuracy of our method was 87.7%, 83.6%, and 95.9% for uninfected, post-eradication, and currently infected cases, respectively, whereas that of the physicians was 81.7%, 76.5%, and 90.3%, respectively. When including the patient's H. pylori eradication history information, the classification accuracy of the method was 92.5%, 91.1%, and 98.6% for uninfected, post-eradication, and currently infected cases, respectively, whereas that of the physicians was 85.6%, 85.1%, and 97.4%, respectively.

CONCLUSION: The new multistage deep learning method shows potential for an innovative approach to gastric cancer screening. It can evaluate individual subjects' cancer risk based on endoscopic images and reduce the burden of physicians.

PMID:39815116 | DOI:10.1007/s00535-024-02209-5

Categories: Literature Watch

GestaltGAN: synthetic photorealistic portraits of individuals with rare genetic disorders

Wed, 2025-01-15 06:00

Eur J Hum Genet. 2025 Jan 15. doi: 10.1038/s41431-025-01787-z. Online ahead of print.

ABSTRACT

The facial gestalt (overall facial morphology) is a characteristic clinical feature in many genetic disorders that is often essential for suspecting and establishing a specific diagnosis. Therefore, publishing images of individuals affected by pathogenic variants in disease-associated genes has been an important part of scientific communication. Furthermore, medical imaging data is also crucial for teaching and training deep-learning models such as GestaltMatcher. However, medical data is often sparsely available, and sharing patient images involves risks related to privacy and re-identification. Therefore, we explored whether generative neural networks can be used to synthesize accurate portraits for rare disorders. We modified a StyleGAN architecture and trained it to produce artificial condition-specific portraits for multiple disorders. In addition, we present a technique that generates a sharp and detailed average patient portrait for a given disorder. We trained our GestaltGAN on the 20 most frequent disorders from the GestaltMatcher database. We used REAL-ESRGAN to increase the resolution of portraits from the training data with low-quality and colorized black-and-white images. To augment the model's understanding of human facial features, an unaffected class was introduced to the training data. We tested the validity of our generated portraits with 63 human experts. Our findings demonstrate the model's proficiency in generating photorealistic portraits that capture the characteristic features of a disorder while preserving patient privacy. Overall, the output from our approach holds promise for various applications, including visualizations for publications and educational materials and augmenting training data for deep learning.

PMID:39815041 | DOI:10.1038/s41431-025-01787-z

Categories: Literature Watch

Deep learning reveals diverging effects of altitude on aging

Wed, 2025-01-15 06:00

Geroscience. 2025 Jan 15. doi: 10.1007/s11357-024-01502-8. Online ahead of print.

ABSTRACT

Aging is influenced by a complex interplay of multifarious factors, including an individual's genetics, environment, and lifestyle. Notably, high altitude may impact aging and age-related diseases through exposures such as hypoxia and ultraviolet (UV) radiation. To investigate this, we mined risk exposure data (summary exposure value), disease burden data (disability-adjusted life years (DALYs)), and death rates and life expectancy from the Global Health Data Exchange (GHDx) and National Data Management Center for Health of Ethiopia for each subnational region of Ethiopia, a country with considerable differences in the living altitude. We conducted a cross-sectional clinical trial involving 227 highland and 202 lowland dwellers from the Tigray region in Northern Ethiopia to gain a general insight into the biological aging at high altitudes. Notably, we observed significantly lower risk exposure rates and a reduced disease burden as well as increased life expectancy by lower mortality rates in higher-altitude regions of Ethiopia. When assessing biological aging using facial photographs, we found a faster rate of aging with increasing elevation, likely due to greater UV exposure. Conversely, analysis of nuclear morphologies of peripheral blood mononuclear cells (PBMCs) in blood smears with five different senescence predictors revealed a significant decrease in DNA damage-induced senescence in both monocytes and lymphocytes with increasing elevation. Overall, our findings suggest that disease and DNA damage-induced senescence decreases with altitude in agreement with the idea that oxidative stress may drive aging.

PMID:39815037 | DOI:10.1007/s11357-024-01502-8

Categories: Literature Watch

Targeting protein-ligand neosurfaces with a generalizable deep learning tool

Wed, 2025-01-15 06:00

Nature. 2025 Jan 15. doi: 10.1038/s41586-024-08435-4. Online ahead of print.

ABSTRACT

Molecular recognition events between proteins drive biological processes in living systems1. However, higher levels of mechanistic regulation have emerged, in which protein-protein interactions are conditioned to small molecules2-5. Despite recent advances, computational tools for the design of new chemically induced protein interactions have remained a challenging task for the field6,7. Here we present a computational strategy for the design of proteins that target neosurfaces, that is, surfaces arising from protein-ligand complexes. To develop this strategy, we leveraged a geometric deep learning approach based on learned molecular surface representations8,9 and experimentally validated binders against three drug-bound protein complexes: Bcl2-venetoclax, DB3-progesterone and PDF1-actinonin. All binders demonstrated high affinities and accurate specificities, as assessed by mutational and structural characterization. Remarkably, surface fingerprints previously trained only on proteins could be applied to neosurfaces induced by interactions with small molecules, providing a powerful demonstration of generalizability that is uncommon in other deep learning approaches. We anticipate that such designed chemically induced protein interactions will have the potential to expand the sensing repertoire and the assembly of new synthetic pathways in engineered cells for innovative drug-controlled cell-based therapies10.

PMID:39814890 | DOI:10.1038/s41586-024-08435-4

Categories: Literature Watch

A novel multi-user collaborative cognitive radio spectrum sensing model: Based on a CNN-LSTM model

Wed, 2025-01-15 06:00

PLoS One. 2025 Jan 15;20(1):e0316291. doi: 10.1371/journal.pone.0316291. eCollection 2025.

ABSTRACT

Cognitive Radio (CR) technology enables wireless devices to learn about their surrounding spectrum environment through sensing capabilities, thereby facilitating efficient spectrum utilization without interfering with the normal operation of licensed users. This study aims to enhance spectrum sensing in multi-user cooperative cognitive radio systems by leveraging a hybrid model that combines Convolutional Neural Networks (CNN) and Long Short-Term Memory (LSTM) networks. A novel multi-user cooperative spectrum sensing model is developed, utilizing CNN's local feature extraction capability and LSTM's advantage in handling sequential data to optimize sensing accuracy and efficiency. Furthermore, a multi-head self-attention mechanism is incorporated to improve information flow, enhancing the model's adaptability and robustness in dynamic and complex environments. Simulation experiments were conducted to quantitatively evaluate the performance of the proposed model. The results demonstrate that the CNN-LSTM model achieves low sensing error rates across various numbers of secondary users (16, 24, 32, 40, 48), with a particularly low sensing error of 9.9658% under the 32-user configuration. Additionally, when comparing the sensing errors of different deep learning models, the proposed model consistently outperformed others, showing a 12% lower sensing error under low-power conditions (100 mW). This study successfully develops a CNN-LSTM-based cooperative spectrum sensing model for multi-user cognitive radio systems, significantly improving sensing accuracy and efficiency. By integrating CNN and LSTM technologies, the model not only enhances sensing performance but also improves the handling of long-term dependencies in time-series data, offering a novel technical approach and theoretical support for cognitive radio research. Moreover, the introduction of the multi-head self-attention mechanism further optimizes the model's adaptability to complex environments, demonstrating significant potential for practical applications.

PMID:39813223 | DOI:10.1371/journal.pone.0316291

Categories: Literature Watch

An ensemble deep learning framework for energy demand forecasting using genetic algorithm-based feature selection

Wed, 2025-01-15 06:00

PLoS One. 2025 Jan 15;20(1):e0310465. doi: 10.1371/journal.pone.0310465. eCollection 2025.

ABSTRACT

Accurate energy demand forecasting is critical for efficient energy management and planning. Recent advancements in computing power and the availability of large datasets have fueled the development of machine learning models. However, selecting the most appropriate features to enhance prediction accuracy and robustness remains a key challenge. This study proposes an ensemble approach that integrates a genetic algorithm with multiple forecasting models to optimize feature selection. The genetic algorithm identifies the optimal subset of features from a dataset that includes historical energy consumption, weather variables, and temporal characteristics. These selected features are then used to train three base learners: Long Short-Term Memory (LSTM), Bi-directional Long Short-Term Memory (BiLSTM), and Gated Recurrent Unit (GRU). The predictions from these models are combined using a stacking ensemble technique to generate the final forecast. To enhance model evaluation, we divided the dataset into weekday and weekend subsets, allowing for a more detailed analysis of energy consumption patterns. To ensure the reliability of our findings, we conducted ten simulations and applied the Wilcoxon Signed Rank Test to the results. The proposed model demonstrated exceptional precision, achieving a Root Mean Square Error (RMSE) of 130.6, a Mean Absolute Percentage Error (MAPE) of 0.38%, and a Mean Absolute Error (MAE) of 99.41 for weekday data. The model also maintained high accuracy for weekend predictions, with an RMSE of 137.41, a MAPE of 0.42%, and an MAE of 105.67. This research provides valuable insights for energy analysts and contributes to developing more sophisticated demand forecasting methods.

PMID:39813218 | DOI:10.1371/journal.pone.0310465

Categories: Literature Watch

Pages