Deep learning

Deep learning in integrating spatial transcriptomics with other modalities

Sun, 2025-01-12 06:00

Brief Bioinform. 2024 Nov 22;26(1):bbae719. doi: 10.1093/bib/bbae719.

ABSTRACT

Spatial transcriptomics technologies have been extensively applied in biological research, enabling the study of transcriptome while preserving the spatial context of tissues. Paired with spatial transcriptomics data, platforms often provide histology and (or) chromatin images, which capture cellular morphology and chromatin organization. Additionally, single-cell RNA sequencing (scRNA-seq) data from matching tissues often accompany spatial data, offering a transcriptome-wide gene expression profile of individual cells. Integrating such additional data from other modalities can effectively enhance spatial transcriptomics data, and, conversely, spatial transcriptomics data can supplement scRNA-seq with spatial information. Moreover, the rapid development of spatial multi-omics technology has spurred the demand for the integration of spatial multi-omics data to present a more detailed molecular landscape within tissues. Numerous deep learning (DL) methods have been developed for integrating spatial transcriptomics with other modalities. However, a comprehensive review of DL approaches for integrating spatial transcriptomics data with other modalities remains absent. In this study, we systematically review the applications of DL in integrating spatial transcriptomics data with other modalities. We first delineate the DL techniques applied in this integration and the key tasks involved. Next, we detail these methods and categorize them based on integrated modality and key task. Furthermore, we summarize the integration strategies of these integration methods. Finally, we discuss the challenges and future directions in integrating spatial transcriptomics with other modalities, aiming to facilitate the development of robust computational methods that more comprehensively exploit multimodal information.

PMID:39800876 | DOI:10.1093/bib/bbae719

Categories: Literature Watch

DD-PRiSM: a deep learning framework for decomposition and prediction of synergistic drug combinations

Sun, 2025-01-12 06:00

Brief Bioinform. 2024 Nov 22;26(1):bbae717. doi: 10.1093/bib/bbae717.

ABSTRACT

Combination therapies have emerged as a promising approach for treating complex diseases, particularly cancer. However, predicting the efficacy and safety profiles of these therapies remains a significant challenge, primarily because of the complex interactions among drugs and their wide-ranging effects. To address this issue, we introduce DD-PRiSM (Decomposition of Drug-Pair Response into Synergy and Monotherapy effect), a deep-learning pipeline that predicts the effects of combination therapy. DD-PRiSM consists of two predictive models. The first is the Monotherapy model, which predicts parameters of the drug response curve based on drug structure and cell line gene expression. This reconstructed curve is then used to predict cell viability at the given drug dosage. The second is the Combination therapy model, which predicts the efficacy of drug combinations by analyzing individual drug effects and their synergistic interactions with a specific dosage level of individual drugs. The efficacy of DD-PRiSM is demonstrated through its performance metrics, achieving a root mean square error of 0.0854, a Pearson correlation coefficient of 0.9063, and an R2 of 0.8209 for unseen pairs. Furthermore, DD-PRiSM distinguishes itself by its capability to decompose combination therapy efficacy, successfully identifying synergistic drug pairs. We demonstrated synergistic responses vary across cancer types and identified hub drugs that trigger synergistic effects. Finally, we suggested a promising drug pair through our case study.

PMID:39800875 | DOI:10.1093/bib/bbae717

Categories: Literature Watch

Integrating scRNA-seq and scATAC-seq with inter-type attention heterogeneous graph neural networks

Sun, 2025-01-12 06:00

Brief Bioinform. 2024 Nov 22;26(1):bbae711. doi: 10.1093/bib/bbae711.

ABSTRACT

Single-cell multi-omics techniques, which enable the simultaneous measurement of multiple modalities such as RNA gene expression and Assay for Transposase-Accessible Chromatin (ATAC) within individual cells, have become a powerful tool for deciphering the intricate complexity of cellular systems. Most current methods rely on motif databases to establish cross-modality relationships between genes from RNA-seq data and peaks from ATAC-seq data. However, these approaches are constrained by incomplete database coverage, particularly for novel or poorly characterized relationships. To address these limitations, we introduce single-cell Multi-omics Integration (scMI), a heterogeneous graph embedding method that encodes both cells and modality features from single-cell RNA-seq and ATAC-seq data into a shared latent space by learning cross-modality relationships. By modeling cells and modality features as distinct node types, we design an inter-type attention mechanism to effectively capture long-range cross-modality interactions between genes and peaks. Benchmark results demonstrate that embeddings learned by scMI preserve more biological information and achieve comparable or superior performance in downstream tasks including modality prediction, cell clustering, and gene regulatory network inference compared to methods that rely on databases. Furthermore, scMI significantly improves the alignment and integration of unmatched multi-omics data, enabling more accurate embedding and improved outcomes in downstream tasks.

PMID:39800872 | DOI:10.1093/bib/bbae711

Categories: Literature Watch

An examination of daily CO(2) emissions prediction through a comparative analysis of machine learning, deep learning, and statistical models

Sun, 2025-01-12 06:00

Environ Sci Pollut Res Int. 2025 Jan 13. doi: 10.1007/s11356-024-35764-8. Online ahead of print.

ABSTRACT

Human-induced global warming, primarily attributed to the rise in atmospheric CO2, poses a substantial risk to the survival of humanity. While most research focuses on predicting annual CO2 emissions, which are crucial for setting long-term emission mitigation targets, the precise prediction of daily CO2 emissions is equally vital for setting short-term targets. This study examines the performance of 14 models in predicting daily CO2 emissions data from 1/1/2022 to 30/9/2023 across the top four polluting regions (China, India, the USA, and the EU27&UK). The 14 models used in the study include four statistical models (ARMA, ARIMA, SARMA, and SARIMA), three machine learning models (support vector machine (SVM), random forest (RF), and gradient boosting (GB)), and seven deep learning models (artificial neural network (ANN), recurrent neural network variations such as gated recurrent unit (GRU), long short-term memory (LSTM), bidirectional-LSTM (BILSTM), and three hybrid combinations of CNN-RNN). Performance evaluation employs four metrics (R2, MAE, RMSE, and MAPE). The results show that the machine learning (ML) and deep learning (DL) models, with higher R2 (0.714-0.932) and lower RMSE (0.480-0.247) values, respectively, outperformed the statistical model, which had R2 (- 0.060-0.719) and RMSE (1.695-0.537) values, in predicting daily CO2 emissions across all four regions. The performance of the ML and DL models was further enhanced by differencing, a technique that improves accuracy by ensuring stationarity and creating additional features and patterns from which the model can learn. Additionally, applying ensemble techniques such as bagging and voting improved the performance of the ML models by approximately 9.6%, whereas hybrid combinations of CNN-RNN enhanced the performance of the RNN models. In summary, the performance of both the ML and DL models was relatively similar. However, due to the high computational requirements associated with DL models, the recommended models for daily CO2 emission prediction are ML models using the ensemble technique of voting and bagging. This model can assist in accurately forecasting daily emissions, aiding authorities in setting targets for CO2 emission reduction.

PMID:39800837 | DOI:10.1007/s11356-024-35764-8

Categories: Literature Watch

Annotation-free deep learning algorithm trained on hematoxylin & eosin images predicts epithelial-to-mesenchymal transition phenotype and endocrine response in estrogen receptor-positive breast cancer

Sun, 2025-01-12 06:00

Breast Cancer Res. 2025 Jan 12;27(1):6. doi: 10.1186/s13058-025-01959-1.

ABSTRACT

Recent evidence indicates that endocrine resistance in estrogen receptor-positive (ER+) breast cancer is closely correlated with phenotypic characteristics of epithelial-to-mesenchymal transition (EMT). Nonetheless, identifying tumor tissues with a mesenchymal phenotype remains challenging in clinical practice. In this study, we validated the correlation between EMT status and resistance to endocrine therapy in ER+ breast cancer from a transcriptomic perspective. To confirm the presence of morphological discrepancies in tumor tissues of ER+ breast cancer classified as epithelial- and mesenchymal-phenotypes according to EMT-related transcriptional features, we trained deep learning algorithms based on EfficientNetV2 architecture to assign the phenotypic status for each patient utilizing hematoxylin & eosin (H&E)-stained slides from The Cancer Genome Atlas database. Our classifier model accurately identified the precise phenotypic status, achieving an area under the curve (AUC) of 0.886 at the tile-level and an AUC of 0.910 at the slide-level. Furthermore, we evaluated the efficacy of the classifier in predicting endocrine response using data from an independent ER+ breast cancer patient cohort. Our classifier achieved a predicting accuracy of 81.25%, and 88.7% slides labeled as endocrine resistant were predicted as the mesenchymal-phenotype, while 75.6% slides labeled as sensitive were predicted as the epithelial-phenotype. Our work introduces an H&E-based framework capable of accurately predicting EMT phenotype and endocrine response for ER+ breast cancer, demonstrating its potential for clinical application and benefit.

PMID:39800743 | DOI:10.1186/s13058-025-01959-1

Categories: Literature Watch

Artificial Intelligence for Cervical Spine Fracture Detection: A Systematic Review of Diagnostic Performance and Clinical Potential

Sun, 2025-01-12 06:00

Global Spine J. 2025 Jan 12:21925682251314379. doi: 10.1177/21925682251314379. Online ahead of print.

ABSTRACT

STUDY DESIGN: Systematic review.

OBJECTIVE: Artificial intelligence (AI) and deep learning (DL) models have recently emerged as tools to improve fracture detection, mainly through imaging modalities such as computed tomography (CT) and radiographs. This systematic review evaluates the diagnostic performance of AI and DL models in detecting cervical spine fractures and assesses their potential role in clinical practice.

METHODS: A systematic search of PubMed/Medline, Embase, Scopus, and Web of Science was conducted for studies published between January 2000 and July 2024. Studies that evaluated AI models for cervical spine fracture detection were included. Diagnostic performance metrics were extracted and included sensitivity, specificity, accuracy, and area under the curve. The PROBAST tool assessed bias, and PRISMA criteria were used for study selection and reporting.

RESULTS: Eleven studies published between 2021 and 2024 were included in the review. AI models demonstrated variable performance, with sensitivity ranging from 54.9% to 100% and specificity from 72% to 98.6%. Models applied to CT imaging generally outperformed those applied to radiographs, with convolutional neural networks (CNN) and advanced architectures such as MobileNetV2 and Vision Transformer (ViT) achieving the highest accuracy. However, most studies lacked external validation, raising concerns about the generalizability of their findings.

CONCLUSIONS: AI and DL models show significant potential in improving fracture detection, particularly in CT imaging. While these models offer high diagnostic accuracy, further validation and refinement are necessary before they can be widely integrated into clinical practice. AI should complement, rather than replace, human expertise in diagnostic workflows.

PMID:39800538 | DOI:10.1177/21925682251314379

Categories: Literature Watch

End-to-end deep-learning model for the detection of coronary artery stenosis on coronary CT images

Sun, 2025-01-12 06:00

Open Heart. 2025 Jan 11;12(1):e002998. doi: 10.1136/openhrt-2024-002998.

ABSTRACT

PURPOSE: We examined whether end-to-end deep-learning models could detect moderate (≥50%) or severe (≥70%) stenosis in the left anterior descending artery (LAD), right coronary artery (RCA) or left circumflex artery (LCX) in iodine contrast-enhanced ECG-gated coronary CT angiography (CCTA) scans.

METHODS: From a database of 6293 CCTA scans, we used pre-existing curved multiplanar reformations (CMR) images of the LAD, RCA and LCX arteries to create end-to-end deep-learning models for the detection of moderate or severe stenoses. We preprocessed the images by exploiting domain knowledge and employed a transfer learning approach using EfficientNet, ResNet, DenseNet and Inception-ResNet, with a class-weighted strategy optimised through cross-validation. Heatmaps were generated to indicate critical areas identified by the models, aiding clinicians in understanding the model's decision-making process.

RESULTS: Among the 900 CMR cases, 279 involved the LAD artery, 259 the RCA artery and 253 the LCX artery. EfficientNet models outperformed others, with EfficientNetB3 and EfficientNetB0 demonstrating the highest accuracy for LAD, EfficientNetB2 for RCA and EfficientNetB0 for LCX. The area under the curve for receiver operating characteristic (AUROC) reached 0.95 for moderate and 0.94 for severe stenosis in the LAD. For the RCA, the AUROC was 0.92 for both moderate and severe stenosis detection. The LCX achieved an AUROC of 0.88 for the detection of moderate stenoses, though the calibration curve exhibited significant overestimation. Calibration curves matched probabilities for the LAD but showed discrepancies for the RCA. Heatmap visualisations confirmed the models' precision in delineating stenotic lesions. Decision curve analysis and net reclassification index assessments reinforced the efficacy of EfficientNet models, confirming their superior diagnostic capabilities.

CONCLUSION: Our end-to-end deep-learning model demonstrates, for the LAD artery, excellent discriminatory ability and calibration during internal validation, despite a small dataset used to train the network. The model reliably produces precise, highly interpretable images.

PMID:39800435 | DOI:10.1136/openhrt-2024-002998

Categories: Literature Watch

Clinical Application Of Deep Learning-assisted Needles Reconstruction In Prostate Ultrasound Brachytherapy

Sun, 2025-01-12 06:00

Int J Radiat Oncol Biol Phys. 2025 Jan 10:S0360-3016(25)00002-1. doi: 10.1016/j.ijrobp.2024.12.026. Online ahead of print.

ABSTRACT

PURPOSE: High dose rate (HDR) prostate brachytherapy (BT) procedure requires image-guided needle insertion. Given that general anesthesia is often employed during the procedure, minimizing overall planning time is crucial. In this study, we explore the clinical feasibility and time-saving potential of artificial intelligence (AI)-driven auto-reconstruction of transperineal needles in the context of US-guided prostate BT planning.

MATERIALS AND METHODS: This study included a total of 102 US-planned BT images from a single institution and split into three groups: 50 for model training and validation, 11 to evaluate reconstruction accuracy (test set), and 41 to evaluate the AI tool in a clinical implementation (clinical set). Reconstruction accuracy for the test set was evaluated by comparing the performance of AI-derived and manually reconstructed needles from 5 medical physicists on the 3D-US scans after treatment. The needle total reconstruction time for the clinical set was defined as the timestamp difference from scan acquisition to the start of dose calculations and was compared to values recorded before the clinical implementation of the AI-assisted tool.

RESULTS: A mean error of (0.44±0.32)mm was found between the AI-reconstructed and the human consensus needle positions in the test set, with 95.0% of AI needle points falling below 1mm from their human-made counterparts. Post-hoc analysis showed only one of the human observers' reconstructions were significantly different from the others including the AI's. In the clinical set, the AI algorithm achieved a true positive reconstruction rate of 93.4% with only 4.5% of these needles requiring manual corrections from the planner before dosimetry. Total time required to perform AI-assisted catheter reconstruction on clinical cases was on average 15.2min lower (p < 0.01) compared to procedure without AI assistance.

CONCLUSIONS: This study demonstrates the feasibility of an AI-assisted needle reconstructing tool for 3D-US based HDR prostate BT. This is a step toward treatment planning automation and increased efficiency in HDR prostate BT.

PMID:39800329 | DOI:10.1016/j.ijrobp.2024.12.026

Categories: Literature Watch

Development and routine implementation of deep learning algorithm for automatic brain metastases segmentation on MRI for RANO-BM criteria follow-up

Sun, 2025-01-12 06:00

Neuroimage. 2025 Jan 10:121002. doi: 10.1016/j.neuroimage.2025.121002. Online ahead of print.

ABSTRACT

RATIONALE AND OBJECTIVES: The RANO-BM criteria, which employ a one-dimensional measurement of the largest diameter, are imperfect due to the fact that the lesion volume is neither isotropic nor homogeneous. Furthermore, this approach is inherently time-consuming. Consequently, in clinical practice, monitoring patients in clinical trials in compliance with the RANO-BM criteria is rarely achieved. The objective of this study was to develop and validate an AI solution capable of delineating brain metastases (BM) on MRI to easily obtain, using an in-house solution, RANO-BM criteria as well as BM volume in a routine clinical setting.

MATERIALS (PATIENTS) AND METHODS: A total of 27456 post-Gadolinium-T1 MRI from 132 patients with BM were employed in this study. A deep learning (DL) model was constructed using the PyTorch and PyTorch Lightning frameworks, and the UNETR transfer learning method was employed to segment BM from MRI.

RESULTS: A visual analysis of the AI model results demonstrates confident delineation of the BM lesions. The model shows 100% accuracy in predicting RANO-BM criteria in comparison to that of an expert medical doctor. There was a high degree of overlap between the AI and the doctor's segmentation, with a mean DICE score of 0.77. The diameter and volume of the BM lesions were found to be concordant between the AI and the reference segmentation. The user interface developed in this study can readily provide RANO-BM criteria following AI BM segmentation.

CONCLUSION: The in-house deep learning solution is accessible to everyone without expertise in AI and offers effective BM segmentation and substantial time savings.

PMID:39800174 | DOI:10.1016/j.neuroimage.2025.121002

Categories: Literature Watch

Enhanced detection of atrial fibrillation in single-lead electrocardiograms using a cloud-based artificial intelligence platform

Sun, 2025-01-12 06:00

Heart Rhythm. 2025 Jan 10:S1547-5271(25)00019-0. doi: 10.1016/j.hrthm.2024.12.048. Online ahead of print.

ABSTRACT

BACKGROUND: Although smartphone-based devices have been developed to record 1-lead ECG, existing solutions for automatic atrial fibrillation (AF) detection often has poor positive predictive value.

OBJECTIVE: This study aimed to validate a cloud-based deep learning platform for automatic AF detection in a large cohort of patients using 1-lead ECG records.

METHODS: We analyzed 8,528 patients with 30-second ECG records from a single-lead handheld ECG device. Ground truth for AF presence was established through a benchmark algorithm and expert manual labeling. The Willem Artificial Intelligence (AI) platform, not trained on these ECGs, was used for automatic arrhythmia detection, including AF. A rules-based algorithm was also used for comparison. An expert cardiology committee reviewed false positives and negatives and performance metrics were computed.

RESULTS: The AI platform achieved an accuracy of 96.1% (initial labels) and 96.4% (expert review), with sensitivities of 83.3% and 84.2%, and specificities of 97.3% and 97.6%, respectively. The positive predictive value was 75.2% and 78.0%, and the negative predictive value was 98.4%. Performance of the AI platform largely exceeded the performance of the rules-based algorithm for all metrics. The AI also detected other arrhythmias, such as premature ventricular complexes, premature atrial complexes along with 1-degree atrioventricular blocks.

CONCLUSIONS: The result of this external validation indicates that the AI platform can match cardiologist-level accuracy in AF detection from 1-lead ECGs. Such tools are promising for AF screening and has the potential to improve accuracy in non-cardiology expert healthcare professional interpretation and trigger further tests for effective patient management.

PMID:39800092 | DOI:10.1016/j.hrthm.2024.12.048

Categories: Literature Watch

A real-time approach for surgical activity recognition and prediction based on transformer models in robot-assisted surgery

Sun, 2025-01-12 06:00

Int J Comput Assist Radiol Surg. 2025 Jan 12. doi: 10.1007/s11548-024-03306-9. Online ahead of print.

ABSTRACT

PURPOSE: This paper presents a deep learning approach to recognize and predict surgical activity in robot-assisted minimally invasive surgery (RAMIS). Our primary objective is to deploy the developed model for implementing a real-time surgical risk monitoring system within the realm of RAMIS.

METHODS: We propose a modified Transformer model with the architecture comprising no positional encoding, 5 fully connected layers, 1 encoder, and 3 decoders. This model is specifically designed to address 3 primary tasks in surgical robotics: gesture recognition, prediction, and end-effector trajectory prediction. Notably, it operates solely on kinematic data obtained from the joints of robotic arm.

RESULTS: The model's performance was evaluated on JHU-ISI Gesture and Skill Assessment Working Set dataset, achieving highest accuracy of 94.4% for gesture recognition, 84.82% for gesture prediction, and significantly low distance error of 1.34 mm with a prediction of 1 s in advance. Notably, the computational time per iteration was minimal recorded at only 4.2 ms.

CONCLUSION: The results demonstrated the excellence of our proposed model compared to previous studies highlighting its potential for integration in real-time systems. We firmly believe that our model could significantly elevate realms of surgical activity recognition and prediction within RAS and make a substantial and meaningful contribution to the healthcare sector.

PMID:39799528 | DOI:10.1007/s11548-024-03306-9

Categories: Literature Watch

DDGemb: predicting protein stability change upon single- and multi-point variations with embeddings and deep learning

Sun, 2025-01-12 06:00

Bioinformatics. 2025 Jan 12:btaf019. doi: 10.1093/bioinformatics/btaf019. Online ahead of print.

ABSTRACT

MOTIVATION: The knowledge of protein stability upon residue variation is an important step for functional protein design and for understanding how protein variants can promote disease onset. Computational methods are important to complement experimental approaches and allow a fast screening of large datasets of variations.

RESULTS: In this work we present DDGemb, a novel method combining protein language model embeddings and transformer architectures to predict protein ΔΔG upon both single- and multi-point variations. DDGemb has been trained on a high-quality dataset derived from literature and tested on available benchmark datasets of single- and multi-point variations. DDGemb performs at the state of the art in both single- and multi-point variations.

AVAILABILITY: DDGemb is available as web server at https://ddgemb.biocomp.unibo.it. Datasets used in this study are available at https://ddgemb.biocomp.unibo.it/datasets.

PMID:39799516 | DOI:10.1093/bioinformatics/btaf019

Categories: Literature Watch

Development of a model for measuring sagittal plane parameters in 10-18-year old adolescents with idiopathic scoliosis based on RTMpose deep learning technology

Sat, 2025-01-11 06:00

J Orthop Surg Res. 2025 Jan 11;20(1):41. doi: 10.1186/s13018-024-05334-2.

ABSTRACT

PURPOSE: The study aimed to develop a deep learning model for rapid, automated measurement of full-spine X-rays in adolescents with Adolescent Idiopathic Scoliosis (AIS). A significant challenge in this field is the time-consuming nature of manual measurements and the inter-individual variability in these measurements. To address these challenges, we utilized RTMpose deep learning technology to automate the process.

METHODS: We conducted a retrospective multicenter diagnostic study using 560 full-spine sagittal plane X-ray images from five hospitals in Inner Mongolia. The model was trained and validated using 500 images, with an additional 60 images for independent external validation. We evaluated the consistency of keypoint annotations among different physicians, the accuracy of model-predicted keypoints, and the accuracy of model measurement results compared to manual measurements.

RESULTS: The consistency percentages of keypoint annotations among different physicians and the model were 90-97% within the 4-mm range. The model's prediction accuracies for key points were 91-100% within the 4-mm range compared to the reference standards. The model's predictions for 15 anatomical parameters showed high consistency with experienced physicians, with intraclass correlation coefficients ranging from 0.892 to 0.991. The mean absolute error for SVA was 1.16 mm, and for other parameters, it ranged from 0.22° to 3.32°. A significant challenge we faced was the variability in data formats and specifications across different hospitals, which we addressed through data augmentation techniques. The model took an average of 9.27 s to automatically measure the 15 anatomical parameters per X-ray image.

CONCLUSION: The deep learning model based on RTMpose can effectively enhance clinical efficiency by automatically measuring the sagittal plane parameters of the spine in X-rays of patients with AIS. The model's performance was found to be highly consistent with manual measurements by experienced physicians, offering a valuable tool for clinical diagnostics.

PMID:39799363 | DOI:10.1186/s13018-024-05334-2

Categories: Literature Watch

UniAMP: enhancing AMP prediction using deep neural networks with inferred information of peptides

Sat, 2025-01-11 06:00

BMC Bioinformatics. 2025 Jan 11;26(1):10. doi: 10.1186/s12859-025-06033-3.

ABSTRACT

Antimicrobial peptides (AMPs) have been widely recognized as a promising solution to combat antimicrobial resistance of microorganisms due to the increasing abuse of antibiotics in medicine and agriculture around the globe. In this study, we propose UniAMP, a systematic prediction framework for discovering AMPs. We observe that feature vectors used in various existing studies constructed from peptide information, such as sequence, composition, and structure, can be augmented and even replaced by information inferred by deep learning models. Specifically, we use a feature vector with 2924 values inferred by two deep learning models, UniRep and ProtT5, to demonstrate that such inferred information of peptides suffice for the task, with the help of our proposed deep neural network model composed of fully connected layers and transformer encoders for predicting the antibacterial activity of peptides. Evaluation results demonstrate superior performance of our proposed model on both balanced benchmark datasets and imbalanced test datasets compared with existing studies. Subsequently, we analyze the relations among peptide sequences, manually extracted features, and automatically inferred information by deep learning models, leading to observations that the inferred information is more comprehensive and non-redundant for the task of predicting AMPs. Moreover, this approach alleviates the impact of the scarcity of positive data and demonstrates great potential in future research and applications.

PMID:39799358 | DOI:10.1186/s12859-025-06033-3

Categories: Literature Watch

Improving 3D deep learning segmentation with biophysically motivated cell synthesis

Sat, 2025-01-11 06:00

Commun Biol. 2025 Jan 11;8(1):43. doi: 10.1038/s42003-025-07469-2.

ABSTRACT

Biomedical research increasingly relies on three-dimensional (3D) cell culture models and artificial-intelligence-based analysis can potentially facilitate a detailed and accurate feature extraction on a single-cell level. However, this requires for a precise segmentation of 3D cell datasets, which in turn demands high-quality ground truth for training. Manual annotation, the gold standard for ground truth data, is too time-consuming and thus not feasible for the generation of large 3D training datasets. To address this, we present a framework for generating 3D training data, which integrates biophysical modeling for realistic cell shape and alignment. Our approach allows the in silico generation of coherent membrane and nuclei signals, that enable the training of segmentation models utilizing both channels for improved performance. Furthermore, we present a generative adversarial network (GAN) training scheme that generates not only image data but also matching labels. Quantitative evaluation shows superior performance of biophysical motivated synthetic training data, even outperforming manual annotation and pretrained models. This underscores the potential of incorporating biophysical modeling for enhancing synthetic training data quality.

PMID:39799275 | DOI:10.1038/s42003-025-07469-2

Categories: Literature Watch

Deep learning for predicting prognostic consensus molecular subtypes in cervical cancer from histology images

Sat, 2025-01-11 06:00

NPJ Precis Oncol. 2025 Jan 11;9(1):11. doi: 10.1038/s41698-024-00778-5.

ABSTRACT

Cervical cancer remains the fourth most common cancer among women worldwide. This study proposes an end-to-end deep learning framework to predict consensus molecular subtypes (CMS) in HPV-positive cervical squamous cell carcinoma (CSCC) from H&E-stained histology slides. Analysing three CSCC cohorts (n = 545), we show our Digital-CMS scores significantly stratify patients by both disease-specific (TCGA p = 0.0022, Oslo p = 0.0495) and disease-free (TCGA p = 0.0495, Oslo p = 0.0282) survival. In addition, our extensive tumour microenvironment analysis reveals differences between the two CMS subtypes, with CMS-C1 tumours exhibit increased lymphocyte presence, while CMS-C2 tumours show high nuclear pleomorphism, elevated neutrophil-to-lymphocyte ratio, and higher malignancy, correlating with poor prognosis. This study introduces a potentially clinically advantageous Digital-CMS score derived from digitised WSIs of routine H&E-stained tissue sections, offers new insights into TME differences impacting patient prognosis and potential therapeutic targets, and identifies histological patterns serving as potential surrogate markers of the CMS subtypes for clinical application.

PMID:39799271 | DOI:10.1038/s41698-024-00778-5

Categories: Literature Watch

Unsupervised deep learning of electrocardiograms enables scalable human disease profiling

Sat, 2025-01-11 06:00

NPJ Digit Med. 2025 Jan 12;8(1):23. doi: 10.1038/s41746-024-01418-9.

ABSTRACT

The 12-lead electrocardiogram (ECG) is inexpensive and widely available. Whether conditions across the human disease landscape can be detected using the ECG is unclear. We developed a deep learning denoising autoencoder and systematically evaluated associations between ECG encodings and ~1,600 Phecode-based diseases in three datasets separate from model development, and meta-analyzed the results. The latent space ECG model identified associations with 645 prevalent and 606 incident Phecodes. Associations were most enriched in the circulatory (n = 140, 82% of category-specific Phecodes), respiratory (n = 53, 62%) and endocrine/metabolic (n = 73, 45%) categories, with additional associations across the phenome. The strongest ECG association was with hypertension (p < 2.2×10-308). The ECG latent space model demonstrated more associations than models using standard ECG intervals, and offered favorable discrimination of prevalent disease compared to models comprising age, sex, and race. We further demonstrate how latent space models can be used to generate disease-specific ECG waveforms and facilitate individual disease profiling.

PMID:39799251 | DOI:10.1038/s41746-024-01418-9

Categories: Literature Watch

Improving spleen segmentation in ultrasound images using a hybrid deep learning framework

Sat, 2025-01-11 06:00

Sci Rep. 2025 Jan 11;15(1):1670. doi: 10.1038/s41598-025-85632-9.

ABSTRACT

This paper introduces a novel method for spleen segmentation in ultrasound images, using a two-phase training approach. In the first phase, the SegFormerB0 network is trained to provide an initial segmentation. In the second phase, the network is further refined using the Pix2Pix structure, which enhances attention to details and corrects any erroneous or additional segments in the output. This hybrid method effectively combines the strengths of both SegFormer and Pix2Pix to produce highly accurate segmentation results. We have assembled the Spleenex dataset, consisting of 450 ultrasound images of the spleen, which is the first dataset of its kind in this field. Our method has been validated on this dataset, and the experimental results show that it outperforms existing state-of-the-art models. Specifically, our approach achieved a mean Intersection over Union (mIoU) of 94.17% and a mean Dice (mDice) score of 96.82%, surpassing models such as Splenomegaly Segmentation Network (SSNet), U-Net, and Variational autoencoder based methods. The proposed method also achieved a Mean Percentage Length Error (MPLE) of 3.64%, further demonstrating its accuracy. Furthermore, the proposed method has demonstrated strong performance even in the presence of noise in ultrasound images, highlighting its practical applicability in clinical environments.

PMID:39799236 | DOI:10.1038/s41598-025-85632-9

Categories: Literature Watch

A benchmark of deep learning approaches to predict lung cancer risk using national lung screening trial cohort

Sat, 2025-01-11 06:00

Sci Rep. 2025 Jan 11;15(1):1736. doi: 10.1038/s41598-024-84193-7.

ABSTRACT

Deep learning (DL) methods have demonstrated remarkable effectiveness in assisting with lung cancer risk prediction tasks using computed tomography (CT) scans. However, the lack of comprehensive comparison and validation of state-of-the-art (SOTA) models in practical settings limits their clinical application. This study aims to review and analyze current SOTA deep learning models for lung cancer risk prediction (malignant-benign classification). To evaluate our model's general performance, we selected 253 out of 467 patients from a subset of the National Lung Screening Trial (NLST) who had CT scans without contrast, which are the most commonly used, and divided them into training and test cohorts. The CT scans were preprocessed into 2D-image and 3D-volume formats according to their nodule annotations. We evaluated ten 3D and eleven 2D SOTA deep learning models, which were pretrained on large-scale general-purpose datasets (Kinetics and ImageNet) and radiological datasets (3DSeg-8, nnUnet and RadImageNet), for their lung cancer risk prediction performance. Our results showed that 3D-based deep learning models generally perform better than 2D models. On the test cohort, the best-performing 3D model achieved an AUROC of 0.86, while the best 2D model reached 0.79. The lowest AUROCs for the 3D and 2D models were 0.70 and 0.62, respectively. Furthermore, pretraining on large-scale radiological image datasets did not show the expected performance advantage over pretraining on general-purpose datasets. Both 2D and 3D deep learning models can handle lung cancer risk prediction tasks effectively, although 3D models generally have superior performance than their 2D competitors. Our findings highlight the importance of carefully selecting pretrained datasets and model architectures for lung cancer risk prediction. Overall, these results have important implications for the development and clinical integration of DL-based tools in lung cancer screening.

PMID:39799226 | DOI:10.1038/s41598-024-84193-7

Categories: Literature Watch

Pages