Deep learning

AFMDD: Analyzing Functional Connectivity Feature of Major Depressive Disorder by Graph Neural Network-Based Model

Mon, 2025-02-03 06:00

J Comput Biol. 2025 Feb 3. doi: 10.1089/cmb.2024.0505. Online ahead of print.

ABSTRACT

The extraction of biomarkers from functional connectivity (FC) in the brain is of great significance for the diagnosis of mental disorders. In recent years, with the development of deep learning, several methods have been proposed to assist in the diagnosis of depression and promote its automatic identification. However, these methods still have some limitations. The current approaches overlook the importance of subgraphs in brain graphs, resulting in low accuracy. Using these methods with low accuracy for FC analysis may lead to unreliable results. To address these issues, we have designed a graph neural network-based model called AFMDD, specifically for analyzing FC features of depression and depression identification. Through experimental validation, our model has demonstrated excellent performance in depression diagnosis, achieving an accuracy of 73.15%, surpassing many state-of-the-art methods. In our study, we conducted visual analysis of nodes and edges in the FC networks of depression and identified several novel FC features. Those findings may provide valuable clues for the development of biomarkers for the clinical diagnosis of depression.

PMID:39899351 | DOI:10.1089/cmb.2024.0505

Categories: Literature Watch

Automated Patient-specific Quality Assurance for Automated Segmentation of Organs at Risk in Nasopharyngeal Carcinoma Radiotherapy

Mon, 2025-02-03 06:00

Cancer Control. 2025 Jan-Dec;32:10732748251318387. doi: 10.1177/10732748251318387.

ABSTRACT

INTRODUCTION: Precision radiotherapy relies on accurate segmentation of tumor targets and organs at risk (OARs). Clinicians manually review automatically delineated structures on a case-by-case basis, a time-consuming process dependent on reviewer experience and alertness. This study proposes a general process for automated threshold generation for structural evaluation indicators and patient-specific quality assurance (QA) for automated segmentation of nasopharyngeal carcinoma (NPC).

METHODS: The patient-specific QA process for automated segmentation involves determining the confidence limit and error structure highlight stage. Three expert physicians segmented 17 OARs using computed tomography images of NPC and compared them using the Dice similarity coefficient, the maximum Hausdorff distance, and the mean distance to agreement. For each OAR, the 95% confidence interval was calculated as the confidence limit for each indicator. If two or more evaluation indicators (N2) or one or more evaluation indicators (N1) exceeded the confidence limits, the structure segmentation result was considered abnormal. The quantitative performances of these two methods were compared with those obtained by artificially introducing small/medium and serious errors.

RESULTS: The sensitivity, specificity, balanced accuracy, and F-score values for N2 were 0.944 ± 0.052, 0.827 ± 0.149, 0.886 ± 0.076, and 0.936 ± 0.045, respectively, whereas those for N1 were 0.955 ± 0.045, 0.788 ± 0.189, 0.878 ± 0.096, and 0.948 ± 0.035, respectively. N2 and N1 had small/medium error detection rates of 97.67 ± 0.04% and 98.67 ± 0.04%, respectively, with a serious error detection rate of 100%.

CONCLUSION: The proposed automated patient-specific QA process effectively detected segmentation abnormalities, particularly serious errors. These are crucial for enhancing review efficiency and automated segmentation, and for improving physician confidence in automated segmentation.

PMID:39899269 | DOI:10.1177/10732748251318387

Categories: Literature Watch

A Multi-View Feature-Based Interpretable Deep Learning Framework for Drug-Drug Interaction Prediction

Mon, 2025-02-03 06:00

Interdiscip Sci. 2025 Feb 3. doi: 10.1007/s12539-025-00687-6. Online ahead of print.

ABSTRACT

Drug-drug interactions (DDIs) can result in deleterious consequences when patients take multiple medications simultaneously, emphasizing the critical need for accurate DDI prediction. Computational methods for DDI prediction have garnered recent attention. However, current approaches concentrate solely on single-view features, such as atomic-view or substructure-view features, limiting predictive capacity. The scarcity of research on interpretability studies based on multi-view features is crucial for tracing interactions. Addressing this gap, we present MI-DDI, a multi-view feature-based interpretable deep learning framework for DDI. To fully extract multi-view features, we employ a Message Passing Neural Network (MPNN) to learn atomic features from molecular graphs generated by RDkit, and transformer encoders are used to learn substructure-view embeddings from drug SMILES simultaneously. These atomic-view and substructure-view features are then amalgamated into a holistic drug embedding matrix. Subsequently, an intricately designed interaction module not only establishes a tractable path for understanding interactions but also directly informs the construction of weight matrices, enabling precise and interpretable interaction predictions. Validation on the BIOSNAP dataset and DrugBank dataset demonstrates MI-DDI's superiority. It surpasses the current benchmarks by a substantial average of 3% on BIOSNAP and 1% on DrugBank. Additional experiments underscore the significance of atomic-view information for DDI prediction and confirm that our interaction module indeed learns more effective information for DDI prediction. The source codes are available at https://github.com/ZihuiCheng/MI-DDI .

PMID:39899225 | DOI:10.1007/s12539-025-00687-6

Categories: Literature Watch

Multi-modal dataset creation for federated learning with DICOM-structured reports

Mon, 2025-02-03 06:00

Int J Comput Assist Radiol Surg. 2025 Feb 3. doi: 10.1007/s11548-025-03327-y. Online ahead of print.

ABSTRACT

Purpose Federated training is often challenging on heterogeneous datasets due to divergent data storage options, inconsistent naming schemes, varied annotation procedures, and disparities in label quality. This is particularly evident in the emerging multi-modal learning paradigms, where dataset harmonization including a uniform data representation and filtering options are of paramount importance.Methods DICOM-structured reports enable the standardized linkage of arbitrary information beyond the imaging domain and can be used within Python deep learning pipelines with highdicom. Building on this, we developed an open platform for data integration with interactive filtering capabilities, thereby simplifying the process of creation of patient cohorts over several sites with consistent multi-modal data.Results In this study, we extend our prior work by showing its applicability to more and divergent data types, as well as streamlining datasets for federated training within an established consortium of eight university hospitals in Germany. We prove its concurrent filtering ability by creating harmonized multi-modal datasets across all locations for predicting the outcome after minimally invasive heart valve replacement. The data include imaging and waveform data (i.e., computed tomography images, electrocardiography scans) as well as annotations (i.e., calcification segmentations, and pointsets), and metadata (i.e., prostheses and pacemaker dependency).Conclusion Structured reports bridge the traditional gap between imaging systems and information systems. Utilizing the inherent DICOM reference system arbitrary data types can be queried concurrently to create meaningful cohorts for multi-centric data analysis. The graphical interface as well as example structured report templates are available at https://github.com/Cardio-AI/fl-multi-modal-dataset-creation .

PMID:39899185 | DOI:10.1007/s11548-025-03327-y

Categories: Literature Watch

Multi-scale dual attention embedded U-shaped network for accurate segmentation of coronary vessels in digital subtraction angiography

Mon, 2025-02-03 06:00

Med Phys. 2025 Feb 3. doi: 10.1002/mp.17618. Online ahead of print.

ABSTRACT

BACKGROUND: Most attention-based networks fall short in effectively integrating spatial and channel-wise information across different scales, which results in suboptimal performance for segmenting coronary vessels in x-ray digital subtraction angiography (DSA) images. This limitation becomes particularly evident when attempting to identify tiny sub-branches.

PURPOSE: To address this limitation, a multi-scale dual attention embedded network (named MDA-Net) is proposed to consolidate contextual spatial and channel information across contiguous levels and scales.

METHODS: MDA-Net employs five cascaded double-convolution blocks within its encoder to adeptly extract multi-scale features. It incorporates skip connections that facilitate the retention of low-level feature details throughout the decoding phase, thereby enhancing the reconstruction of detailed image information. Furthermore, MDA modules, which take in features from neighboring scales and hierarchical levels, are tasked with discerning subtle distinctions between foreground elements, such as coronary vessels of diverse morphologies and dimensions, and the complex background, which includes structures like catheters or other tissues with analogous intensities. To sharpen the segmentation accuracy, the network utilizes a composite loss function that integrates intersection over union (IoU) loss with binary cross-entropy loss, ensuring the precision of the segmentation outcomes and maintaining an equilibrium between positive and negative classifications.

RESULTS: Experimental results demonstrate that MDA-Net not only performs more robustly and effectively on DSA images under various image conditions, but also achieves significant advantages over state-of-the-art methods, achieving the optimal scores in terms of IoU, Dice, accuracy, and Hausdorff distance 95%.

CONCLUSIONS: MDA-Net has high robustness for coronary vessels segmentation, providing an active strategy for early diagnosis of cardiovascular diseases. The code is publicly available at https://github.com/30410B/MDA-Net.git.

PMID:39899182 | DOI:10.1002/mp.17618

Categories: Literature Watch

Redefining healthcare - The transformative power of generative AI in modern medicine

Mon, 2025-02-03 06:00

Rev Esp Enferm Dig. 2025 Feb 3. doi: 10.17235/reed.2025.11081/2024. Online ahead of print.

ABSTRACT

Over the last decade, technological advances in deep learning (artificial neural networks, big data and computing power) have made possible to build digital solutions that imitate human cognitive process (language, vision, hearing, etc) and are able to generate new content when prompted. This generative AI is going to disrupt healthcare. Healthcare professionals must get prepared because there are ethical and legal challenges that must be identified and tackled.

PMID:39898717 | DOI:10.17235/reed.2025.11081/2024

Categories: Literature Watch

Accuracy of a Cascade Network for Semi-Supervised Maxillary Sinus Detection and Sinus Cyst Classification

Mon, 2025-02-03 06:00

Clin Implant Dent Relat Res. 2025 Feb;27(1):e13431. doi: 10.1111/cid.13431.

ABSTRACT

OBJECTIVE: Maxillary sinus mucosal cysts represent prevalent oral and maxillofacial diseases, and their precise diagnosis is essential for surgical planning in maxillary sinus floor elevation. This study aimed to develop a deep learning-based pipeline for the classification of maxillary sinus lesions in cone beam computed tomography (CBCT) images to provide auxiliary support for clinical diagnosis.

METHODS: This study utilized 45 136 maxillary sinus images from CBCT scans of 541 patients. A cascade network was designed, comprising a semi-supervised maxillary sinus area object detection module and a maxillary sinus lesions classification module. The object detection module employed a semi-supervised pseudo-labelling training strategy to expand the maxillary sinus annotation dataset. In the classification module, the performance of Convolutional Neural Network and Transformer architectures was compared for maxillary sinus mucosal lesion classification. The object detection and classification modules were evaluated using metrics including Accuracy, Precision, Recall, F1 score, and Average Precision, with the object detection module additionally assessed using Precision-Recall Curve.

RESULTS: The fully supervised pseudo-label generation model achieved an average accuracy of 0.9433, while the semi-supervised maxillary sinus detection model attained 0.9403. ResNet-50 outperformed in classification, with accuracies of 0.9836 (sagittal) and 0.9797 (coronal). Grad-CAM visualization confirmed accurate focus on clinically relevant lesion features.

CONCLUSION: The proposed pipeline achieves high-precision detection and classification of maxillary sinus mucosal lesions, reducing manual annotation while maintaining accuracy.

PMID:39898709 | DOI:10.1111/cid.13431

Categories: Literature Watch

Enhancing feature-aided data association tracking in passive sonar arrays: An advanced Siamese network approach

Mon, 2025-02-03 06:00

J Acoust Soc Am. 2025 Feb 1;157(2):681-698. doi: 10.1121/10.0035577.

ABSTRACT

Feature-aided tracking integrates supplementary features into traditional methods and improves the accuracy of data association methods that rely solely on kinematic measurements. However, previous applications of feature-aided data association methods in multi-target tracking of passive sonar arrays directly utilized raw features for likelihood calculations, causing performance degradation in complex marine scenarios with low signal-to-noise ratio and close-proximity trajectories. Inspired by the successful application of deep learning, this study proposes BiChannel-SiamDinoNet, an advanced network derived from the Siamese network and integrated into the joint probability data association framework to calculate feature measurement likelihood. This method forms an embedding space through the feature structure of acoustic targets, bringing similar targets closer together. This makes the system more robust to variations, capable of capturing complex relationships between measurements and targets and effectively discriminating discrepancies between them. Additionally, this study refines the network's feature extraction module to address underwater acoustic signals' unique line spectrum and implement the knowledge distillation training method to improve the network's capability to assess consistency between features through local representations. The performance of the proposed method is assessed through simulation analysis and marine experiments.

PMID:39898705 | DOI:10.1121/10.0035577

Categories: Literature Watch

Enhancing U-Net-based Pseudo-CT generation from MRI using CT-guided bone segmentation for radiation treatment planning in head & neck cancer patients

Mon, 2025-02-03 06:00

Phys Med Biol. 2025 Jan 31. doi: 10.1088/1361-6560/adb124. Online ahead of print.

ABSTRACT

OBJECTIVE: This study investigates the effects of various training protocols on enhancing the precision of MRI-only Pseudo-CT generation for radiation treatment planning and adaptation in head & neck cancer patients. It specifically tackles the challenge of differentiating bone from air, a limitation that frequently results in substantial deviations in the representation of bony structures on Pseudo-CT images.

APPROACH: The study included 25 patients, utilizing pre-treatment MRI-CT image pairs. Five cases were randomly selected for testing, with the remaining 20 used for model training and validation. A 3D U-Net deep learning model was employed, trained on patches of size 643with an overlap of 323. MRI scans were acquired using the Dixon gradient echo (GRE) technique, and various contrasts were explored to improve Pseudo-CT accuracy, including in-phase, water-only, and combined water-only and fat-only images. Additionally, bone extraction from the fat-only image was integrated as an additional channel to better capture bone structures on Pseudo-CTs. The evaluation involved both image quality and dosimetric metrics.

MAIN RESULTS: The generated Pseudo-CTs were compared with their corresponding registered target CTs. The mean absolute error (MAE) and peak signal-to-noise ratio (PSNR) for the base model using combined water-only and fat-only images were 19.20 ± 5.30 HU and 57.24 ± 1.44 dB, respectively. Following the integration of an additional channel using a CT-guided bone segmentation, the model's performance improved, achieving MAE and PSNR of 18.32 ± 5.51 HU and 57.82 ± 1.31 dB, respectively. The dosimetric assessment confirmed that radiation treatment planning on Pseudo-CT achieved accuracy comparable to conventional CT. The measured results are statistically significant, with ap-value < 0.05.

SIGNIFICANCE: This study demonstrates improved accuracy in bone representation on Pseudo-CTs achieved through a combination of water-only, fat-only and extracted bone images; thus, enhancing feasibility of MRI-based simulation for radiation treatment planning.

PMID:39898433 | DOI:10.1088/1361-6560/adb124

Categories: Literature Watch

Automated Detection and Severity Prediction of Wheat Rust Using Cost-Effective Xception Architecture

Mon, 2025-02-03 06:00

Plant Cell Environ. 2025 Feb 3. doi: 10.1111/pce.15413. Online ahead of print.

ABSTRACT

Wheat crop production is under constant threat from leaf and stripe rust, an airborne fungal disease caused by the pathogen Puccinia triticina. Early detection and efficient crop phenotyping are crucial for managing and controlling the spread of this disease in susceptible wheat varieties. Current detection methods are predominantly manual and labour-intensive. Traditional strategies such as cultivating resistant varieties, applying fungicides and practicing good agricultural techniques often fall short in effectively identifying and responding to wheat rust outbreaks. To address these challenges, we propose an innovative computer vision-based disease severity prediction pipeline. Our approach utilizes a deep learning-based classifier to differentiate between healthy and rust-infected wheat leaves. Upon identifying an infected leaf, we apply Grabcut-based segmentation to isolate the foreground mask. This mask is then processed in the CIELAB color space to distinguish leaf rust stripes and spores. The disease severity ratio is calculated to measure the extent of infection on each test leaf. This paper introduces a ground-breaking disease severity prediction method, offering a low-cost, accessible and automated solution for wheat rust disease screening in field conditions using digital colour images. Our approach represents a significant advancement in crop disease management, promising timely interventions and better control measures for wheat rust.

PMID:39898421 | DOI:10.1111/pce.15413

Categories: Literature Watch

Development and Evaluation of a Deep Learning-Based Pulmonary Hypertension Screening Algorithm Using a Digital Stethoscope

Mon, 2025-02-03 06:00

J Am Heart Assoc. 2025 Feb 3:e036882. doi: 10.1161/JAHA.124.036882. Online ahead of print.

ABSTRACT

BACKGROUND: Despite the poor outcomes related to the presence of pulmonary hypertension, it often goes undiagnosed in part because of low suspicion and screening tools not being easily accessible such as echocardiography. A new readily available screening tool to identify elevated pulmonary artery systolic pressures is needed to help with the prognosis and timely treatment of underlying causes such as heart failure or pulmonary vascular remodeling. We developed a deep learning-based method that uses phonocardiograms (PCGs) for the detection of elevated pulmonary artery systolic pressure, an indicator of pulmonary hypertension.

METHODS: Approximately 6000 PCG recordings with the corresponding echocardiogram-based estimated pulmonary artery systolic pressure values, as well as ≈169 000 PCG recordings without associated echocardiograms, were used for training a deep convolutional network to detect pulmonary artery systolic pressures ≥40 mm Hg in a semisupervised manner. Each 15-second PCG, recorded using a digital stethoscope, was processed to generate 5-second mel-spectrograms. An additional labeled data set of 196 patients was used for testing. GradCAM++ was used to visualize high importance segments contributing to the network decision.

RESULTS: An average area under the receiver operator characteristic curve of 0.79 was obtained across 5 cross-validation folds. The testing data set gave a sensitivity of 0.71 and a specificity of 0.73, with pulmonic and left subclavicular locations having higher sensitivities. GradCAM++ technique highlighted physiologically meaningful PCG segments in example pulmonary hypertension recordings.

CONCLUSIONS: We demonstrated the feasibility of using digital stethoscopes in conjunction with deep learning algorithms as a low-cost, noninvasive, and easily accessible screening tool for early detection of pulmonary hypertension.

PMID:39895552 | DOI:10.1161/JAHA.124.036882

Categories: Literature Watch

CT-based radiomics: A potential indicator of KRAS mutation in pulmonary adenocarcinoma

Mon, 2025-02-03 06:00

Tumori. 2025 Feb 2:3008916251314659. doi: 10.1177/03008916251314659. Online ahead of print.

ABSTRACT

PURPOSE: This study aimed to validate a CT-based radiomics signature for predicting Kirsten rat sarcoma (KRAS) mutation status in lung adenocarcinoma (LADC).

MATERIALS AND METHODS: A total of 815 LADC patients were included. Radiomics features were extracted from non-contrast-enhanced CT (NECT) and contrast-enhanced CT (CECT) images using Pyradiomics. CT-based radiomics were combined with clinical features to distinguish KRAS mutation status. Four feature selection methods and four deep learning classifiers were employed. Data was split into 70% training and 30% test sets, with SMOTE addressing imbalance in the training set. Model performance was evaluated using AUC, accuracy, precision, F1 score, and recall.

RESULTS: The analysis revealed that 10.4% of patients showed KRAS mutations. The study extracted 1061 radiomics features and combined them with 17 clinical features. After feature selection, two signatures were constructed using top 10, 20, and 50 features. The best performance was achieved using Multilayer Perceptron with 20 features. CECT, it showed 66% precision, 76% recall, 69% F1-score, 84% accuracy, and AUC of 93.3% and 87.4% for train and test sets, respectively. For NECT, accuracy was 85% and 82%, with AUC of 90.7% and 87.6% for train and test sets, respectively.

CONCLUSIONS: CT-based radiomics signature is a noninvasive method that can predict KRAS mutation status of LADC when mutational profiling is unavailable.

PMID:39894961 | DOI:10.1177/03008916251314659

Categories: Literature Watch

Unveiling encephalopathy signatures: A deep learning approach with locality-preserving features and hybrid neural network for EEG analysis

Sun, 2025-02-02 06:00

Neurosci Lett. 2025 Jan 31:138146. doi: 10.1016/j.neulet.2025.138146. Online ahead of print.

ABSTRACT

EEG signals exhibit spatio-temporal characteristics due to the neural activity dispersion in space over the brain and the dynamic temporal patterns of electrical activity in neurons. This study tries to effectively utilize the spatio-temporal nature of EEG signals for diagnosing encephalopathy using a combination of novel locality preserving feature extraction using Local Binary Patterns (LBP) and a custom fine-tuned Long Short-Term Memory (LSTM) neural network. A carefully curated primary EEG dataset is used to assess the effectiveness of the technique for treatment of encephalopathies. EEG signals of all electrodes are mapped onto a spatial matrix from which the custom feature extraction method isolates spatial features of the signals. These spatial features are further given to the neural network, which learns to combine the spatial information with temporal dynamics summarizing pertinent details from the raw EEG data. Such a unified representation is key to perform reliable disease classification at the output layer of the neural network, leading to a robust classification system, potentially providing improved diagnosis and treatment. The proposed method shows promising potential for enhancing the automated diagnosis of encephalopathy, with a remarkable accuracy rate of 90.5%. To the best of our knowledge, this is the first attempt to compress and represent both spatial and temporal features into a single vector for encephalopathy detection, simplifying visual diagnosis and providing a robust feature for automated predictions. This advancement holds significant promise for ensuring early detection and intervention strategies in the clinical environment, which in turn enhances patient care.

PMID:39894198 | DOI:10.1016/j.neulet.2025.138146

Categories: Literature Watch

NLP for Analyzing Electronic Health Records and Clinical Notes in Cancer Research: A Review

Sun, 2025-02-02 06:00

J Pain Symptom Manage. 2025 Jan 31:S0885-3924(25)00037-5. doi: 10.1016/j.jpainsymman.2025.01.019. Online ahead of print.

ABSTRACT

This review examines the application of natural language processing (NLP) techniques in cancer research using electronic health records (EHRs) and clinical notes. It addresses gaps in existing literature by providing a broader perspective than previous studies focused on specific cancer types or applications. A comprehensive literature search in the Scopus database identified 94 relevant studies published between 2019 and 2024. The analysis revealed a growing trend in NLP applications for cancer research, with information extraction (47 studies) and text classification (40 studies) emerging as predominant NLP tasks, followed by named entity recognition (7 studies). Among cancer types, breast, lung, and colorectal cancers were found to be the most studied. A significant shift from rule-based and traditional machine learning approaches to advanced deep learning techniques and transformer-based models was observed. It was found that dataset sizes used in existing studies varied widely, ranging from small, manually annotated datasets to large-scale EHRs. The review highlighted key challenges, including the limited generalizability of proposed solutions and the need for improved integration into clinical workflows. While NLP techniques show significant potential in analyzing EHRs and clinical notes for cancer research, future work should focus on improving model generalizability, enhancing robustness in handling complex clinical language, and expanding applications to understudied cancer types. The integration of NLP tools into palliative medicine and addressing ethical considerations remain crucial for utilizing the full potential of NLP in enhancing cancer diagnosis, treatment, and patient outcomes. This review provides valuable insights into the current state and future directions of NLP applications in cancer research.

PMID:39894080 | DOI:10.1016/j.jpainsymman.2025.01.019

Categories: Literature Watch

ABIET: An explainable transformer for identifying functional groups in biological active molecules

Sun, 2025-02-02 06:00

Comput Biol Med. 2025 Feb 1;187:109740. doi: 10.1016/j.compbiomed.2025.109740. Online ahead of print.

ABSTRACT

Recent advancements in deep learning have revolutionized the field of drug discovery, with Transformer-based models emerging as powerful tools for molecular design and property prediction. However, the lack of explainability in such models remains a significant challenge. In this study, we introduce ABIET (Attention-Based Importance Estimation Tool), an explainable Transformer model designed to identify the most critical regions for drug-target interactions - functional groups (FGs) - in biologically active molecules. Functional groups play a pivotal role in determining chemical behavior and biological interactions. Our approach leverages attention weights from Transformer-encoder architectures trained on SMILES representations to assess the relative importance of molecular subregions. By processing attention scores using a specific strategy - considering bidirectional interactions, layer-based extraction, and activation transformations - we effectively distinguish FGs from non-FG atoms. Experimental validation on diverse datasets targeting pharmacological receptors, including VEGFR2, AA2A, GSK3, JNK3, and DRD2, demonstrates the model's robustness and interpretability. Comparative analysis with state-of-the-art gradient-based and perturbation-based methods confirms ABIET's superior performance, with functional groups receiving statistically higher importance scores. This work enhances the transparency of Transformer predictions, providing critical insights for molecular design, structure-activity analysis, and targeted drug development.

PMID:39894011 | DOI:10.1016/j.compbiomed.2025.109740

Categories: Literature Watch

Attention-based deep learning models for predicting anomalous shock of wastewater treatment plants

Sun, 2025-02-02 06:00

Water Res. 2025 Jan 23;275:123192. doi: 10.1016/j.watres.2025.123192. Online ahead of print.

ABSTRACT

Quickly grasping the time-consuming water quality indicators (WQIs) such as total nitrogen (TN) and total phosphorus (TP) of influent is an essential prerequisite for wastewater treatment plants (WWTPs) to prompt respond to sudden shock loads. Soft detection methods based on machine learning models, especially deep learning models, perform well in predicting the normal fluctuations of these time-consuming WQIs but hardly predict their sudden fluctuations mainly due to the lack of extreme fluctuation data for model training. This work employs attention mechanisms to aid deep learning models in learning patterns of anomalous water quality. The lack of interpretability has always hindered deep learning models from optimizing for different application scenarios. Therefore, the local and global sensitivity analyses are performed based on the best-performing attention-based deep learning and ordinary machine learning models, respectively, allowing for reliable feature importance quantification with a small computational burden. In the case study, three types of attention-based deep learning models were developed, including attention-based multilayer perceptron (A-MLP), Transformer composed of stacked A-MLP encoder and A-MLP decoder, and feature-temporal attention-based long short-term memory (FTA-LSTM) neural network with encoder-decoder architecture. These developed attention-based deep learning models consistently outperform the corresponding baseline models in predicting the testing set of TN, TP, and chemical oxygen demand (COD) time series and the anomalous values therein, clearly demonstrating the positive effect of the integrated attention mechanism. Among them, the prediction performance of FTA-LSTM outperforms A-MLP and Transformer (2.01-38.48 % higher R2, 0-85.14 % higher F1-score, 0-62.57 % higher F2-score). Predicting anomalous water quality using attention-based deep learning models is a novel attempt that drives the WWTPs' operation towards being safer, cleaner, and more cost-efficient.

PMID:39893907 | DOI:10.1016/j.watres.2025.123192

Categories: Literature Watch

Deep learning to decode sites of RNA translation in normal and cancerous tissues

Sun, 2025-02-02 06:00

Nat Commun. 2025 Feb 2;16(1):1275. doi: 10.1038/s41467-025-56543-0.

ABSTRACT

The biological process of RNA translation is fundamental to cellular life and has wide-ranging implications for human disease. Accurate delineation of RNA translation variation represents a significant challenge due to the complexity of the process and technical limitations. Here, we introduce RiboTIE, a transformer model-based approach designed to enhance the analysis of ribosome profiling data. Unlike existing methods, RiboTIE leverages raw ribosome profiling counts directly to robustly detect translated open reading frames (ORFs) with high precision and sensitivity, evaluated on a diverse set of datasets. We demonstrate that RiboTIE successfully recapitulates known findings and provides novel insights into the regulation of RNA translation in both normal brain and medulloblastoma cancer samples. Our results suggest that RiboTIE is a versatile tool that can significantly improve the accuracy and depth of Ribo-Seq data analysis, thereby advancing our understanding of protein synthesis and its implications in disease.

PMID:39894899 | DOI:10.1038/s41467-025-56543-0

Categories: Literature Watch

3D convolutional deep learning for nonlinear estimation of body composition from whole body morphology

Sun, 2025-02-02 06:00

NPJ Digit Med. 2025 Feb 2;8(1):79. doi: 10.1038/s41746-025-01469-6.

ABSTRACT

Body composition prediction from 3D optical imagery has previously been studied with linear algorithms. In this study, we present a novel application of deep 3D convolutional graph networks and nonlinear Gaussian process regression for human body shape parameterization and body composition estimation. We trained and tested linear and nonlinear models with ablation studies on a novel ensemble body shape dataset containing 4286 scans. Nonlinear GPR produced up to a 20% reduction in prediction error and up to a 30% increase in precision over linear regression for both sexes in 10 tested body composition variables. Deep shape features produced 6-8% reduction in prediction error over linear PCA features for males only, and a 4-14% reduction in precision error for both sexes. All coefficients of determination (R2) for all predicted variables were above 0.86 and achieved lower estimation RMSEs than all previous work on 10 metrics of body composition.

PMID:39894882 | DOI:10.1038/s41746-025-01469-6

Categories: Literature Watch

Functional feature extraction and validation from twelve-lead electrocardiograms to identify atrial fibrillation

Sun, 2025-02-02 06:00

Commun Med (Lond). 2025 Feb 2;5(1):32. doi: 10.1038/s43856-025-00749-2.

ABSTRACT

BACKGROUND: Deep learning methods on standard, 12-lead electrocardiograms (ECG) have resulted in the ability to identify individuals at high-risk for the development of atrial fibrillation. However, the process remains a "black box" and does not help clinicians in understanding the electrocardiographic changes at an individual level. we propose a nonparametric feature extraction approach to identify features that are associated with the development of atrial fibrillation (AF).

METHODS: We apply functional principal component analysis to the raw ECG tracings collected in the Chronic Renal Insufficiency Cohort (CRIC) study. We define and select the features using ECGs from participants enrolled in Phase I (2003-2008) of the study. Cox proportional hazards models are used to evaluate the association of selected ECG features and their changes with the incident risk of AF during study follow-up. The findings are then validated in ECGs from participants enrolled in Phase III (2013-2015).

RESULTS: We identify four features that are related to the P-wave amplitude, QRS complex and ST segment. Both their initial measurement and 3-year changes are associated with the development of AF. In particular, one standard deviation in the 3-year decline of the P-wave amplitude is independently associated with a 29% increased risk of incident AF in the multivariable model (HR: 1.29, 95% CI: [1.16, 1.43]).

CONCLUSIONS: Compared with deep learning methods, our features are intuitive and can provide insights into the longitudinal ECG changes at an individual level that precede the development of AF.

PMID:39894874 | DOI:10.1038/s43856-025-00749-2

Categories: Literature Watch

Optimization of sparse-view CT reconstruction based on convolutional neural network

Sun, 2025-02-02 06:00

Med Phys. 2025 Feb 2. doi: 10.1002/mp.17636. Online ahead of print.

ABSTRACT

BACKGROUND: Sparse-view CT shortens scan time and reduces radiation dose but results in severe streak artifacts due to insufficient sampling data. Deep learning methods can now suppress these artifacts and improve image quality in sparse-view CT reconstruction.

PURPOSE: The quality of sparse-view CT reconstructed images can still be improved. Additionally, the interpretability of deep learning-based optimization methods for these reconstruction images is lacking, and the role of different network layers in artifact removal requires further study. Moreover, the optimization capability of these methods for reconstruction images from various sparse views needs enhancement. This study aims to improve the network's optimization ability for sparse-view reconstructed images, enhance interpretability, and boost generalization by establishing multiple network structures and datasets.

METHODS: In this paper, we developed a sparse-view CT reconstruction images improvement network (SRII-Net) based on U-Net. We added a copy pathway in the network and designed a residual image output block to boost the network's performance. Multiple networks with different connectivity structures were established using SRII-Net to analyze the contribution of each layer to artifact removal, improving the network's interpretability. Additionally, we created multiple datasets with reconstructed images of various sampling views to train and test the proposed network, investigating how these datasets from different sampling views affect the network's generalization ability.

RESULTS: The results show that the proposed method outperforms current networks, with significant improvements in metrics like PSNR and SSIM. Image optimization time is at the millisecond level. By comparing the performance of different network structures, we've identified the impact of various hierarchical structures. The image detail information learned by shallow layers and the high-level abstract feature information learned by deep layers play a crucial role in optimizing sparse-view CT reconstruction images. Training the network with multiple mixed datasets revealed that, under a certain amount of data, selecting the appropriate categories of sampling views and their corresponding samples can effectively enhance the network's optimization ability for reconstructing images with different sampling views.

CONCLUSIONS: The network in this paper effectively suppresses artifacts in reconstructed images with different sparse views, improving generalization. We have also created diverse network structures and datasets to deepen the understanding of artifact removal in deep learning networks, offering insights for noise reduction and image enhancement in other imaging methods.

PMID:39894762 | DOI:10.1002/mp.17636

Categories: Literature Watch

Pages