Deep learning

An examination of daily CO(2) emissions prediction through a comparative analysis of machine learning, deep learning, and statistical models

Sun, 2025-01-12 06:00

Environ Sci Pollut Res Int. 2025 Jan 13. doi: 10.1007/s11356-024-35764-8. Online ahead of print.

ABSTRACT

Human-induced global warming, primarily attributed to the rise in atmospheric CO2, poses a substantial risk to the survival of humanity. While most research focuses on predicting annual CO2 emissions, which are crucial for setting long-term emission mitigation targets, the precise prediction of daily CO2 emissions is equally vital for setting short-term targets. This study examines the performance of 14 models in predicting daily CO2 emissions data from 1/1/2022 to 30/9/2023 across the top four polluting regions (China, India, the USA, and the EU27&UK). The 14 models used in the study include four statistical models (ARMA, ARIMA, SARMA, and SARIMA), three machine learning models (support vector machine (SVM), random forest (RF), and gradient boosting (GB)), and seven deep learning models (artificial neural network (ANN), recurrent neural network variations such as gated recurrent unit (GRU), long short-term memory (LSTM), bidirectional-LSTM (BILSTM), and three hybrid combinations of CNN-RNN). Performance evaluation employs four metrics (R2, MAE, RMSE, and MAPE). The results show that the machine learning (ML) and deep learning (DL) models, with higher R2 (0.714-0.932) and lower RMSE (0.480-0.247) values, respectively, outperformed the statistical model, which had R2 (- 0.060-0.719) and RMSE (1.695-0.537) values, in predicting daily CO2 emissions across all four regions. The performance of the ML and DL models was further enhanced by differencing, a technique that improves accuracy by ensuring stationarity and creating additional features and patterns from which the model can learn. Additionally, applying ensemble techniques such as bagging and voting improved the performance of the ML models by approximately 9.6%, whereas hybrid combinations of CNN-RNN enhanced the performance of the RNN models. In summary, the performance of both the ML and DL models was relatively similar. However, due to the high computational requirements associated with DL models, the recommended models for daily CO2 emission prediction are ML models using the ensemble technique of voting and bagging. This model can assist in accurately forecasting daily emissions, aiding authorities in setting targets for CO2 emission reduction.

PMID:39800837 | DOI:10.1007/s11356-024-35764-8

Categories: Literature Watch

Annotation-free deep learning algorithm trained on hematoxylin & eosin images predicts epithelial-to-mesenchymal transition phenotype and endocrine response in estrogen receptor-positive breast cancer

Sun, 2025-01-12 06:00

Breast Cancer Res. 2025 Jan 12;27(1):6. doi: 10.1186/s13058-025-01959-1.

ABSTRACT

Recent evidence indicates that endocrine resistance in estrogen receptor-positive (ER+) breast cancer is closely correlated with phenotypic characteristics of epithelial-to-mesenchymal transition (EMT). Nonetheless, identifying tumor tissues with a mesenchymal phenotype remains challenging in clinical practice. In this study, we validated the correlation between EMT status and resistance to endocrine therapy in ER+ breast cancer from a transcriptomic perspective. To confirm the presence of morphological discrepancies in tumor tissues of ER+ breast cancer classified as epithelial- and mesenchymal-phenotypes according to EMT-related transcriptional features, we trained deep learning algorithms based on EfficientNetV2 architecture to assign the phenotypic status for each patient utilizing hematoxylin & eosin (H&E)-stained slides from The Cancer Genome Atlas database. Our classifier model accurately identified the precise phenotypic status, achieving an area under the curve (AUC) of 0.886 at the tile-level and an AUC of 0.910 at the slide-level. Furthermore, we evaluated the efficacy of the classifier in predicting endocrine response using data from an independent ER+ breast cancer patient cohort. Our classifier achieved a predicting accuracy of 81.25%, and 88.7% slides labeled as endocrine resistant were predicted as the mesenchymal-phenotype, while 75.6% slides labeled as sensitive were predicted as the epithelial-phenotype. Our work introduces an H&E-based framework capable of accurately predicting EMT phenotype and endocrine response for ER+ breast cancer, demonstrating its potential for clinical application and benefit.

PMID:39800743 | DOI:10.1186/s13058-025-01959-1

Categories: Literature Watch

Artificial Intelligence for Cervical Spine Fracture Detection: A Systematic Review of Diagnostic Performance and Clinical Potential

Sun, 2025-01-12 06:00

Global Spine J. 2025 Jan 12:21925682251314379. doi: 10.1177/21925682251314379. Online ahead of print.

ABSTRACT

STUDY DESIGN: Systematic review.

OBJECTIVE: Artificial intelligence (AI) and deep learning (DL) models have recently emerged as tools to improve fracture detection, mainly through imaging modalities such as computed tomography (CT) and radiographs. This systematic review evaluates the diagnostic performance of AI and DL models in detecting cervical spine fractures and assesses their potential role in clinical practice.

METHODS: A systematic search of PubMed/Medline, Embase, Scopus, and Web of Science was conducted for studies published between January 2000 and July 2024. Studies that evaluated AI models for cervical spine fracture detection were included. Diagnostic performance metrics were extracted and included sensitivity, specificity, accuracy, and area under the curve. The PROBAST tool assessed bias, and PRISMA criteria were used for study selection and reporting.

RESULTS: Eleven studies published between 2021 and 2024 were included in the review. AI models demonstrated variable performance, with sensitivity ranging from 54.9% to 100% and specificity from 72% to 98.6%. Models applied to CT imaging generally outperformed those applied to radiographs, with convolutional neural networks (CNN) and advanced architectures such as MobileNetV2 and Vision Transformer (ViT) achieving the highest accuracy. However, most studies lacked external validation, raising concerns about the generalizability of their findings.

CONCLUSIONS: AI and DL models show significant potential in improving fracture detection, particularly in CT imaging. While these models offer high diagnostic accuracy, further validation and refinement are necessary before they can be widely integrated into clinical practice. AI should complement, rather than replace, human expertise in diagnostic workflows.

PMID:39800538 | DOI:10.1177/21925682251314379

Categories: Literature Watch

End-to-end deep-learning model for the detection of coronary artery stenosis on coronary CT images

Sun, 2025-01-12 06:00

Open Heart. 2025 Jan 11;12(1):e002998. doi: 10.1136/openhrt-2024-002998.

ABSTRACT

PURPOSE: We examined whether end-to-end deep-learning models could detect moderate (≥50%) or severe (≥70%) stenosis in the left anterior descending artery (LAD), right coronary artery (RCA) or left circumflex artery (LCX) in iodine contrast-enhanced ECG-gated coronary CT angiography (CCTA) scans.

METHODS: From a database of 6293 CCTA scans, we used pre-existing curved multiplanar reformations (CMR) images of the LAD, RCA and LCX arteries to create end-to-end deep-learning models for the detection of moderate or severe stenoses. We preprocessed the images by exploiting domain knowledge and employed a transfer learning approach using EfficientNet, ResNet, DenseNet and Inception-ResNet, with a class-weighted strategy optimised through cross-validation. Heatmaps were generated to indicate critical areas identified by the models, aiding clinicians in understanding the model's decision-making process.

RESULTS: Among the 900 CMR cases, 279 involved the LAD artery, 259 the RCA artery and 253 the LCX artery. EfficientNet models outperformed others, with EfficientNetB3 and EfficientNetB0 demonstrating the highest accuracy for LAD, EfficientNetB2 for RCA and EfficientNetB0 for LCX. The area under the curve for receiver operating characteristic (AUROC) reached 0.95 for moderate and 0.94 for severe stenosis in the LAD. For the RCA, the AUROC was 0.92 for both moderate and severe stenosis detection. The LCX achieved an AUROC of 0.88 for the detection of moderate stenoses, though the calibration curve exhibited significant overestimation. Calibration curves matched probabilities for the LAD but showed discrepancies for the RCA. Heatmap visualisations confirmed the models' precision in delineating stenotic lesions. Decision curve analysis and net reclassification index assessments reinforced the efficacy of EfficientNet models, confirming their superior diagnostic capabilities.

CONCLUSION: Our end-to-end deep-learning model demonstrates, for the LAD artery, excellent discriminatory ability and calibration during internal validation, despite a small dataset used to train the network. The model reliably produces precise, highly interpretable images.

PMID:39800435 | DOI:10.1136/openhrt-2024-002998

Categories: Literature Watch

Clinical Application Of Deep Learning-assisted Needles Reconstruction In Prostate Ultrasound Brachytherapy

Sun, 2025-01-12 06:00

Int J Radiat Oncol Biol Phys. 2025 Jan 10:S0360-3016(25)00002-1. doi: 10.1016/j.ijrobp.2024.12.026. Online ahead of print.

ABSTRACT

PURPOSE: High dose rate (HDR) prostate brachytherapy (BT) procedure requires image-guided needle insertion. Given that general anesthesia is often employed during the procedure, minimizing overall planning time is crucial. In this study, we explore the clinical feasibility and time-saving potential of artificial intelligence (AI)-driven auto-reconstruction of transperineal needles in the context of US-guided prostate BT planning.

MATERIALS AND METHODS: This study included a total of 102 US-planned BT images from a single institution and split into three groups: 50 for model training and validation, 11 to evaluate reconstruction accuracy (test set), and 41 to evaluate the AI tool in a clinical implementation (clinical set). Reconstruction accuracy for the test set was evaluated by comparing the performance of AI-derived and manually reconstructed needles from 5 medical physicists on the 3D-US scans after treatment. The needle total reconstruction time for the clinical set was defined as the timestamp difference from scan acquisition to the start of dose calculations and was compared to values recorded before the clinical implementation of the AI-assisted tool.

RESULTS: A mean error of (0.44±0.32)mm was found between the AI-reconstructed and the human consensus needle positions in the test set, with 95.0% of AI needle points falling below 1mm from their human-made counterparts. Post-hoc analysis showed only one of the human observers' reconstructions were significantly different from the others including the AI's. In the clinical set, the AI algorithm achieved a true positive reconstruction rate of 93.4% with only 4.5% of these needles requiring manual corrections from the planner before dosimetry. Total time required to perform AI-assisted catheter reconstruction on clinical cases was on average 15.2min lower (p < 0.01) compared to procedure without AI assistance.

CONCLUSIONS: This study demonstrates the feasibility of an AI-assisted needle reconstructing tool for 3D-US based HDR prostate BT. This is a step toward treatment planning automation and increased efficiency in HDR prostate BT.

PMID:39800329 | DOI:10.1016/j.ijrobp.2024.12.026

Categories: Literature Watch

Development and routine implementation of deep learning algorithm for automatic brain metastases segmentation on MRI for RANO-BM criteria follow-up

Sun, 2025-01-12 06:00

Neuroimage. 2025 Jan 10:121002. doi: 10.1016/j.neuroimage.2025.121002. Online ahead of print.

ABSTRACT

RATIONALE AND OBJECTIVES: The RANO-BM criteria, which employ a one-dimensional measurement of the largest diameter, are imperfect due to the fact that the lesion volume is neither isotropic nor homogeneous. Furthermore, this approach is inherently time-consuming. Consequently, in clinical practice, monitoring patients in clinical trials in compliance with the RANO-BM criteria is rarely achieved. The objective of this study was to develop and validate an AI solution capable of delineating brain metastases (BM) on MRI to easily obtain, using an in-house solution, RANO-BM criteria as well as BM volume in a routine clinical setting.

MATERIALS (PATIENTS) AND METHODS: A total of 27456 post-Gadolinium-T1 MRI from 132 patients with BM were employed in this study. A deep learning (DL) model was constructed using the PyTorch and PyTorch Lightning frameworks, and the UNETR transfer learning method was employed to segment BM from MRI.

RESULTS: A visual analysis of the AI model results demonstrates confident delineation of the BM lesions. The model shows 100% accuracy in predicting RANO-BM criteria in comparison to that of an expert medical doctor. There was a high degree of overlap between the AI and the doctor's segmentation, with a mean DICE score of 0.77. The diameter and volume of the BM lesions were found to be concordant between the AI and the reference segmentation. The user interface developed in this study can readily provide RANO-BM criteria following AI BM segmentation.

CONCLUSION: The in-house deep learning solution is accessible to everyone without expertise in AI and offers effective BM segmentation and substantial time savings.

PMID:39800174 | DOI:10.1016/j.neuroimage.2025.121002

Categories: Literature Watch

Enhanced detection of atrial fibrillation in single-lead electrocardiograms using a cloud-based artificial intelligence platform

Sun, 2025-01-12 06:00

Heart Rhythm. 2025 Jan 10:S1547-5271(25)00019-0. doi: 10.1016/j.hrthm.2024.12.048. Online ahead of print.

ABSTRACT

BACKGROUND: Although smartphone-based devices have been developed to record 1-lead ECG, existing solutions for automatic atrial fibrillation (AF) detection often has poor positive predictive value.

OBJECTIVE: This study aimed to validate a cloud-based deep learning platform for automatic AF detection in a large cohort of patients using 1-lead ECG records.

METHODS: We analyzed 8,528 patients with 30-second ECG records from a single-lead handheld ECG device. Ground truth for AF presence was established through a benchmark algorithm and expert manual labeling. The Willem Artificial Intelligence (AI) platform, not trained on these ECGs, was used for automatic arrhythmia detection, including AF. A rules-based algorithm was also used for comparison. An expert cardiology committee reviewed false positives and negatives and performance metrics were computed.

RESULTS: The AI platform achieved an accuracy of 96.1% (initial labels) and 96.4% (expert review), with sensitivities of 83.3% and 84.2%, and specificities of 97.3% and 97.6%, respectively. The positive predictive value was 75.2% and 78.0%, and the negative predictive value was 98.4%. Performance of the AI platform largely exceeded the performance of the rules-based algorithm for all metrics. The AI also detected other arrhythmias, such as premature ventricular complexes, premature atrial complexes along with 1-degree atrioventricular blocks.

CONCLUSIONS: The result of this external validation indicates that the AI platform can match cardiologist-level accuracy in AF detection from 1-lead ECGs. Such tools are promising for AF screening and has the potential to improve accuracy in non-cardiology expert healthcare professional interpretation and trigger further tests for effective patient management.

PMID:39800092 | DOI:10.1016/j.hrthm.2024.12.048

Categories: Literature Watch

A real-time approach for surgical activity recognition and prediction based on transformer models in robot-assisted surgery

Sun, 2025-01-12 06:00

Int J Comput Assist Radiol Surg. 2025 Jan 12. doi: 10.1007/s11548-024-03306-9. Online ahead of print.

ABSTRACT

PURPOSE: This paper presents a deep learning approach to recognize and predict surgical activity in robot-assisted minimally invasive surgery (RAMIS). Our primary objective is to deploy the developed model for implementing a real-time surgical risk monitoring system within the realm of RAMIS.

METHODS: We propose a modified Transformer model with the architecture comprising no positional encoding, 5 fully connected layers, 1 encoder, and 3 decoders. This model is specifically designed to address 3 primary tasks in surgical robotics: gesture recognition, prediction, and end-effector trajectory prediction. Notably, it operates solely on kinematic data obtained from the joints of robotic arm.

RESULTS: The model's performance was evaluated on JHU-ISI Gesture and Skill Assessment Working Set dataset, achieving highest accuracy of 94.4% for gesture recognition, 84.82% for gesture prediction, and significantly low distance error of 1.34 mm with a prediction of 1 s in advance. Notably, the computational time per iteration was minimal recorded at only 4.2 ms.

CONCLUSION: The results demonstrated the excellence of our proposed model compared to previous studies highlighting its potential for integration in real-time systems. We firmly believe that our model could significantly elevate realms of surgical activity recognition and prediction within RAS and make a substantial and meaningful contribution to the healthcare sector.

PMID:39799528 | DOI:10.1007/s11548-024-03306-9

Categories: Literature Watch

DDGemb: predicting protein stability change upon single- and multi-point variations with embeddings and deep learning

Sun, 2025-01-12 06:00

Bioinformatics. 2025 Jan 12:btaf019. doi: 10.1093/bioinformatics/btaf019. Online ahead of print.

ABSTRACT

MOTIVATION: The knowledge of protein stability upon residue variation is an important step for functional protein design and for understanding how protein variants can promote disease onset. Computational methods are important to complement experimental approaches and allow a fast screening of large datasets of variations.

RESULTS: In this work we present DDGemb, a novel method combining protein language model embeddings and transformer architectures to predict protein ΔΔG upon both single- and multi-point variations. DDGemb has been trained on a high-quality dataset derived from literature and tested on available benchmark datasets of single- and multi-point variations. DDGemb performs at the state of the art in both single- and multi-point variations.

AVAILABILITY: DDGemb is available as web server at https://ddgemb.biocomp.unibo.it. Datasets used in this study are available at https://ddgemb.biocomp.unibo.it/datasets.

PMID:39799516 | DOI:10.1093/bioinformatics/btaf019

Categories: Literature Watch

Development of a model for measuring sagittal plane parameters in 10-18-year old adolescents with idiopathic scoliosis based on RTMpose deep learning technology

Sat, 2025-01-11 06:00

J Orthop Surg Res. 2025 Jan 11;20(1):41. doi: 10.1186/s13018-024-05334-2.

ABSTRACT

PURPOSE: The study aimed to develop a deep learning model for rapid, automated measurement of full-spine X-rays in adolescents with Adolescent Idiopathic Scoliosis (AIS). A significant challenge in this field is the time-consuming nature of manual measurements and the inter-individual variability in these measurements. To address these challenges, we utilized RTMpose deep learning technology to automate the process.

METHODS: We conducted a retrospective multicenter diagnostic study using 560 full-spine sagittal plane X-ray images from five hospitals in Inner Mongolia. The model was trained and validated using 500 images, with an additional 60 images for independent external validation. We evaluated the consistency of keypoint annotations among different physicians, the accuracy of model-predicted keypoints, and the accuracy of model measurement results compared to manual measurements.

RESULTS: The consistency percentages of keypoint annotations among different physicians and the model were 90-97% within the 4-mm range. The model's prediction accuracies for key points were 91-100% within the 4-mm range compared to the reference standards. The model's predictions for 15 anatomical parameters showed high consistency with experienced physicians, with intraclass correlation coefficients ranging from 0.892 to 0.991. The mean absolute error for SVA was 1.16 mm, and for other parameters, it ranged from 0.22° to 3.32°. A significant challenge we faced was the variability in data formats and specifications across different hospitals, which we addressed through data augmentation techniques. The model took an average of 9.27 s to automatically measure the 15 anatomical parameters per X-ray image.

CONCLUSION: The deep learning model based on RTMpose can effectively enhance clinical efficiency by automatically measuring the sagittal plane parameters of the spine in X-rays of patients with AIS. The model's performance was found to be highly consistent with manual measurements by experienced physicians, offering a valuable tool for clinical diagnostics.

PMID:39799363 | DOI:10.1186/s13018-024-05334-2

Categories: Literature Watch

UniAMP: enhancing AMP prediction using deep neural networks with inferred information of peptides

Sat, 2025-01-11 06:00

BMC Bioinformatics. 2025 Jan 11;26(1):10. doi: 10.1186/s12859-025-06033-3.

ABSTRACT

Antimicrobial peptides (AMPs) have been widely recognized as a promising solution to combat antimicrobial resistance of microorganisms due to the increasing abuse of antibiotics in medicine and agriculture around the globe. In this study, we propose UniAMP, a systematic prediction framework for discovering AMPs. We observe that feature vectors used in various existing studies constructed from peptide information, such as sequence, composition, and structure, can be augmented and even replaced by information inferred by deep learning models. Specifically, we use a feature vector with 2924 values inferred by two deep learning models, UniRep and ProtT5, to demonstrate that such inferred information of peptides suffice for the task, with the help of our proposed deep neural network model composed of fully connected layers and transformer encoders for predicting the antibacterial activity of peptides. Evaluation results demonstrate superior performance of our proposed model on both balanced benchmark datasets and imbalanced test datasets compared with existing studies. Subsequently, we analyze the relations among peptide sequences, manually extracted features, and automatically inferred information by deep learning models, leading to observations that the inferred information is more comprehensive and non-redundant for the task of predicting AMPs. Moreover, this approach alleviates the impact of the scarcity of positive data and demonstrates great potential in future research and applications.

PMID:39799358 | DOI:10.1186/s12859-025-06033-3

Categories: Literature Watch

Improving 3D deep learning segmentation with biophysically motivated cell synthesis

Sat, 2025-01-11 06:00

Commun Biol. 2025 Jan 11;8(1):43. doi: 10.1038/s42003-025-07469-2.

ABSTRACT

Biomedical research increasingly relies on three-dimensional (3D) cell culture models and artificial-intelligence-based analysis can potentially facilitate a detailed and accurate feature extraction on a single-cell level. However, this requires for a precise segmentation of 3D cell datasets, which in turn demands high-quality ground truth for training. Manual annotation, the gold standard for ground truth data, is too time-consuming and thus not feasible for the generation of large 3D training datasets. To address this, we present a framework for generating 3D training data, which integrates biophysical modeling for realistic cell shape and alignment. Our approach allows the in silico generation of coherent membrane and nuclei signals, that enable the training of segmentation models utilizing both channels for improved performance. Furthermore, we present a generative adversarial network (GAN) training scheme that generates not only image data but also matching labels. Quantitative evaluation shows superior performance of biophysical motivated synthetic training data, even outperforming manual annotation and pretrained models. This underscores the potential of incorporating biophysical modeling for enhancing synthetic training data quality.

PMID:39799275 | DOI:10.1038/s42003-025-07469-2

Categories: Literature Watch

Deep learning for predicting prognostic consensus molecular subtypes in cervical cancer from histology images

Sat, 2025-01-11 06:00

NPJ Precis Oncol. 2025 Jan 11;9(1):11. doi: 10.1038/s41698-024-00778-5.

ABSTRACT

Cervical cancer remains the fourth most common cancer among women worldwide. This study proposes an end-to-end deep learning framework to predict consensus molecular subtypes (CMS) in HPV-positive cervical squamous cell carcinoma (CSCC) from H&E-stained histology slides. Analysing three CSCC cohorts (n = 545), we show our Digital-CMS scores significantly stratify patients by both disease-specific (TCGA p = 0.0022, Oslo p = 0.0495) and disease-free (TCGA p = 0.0495, Oslo p = 0.0282) survival. In addition, our extensive tumour microenvironment analysis reveals differences between the two CMS subtypes, with CMS-C1 tumours exhibit increased lymphocyte presence, while CMS-C2 tumours show high nuclear pleomorphism, elevated neutrophil-to-lymphocyte ratio, and higher malignancy, correlating with poor prognosis. This study introduces a potentially clinically advantageous Digital-CMS score derived from digitised WSIs of routine H&E-stained tissue sections, offers new insights into TME differences impacting patient prognosis and potential therapeutic targets, and identifies histological patterns serving as potential surrogate markers of the CMS subtypes for clinical application.

PMID:39799271 | DOI:10.1038/s41698-024-00778-5

Categories: Literature Watch

Unsupervised deep learning of electrocardiograms enables scalable human disease profiling

Sat, 2025-01-11 06:00

NPJ Digit Med. 2025 Jan 12;8(1):23. doi: 10.1038/s41746-024-01418-9.

ABSTRACT

The 12-lead electrocardiogram (ECG) is inexpensive and widely available. Whether conditions across the human disease landscape can be detected using the ECG is unclear. We developed a deep learning denoising autoencoder and systematically evaluated associations between ECG encodings and ~1,600 Phecode-based diseases in three datasets separate from model development, and meta-analyzed the results. The latent space ECG model identified associations with 645 prevalent and 606 incident Phecodes. Associations were most enriched in the circulatory (n = 140, 82% of category-specific Phecodes), respiratory (n = 53, 62%) and endocrine/metabolic (n = 73, 45%) categories, with additional associations across the phenome. The strongest ECG association was with hypertension (p < 2.2×10-308). The ECG latent space model demonstrated more associations than models using standard ECG intervals, and offered favorable discrimination of prevalent disease compared to models comprising age, sex, and race. We further demonstrate how latent space models can be used to generate disease-specific ECG waveforms and facilitate individual disease profiling.

PMID:39799251 | DOI:10.1038/s41746-024-01418-9

Categories: Literature Watch

Improving spleen segmentation in ultrasound images using a hybrid deep learning framework

Sat, 2025-01-11 06:00

Sci Rep. 2025 Jan 11;15(1):1670. doi: 10.1038/s41598-025-85632-9.

ABSTRACT

This paper introduces a novel method for spleen segmentation in ultrasound images, using a two-phase training approach. In the first phase, the SegFormerB0 network is trained to provide an initial segmentation. In the second phase, the network is further refined using the Pix2Pix structure, which enhances attention to details and corrects any erroneous or additional segments in the output. This hybrid method effectively combines the strengths of both SegFormer and Pix2Pix to produce highly accurate segmentation results. We have assembled the Spleenex dataset, consisting of 450 ultrasound images of the spleen, which is the first dataset of its kind in this field. Our method has been validated on this dataset, and the experimental results show that it outperforms existing state-of-the-art models. Specifically, our approach achieved a mean Intersection over Union (mIoU) of 94.17% and a mean Dice (mDice) score of 96.82%, surpassing models such as Splenomegaly Segmentation Network (SSNet), U-Net, and Variational autoencoder based methods. The proposed method also achieved a Mean Percentage Length Error (MPLE) of 3.64%, further demonstrating its accuracy. Furthermore, the proposed method has demonstrated strong performance even in the presence of noise in ultrasound images, highlighting its practical applicability in clinical environments.

PMID:39799236 | DOI:10.1038/s41598-025-85632-9

Categories: Literature Watch

A benchmark of deep learning approaches to predict lung cancer risk using national lung screening trial cohort

Sat, 2025-01-11 06:00

Sci Rep. 2025 Jan 11;15(1):1736. doi: 10.1038/s41598-024-84193-7.

ABSTRACT

Deep learning (DL) methods have demonstrated remarkable effectiveness in assisting with lung cancer risk prediction tasks using computed tomography (CT) scans. However, the lack of comprehensive comparison and validation of state-of-the-art (SOTA) models in practical settings limits their clinical application. This study aims to review and analyze current SOTA deep learning models for lung cancer risk prediction (malignant-benign classification). To evaluate our model's general performance, we selected 253 out of 467 patients from a subset of the National Lung Screening Trial (NLST) who had CT scans without contrast, which are the most commonly used, and divided them into training and test cohorts. The CT scans were preprocessed into 2D-image and 3D-volume formats according to their nodule annotations. We evaluated ten 3D and eleven 2D SOTA deep learning models, which were pretrained on large-scale general-purpose datasets (Kinetics and ImageNet) and radiological datasets (3DSeg-8, nnUnet and RadImageNet), for their lung cancer risk prediction performance. Our results showed that 3D-based deep learning models generally perform better than 2D models. On the test cohort, the best-performing 3D model achieved an AUROC of 0.86, while the best 2D model reached 0.79. The lowest AUROCs for the 3D and 2D models were 0.70 and 0.62, respectively. Furthermore, pretraining on large-scale radiological image datasets did not show the expected performance advantage over pretraining on general-purpose datasets. Both 2D and 3D deep learning models can handle lung cancer risk prediction tasks effectively, although 3D models generally have superior performance than their 2D competitors. Our findings highlight the importance of carefully selecting pretrained datasets and model architectures for lung cancer risk prediction. Overall, these results have important implications for the development and clinical integration of DL-based tools in lung cancer screening.

PMID:39799226 | DOI:10.1038/s41598-024-84193-7

Categories: Literature Watch

Signature-based intrusion detection using machine learning and deep learning approaches empowered with fuzzy clustering

Sat, 2025-01-11 06:00

Sci Rep. 2025 Jan 11;15(1):1726. doi: 10.1038/s41598-025-85866-7.

ABSTRACT

Network security is crucial in today's digital world, since there are multiple ongoing threats to sensitive data and vital infrastructure. The aim of this study to improve network security by combining methods for instruction detection from machine learning (ML) and deep learning (DL). Attackers have tried to breach security systems by accessing networks and obtaining sensitive information.Intrusion detection systems (IDSs) are one of the significant aspect of cybersecurity that involve the monitoring and analysis, with the intention of identifying and reporting of dangerous activities that would help to prevent the attack.Support Vector Machine (SVM), K-Nearest Neighbors (KNN), Random Forest (RF), Decision Tree (DT), Long Short-Term Memory (LSTM), and Artificial Neural Network (ANN) are the vector figures incorporated into the study through the results. These models are subjected to various test to established the best results on the identification and prevention of network violation. Based on the obtained results, it can be stated that all the tested models are capable of organizing data originating from network traffic. thus, recognizing the difference between normal and intrusive behaviors, models such as SVM, KNN, RF, and DT showed effective results. Deep learning models LSTM and ANN rapidly find long-term and complex pattern in network data. It is extremely effective when dealing with complex intrusions since it is characterised by high precision, accuracy and recall.Based on our study, SVM and Random Forest are considered promising solutions for real-world IDS applications because of their versatility and explainability. For the companies seeking IDS solutions which are reliable and at the same time more interpretable, these models can be promising. Additionally, LSTM and ANN, with their ability to catch successive conditions, are suitable for situations involving nuanced, advancing dangers.

PMID:39799225 | DOI:10.1038/s41598-025-85866-7

Categories: Literature Watch

Importance of neural network complexity for the automatic segmentation of individual thigh muscles in MRI images from patients with neuromuscular diseases

Sat, 2025-01-11 06:00

MAGMA. 2025 Jan 11. doi: 10.1007/s10334-024-01221-3. Online ahead of print.

ABSTRACT

OBJECTIVE: Segmentation of individual thigh muscles in MRI images is essential for monitoring neuromuscular diseases and quantifying relevant biomarkers such as fat fraction (FF). Deep learning approaches such as U-Net have demonstrated effectiveness in this field. However, the impact of reducing neural network complexity remains unexplored in the FF quantification in individual muscles.

MATERIAL AND METHODS: U-Net architectures with different complexities have been compared for the quantification of the fat fraction in each muscle group selected in the central part of the thigh region. The corresponding performance has been assessed in terms of Dice score (DSC) and FF quantification error. The database contained 1450 thigh images from 59 patients and 14 healthy subjects (age: 47 ± 17 years, sex: 36F, 37M). Ten individual muscles were segmented in each image. The performance of each model was compared to nnU-Net, a complex architecture with 4.35 × 107 parameters, 12.8 Gigabytes of peak memory usage and 167 h of training time.

RESULTS: As expected, nnU-Net achieved the highest DSC (94.77 ± 0.13%). A simpler U-Net (5.81 × 105 parameters, 2.37 Gigabytes, 14 h of training time) achieved a lower DSC but still above 90%. Surprisingly, both models achieved a comparable FF estimation.

DISCUSSION: The poor correlation between observed DSC and FF indicates that less complex architectures, reducing GPU memory utilization and training time, can still accurately quantify FF.

PMID:39798067 | DOI:10.1007/s10334-024-01221-3

Categories: Literature Watch

Deep learning multi-classification of middle ear diseases using synthetic tympanic images

Sat, 2025-01-11 06:00

Acta Otolaryngol. 2025 Jan 10:1-6. doi: 10.1080/00016489.2024.2448829. Online ahead of print.

ABSTRACT

BACKGROUND: Recent advances in artificial intelligence have facilitated the automatic diagnosis of middle ear diseases using endoscopic tympanic membrane imaging.

AIM: We aimed to develop an automated diagnostic system for middle ear diseases by applying deep learning techniques to tympanic membrane images obtained during routine clinical practice.

MATERIAL AND METHODS: To augment the training dataset, we explored the use of generative adversarial networks (GANs) to produce high-quality synthetic tympanic images that were subsequently added to the training data. Between 2016 and 2021, we collected 472 endoscopic images representing four tympanic membrane conditions: normal, acute otitis media, otitis media with effusion, and chronic suppurative otitis media. These images were utilized for machine learning based on the InceptionV3 model, which was pretrained on ImageNet. Additionally, 200 synthetic images generated using StyleGAN3 and considered appropriate for each disease category were incorporated for retraining.

RESULTS: The inclusion of synthetic images alongside real endoscopic images did not significantly improve the diagnostic accuracy compared to training solely with real images. However, when trained solely on synthetic images, the model achieved a diagnostic accuracy of approximately 70%.

CONCLUSIONS AND SIGNIFICANCE: Synthetic images generated by GANs have potential utility in the development of machine-learning models for medical diagnosis.

PMID:39797517 | DOI:10.1080/00016489.2024.2448829

Categories: Literature Watch

Integrating Model-Informed Drug Development With AI: A Synergistic Approach to Accelerating Pharmaceutical Innovation

Sat, 2025-01-11 06:00

Clin Transl Sci. 2025 Jan;18(1):e70124. doi: 10.1111/cts.70124.

ABSTRACT

The pharmaceutical industry constantly strives to improve drug development processes to reduce costs, increase efficiencies, and enhance therapeutic outcomes for patients. Model-Informed Drug Development (MIDD) uses mathematical models to simulate intricate processes involved in drug absorption, distribution, metabolism, and excretion, as well as pharmacokinetics and pharmacodynamics. Artificial intelligence (AI), encompassing techniques such as machine learning, deep learning, and Generative AI, offers powerful tools and algorithms to efficiently identify meaningful patterns, correlations, and drug-target interactions from big data, enabling more accurate predictions and novel hypothesis generation. The union of MIDD with AI enables pharmaceutical researchers to optimize drug candidate selection, dosage regimens, and treatment strategies through virtual trials to help derisk drug candidates. However, several challenges, including the availability of relevant, labeled, high-quality datasets, data privacy concerns, model interpretability, and algorithmic bias, must be carefully managed. Standardization of model architectures, data formats, and validation processes is imperative to ensure reliable and reproducible results. Moreover, regulatory agencies have recognized the need to adapt their guidelines to evaluate recommendations from AI-enhanced MIDD methods. In conclusion, integrating model-driven drug development with AI offers a transformative paradigm for pharmaceutical innovation. By integrating the predictive power of computational models and the data-driven insights of AI, the synergy between these approaches has the potential to accelerate drug discovery, optimize treatment strategies, and usher in a new era of personalized medicine, benefiting patients, researchers, and the pharmaceutical industry as a whole.

PMID:39797502 | DOI:10.1111/cts.70124

Categories: Literature Watch

Pages