Deep learning
Computational modeling of tumor invasion from limited and diverse data in Glioblastoma
Comput Med Imaging Graph. 2024 Sep 23;117:102436. doi: 10.1016/j.compmedimag.2024.102436. Online ahead of print.
ABSTRACT
For diseases with high morbidity rates such as Glioblastoma Multiforme, the prognostic and treatment planning pipeline requires a comprehensive analysis of imaging, clinical, and molecular data. Many mutations have been shown to correlate strongly with the median survival rate and response to therapy of patients. Studies have demonstrated that these mutations manifest as specific visual biomarkers in tumor imaging modalities such as MRI. To minimize the number of invasive procedures on a patient and for the overall resource optimization for the prognostic and treatment planning process, the correlation of imaging and molecular features has garnered much interest. While the tumor mass is the most significant feature, the impacted tissue surrounding the tumor is also a significant biomarker contributing to the visual manifestation of mutations - which has not been studied as extensively. The pattern of tumor growth impacts the surrounding tissue accordingly, which is a reflection of tumor properties as well. Modeling how the tumor growth impacts the surrounding tissue can reveal important information about the patterns of tumor enhancement, which in turn has significant diagnostic and prognostic value. This paper presents the first work to automate the computational modeling of the impacted tissue surrounding the tumor using generative deep learning. The paper isolates and quantifies the impact of the Tumor Invasion (TI) on surrounding tissue based on change in mutation status, subsequently assessing its prognostic value. Furthermore, a TI Generative Adversarial Network (TI-GAN) is proposed to model the tumor invasion properties. Extensive qualitative and quantitative analyses, cross-dataset testing, and radiologist blind tests are carried out to demonstrate that TI-GAN can realistically model the tumor invasion under practical challenges of medical datasets such as limited data and high intra-class heterogeneity.
PMID:39342741 | DOI:10.1016/j.compmedimag.2024.102436
A multi-verse optimizer-based CNN-BiLSTM pixel-level detection model for peanut aflatoxins
Food Chem. 2024 Sep 24;463(Pt 3):141393. doi: 10.1016/j.foodchem.2024.141393. Online ahead of print.
ABSTRACT
Peanuts are highly susceptible to contamination by aflatoxins, posing a significant threat to human health. This study aims to enhance the accuracy of pixel-level aflatoxin detection in hyperspectral images using an optimized deep learning method. This study developed a CNN-BiLSTM fusion model optimized by the Multi-Verse Optimizer (MVO) algorithm, specifically designed to detect aflatoxins with high precision. The optimized CNN-BiLSTM model was fine-tuned using aflatoxin spectral data at varying concentrations. The results indicate that the fine-tuned MVO-CNN-BiLSTM model achieved the best performance, with a validation accuracy of 94.92 % and a recall rate of 95.59 %. The accuracy of this model is 6.93 % and 3.6 % higher than machine learning methods such as SVM and AdaBoost, respectively. Additionally, it is 4.18 % and 3.08 % higher than deep learning methods such as CNN and the CNN-LSTM fusion model, respectively. This method enhances pixel-level aflatoxin detection accuracy, supporting the development of online detection devices.
PMID:39342735 | DOI:10.1016/j.foodchem.2024.141393
BOATMAP: Bayesian Optimization Active Targeting for Monomorphic Arrhythmia Pace-mapping
Comput Biol Med. 2024 Sep 28;182:109201. doi: 10.1016/j.compbiomed.2024.109201. Online ahead of print.
ABSTRACT
Recent advances in machine learning and deep learning have presented new opportunities for learning to localize the origin of ventricular activation from 12-lead electrocardiograms (ECGs), an important step in guiding ablation therapies for ventricular tachycardia. Passively learning from population data is faced with challenges due to significant variations among subjects, and building a patient-specific model raises the open question of where to select pace-mapping data for training. This work introduces BOATMAP, a novel active learning approach designed to provide clinicians with interpretable guidance that progressively assists in locating the origin of ventricular activation from 12-lead ECGs. BOATMAP inverts the input-output relationship in traditional machine learning solutions to this problem and learns the similarity between a target ECG and a paced ECG as a function of the pacing site coordinates. Using Gaussian processes (GP) as a surrogate model, BOATMAP iteratively refines the estimated similarity landscape while providing suggestions to clinicians regarding the next optimal pacing site. Furthermore, it can incorporate constraints to avoid suggesting pacing in non-viable regions such as the core of the myocardial scar. Tested in a realistic simulation environment in various heart geometries and tissue properties, BOATMAP demonstrated the ability to accurately localize the origin of activation, achieving an average localization accuracy of 3.9±3.6mm with only 8.0±4.0 pacing sites. BOATMAP offers real-time interpretable guidance for accurate localization and enhancing clinical decision-making.
PMID:39342676 | DOI:10.1016/j.compbiomed.2024.109201
Construction of machine learning diagnostic models for cardiovascular pan-disease based on blood routine and biochemical detection data
Cardiovasc Diabetol. 2024 Sep 28;23(1):351. doi: 10.1186/s12933-024-02439-0.
ABSTRACT
BACKGROUND: Cardiovascular disease, also known as circulation system disease, remains the leading cause of morbidity and mortality worldwide. Traditional methods for diagnosing cardiovascular disease are often expensive and time-consuming. So the purpose of this study is to construct machine learning models for the diagnosis of cardiovascular diseases using easily accessible blood routine and biochemical detection data and explore the unique hematologic features of cardiovascular diseases, including some metabolic indicators.
METHODS: After the data preprocessing, 25,794 healthy people and 32,822 circulation system disease patients with the blood routine and biochemical detection data were utilized for our study. We selected logistic regression, random forest, support vector machine, eXtreme Gradient Boosting (XGBoost), and deep neural network to construct models. Finally, the SHAP algorithm was used to interpret models.
RESULTS: The circulation system disease prediction model constructed by XGBoost possessed the best performance (AUC: 0.9921 (0.9911-0.9930); Acc: 0.9618 (0.9588-0.9645); Sn: 0.9690 (0.9655-0.9723); Sp: 0.9526 (0.9477-0.9572); PPV: 0.9631 (0.9592-0.9668); NPV: 0.9600 (0.9556-0.9644); MCC: 0.9224 (0.9165-0.9279); F1 score: 0.9661 (0.9634-0.9686)). Most models of distinguishing various circulation system diseases also had good performance, the model performance of distinguishing dilated cardiomyopathy from other circulation system diseases was the best (AUC: 0.9267 (0.8663-0.9752)). The model interpretation by the SHAP algorithm indicated features from biochemical detection made major contributions to predicting circulation system disease, such as potassium (K), total protein (TP), albumin (ALB), and indirect bilirubin (NBIL). But for models of distinguishing various circulation system diseases, we found that red blood cell count (RBC), K, direct bilirubin (DBIL), and glucose (GLU) were the top 4 features subdividing various circulation system diseases.
CONCLUSIONS: The present study constructed multiple models using 50 features from the blood routine and biochemical detection data for the diagnosis of various circulation system diseases. At the same time, the unique hematologic features of various circulation system diseases, including some metabolic-related indicators, were also explored. This cost-effective work will benefit more people and help diagnose and prevent circulation system diseases.
PMID:39342281 | DOI:10.1186/s12933-024-02439-0
Dual stage MRI image restoration based on blind spot denoising and hybrid attention
BMC Med Imaging. 2024 Sep 28;24(1):259. doi: 10.1186/s12880-024-01437-8.
ABSTRACT
BACKGROUND: Magnetic Resonance Imaging (MRI) is extensively utilized in clinical diagnostics and medical research, yet the imaging process is often compromised by noise interference. This noise arises from various sources, leading to a reduction in image quality and subsequently hindering the accurate interpretation of image details by clinicians. Traditional denoising methods typically assume that noise follows a Gaussian distribution, thereby neglecting the more complex noise types present in MRI images, such as Rician noise. As a result, denoising remains a challenging and practical task.
METHOD: The main research work of this paper focuses on modifying mask information based on a global mask mapper. The mask mapper samples all blind spot pixels on the denoised image and maps them to the same channel. By incorporating perceptual loss, it utilizes all available information to improve performance while avoiding identity mapping. During the denoising process, the model may mistakenly remove some useful information as noise, resulting in a loss of detail in the denoised image. To address this issue, we train a generative adversarial network (GAN) with adaptive hybrid attention to restore the detailed information in the denoised MRI images.
RESULT: The two-stage model NRAE shows an improvement of nearly 1.4 dB in PSNR and approximately 0.1 in SSIM on clinical datasets compared to other classic models. Specifically, compared to the baseline model, PSNR is increased by about 0.6 dB, and SSIM is only 0.015 lower. From a visual perspective, NRAE more effectively restores the details in the images, resulting in richer and clearer representation of image details.
CONCLUSION: We have developed a deep learning-based two-stage model to address noise issues in medical MRI images. This method not only successfully reduces noise signals but also effectively restores anatomical details. The current results indicate that this is a promising approach. In future work, we plan to replace the current denoising network with more advanced models to further enhance performance.
PMID:39342222 | DOI:10.1186/s12880-024-01437-8
FindCSV: a long-read based method for detecting complex structural variations
BMC Bioinformatics. 2024 Sep 28;25(1):315. doi: 10.1186/s12859-024-05937-w.
ABSTRACT
BACKGROUND: Structural variations play a significant role in genetic diseases and evolutionary mechanisms. Extensive research has been conducted over the past decade to detect simple structural variations, leading to the development of well-established detection methods. However, recent studies have highlighted the potentially greater impact of complex structural variations on individuals compared to simple structural variations. Despite this, the field still lacks precise detection methods specifically designed for complex structural variations. Therefore, the development of a highly efficient and accurate detection method is of utmost importance.
RESULT: In response to this need, we propose a novel method called FindCSV, which leverages deep learning techniques and consensus sequences to enhance the detection of SVs using long-read sequencing data. Compared to current methods, FindCSV performs better in detecting complex and simple structural variations.
CONCLUSIONS: FindCSV is a new method to detect complex and simple structural variations with reasonable accuracy in real and simulated data. The source code for the program is available at https://github.com/nwpuzhengyan/FindCSV .
PMID:39342151 | DOI:10.1186/s12859-024-05937-w
Prediction of short-term adverse clinical outcomes of acute pulmonary embolism using conventional machine learning and deep Learning based on CTPA images
J Thromb Thrombolysis. 2024 Sep 28. doi: 10.1007/s11239-024-03044-4. Online ahead of print.
ABSTRACT
To explore the predictive value of traditional machine learning (ML) and deep learning (DL) algorithms based on computed tomography pulmonary angiography (CTPA) images for short-term adverse outcomes in patients with acute pulmonary embolism (APE). This retrospective study enrolled 132 patients with APE confirmed by CTPA. Thrombus segmentation and texture feature extraction was performed using 3D-Slicer software. The least absolute shrinkage and selection operator (LASSO) algorithm was used for feature dimensionality reduction and selection, with optimal λ values determined using leave-one-fold cross-validation to identify texture features with non-zero coefficients. ML models (logistic regression, random forest, decision tree, support vector machine) and DL models (ResNet 50 and Vgg 19) were used to construct the prediction models. Model performance was evaluated using receiver operating characteristic (ROC) curves and the area under the curve (AUC). The cohort included 84 patients in the good prognosis group and 48 patients in the poor prognosis group. Univariate and multivariate logistic regression analyses showed that diabetes, RV/LV ≥ 1.0, and Qanadli index form independent risk factors predicting poor prognosis in patients with APE(P < 0.05). A total of 750 texture features were extracted, with 4 key features identified through screening. There was a weak positive correlation between texture features and clinical parameters. ROC curves analysis demonstrated AUC values of 0.85 (0.78-0.92), 0.76 (0.67-0.84), and 0.89 (0.83-0.95) for the clinical, texture feature, and combined models, respectively. In the ML models, the random forest model achieved the highest AUC (0.85), and the support vector machine model achieved the lowest AUC (0.62). And the AUCs for the DL models (ResNet 50 and Vgg 19) were 0.91 (95%CI: 0.90-0.92) and 0.94(95%CI: 0.93-0.95), respectively. Vgg 19 model demonstrated exceptional precision (0.93), recall (0.76), specificity (0.95) and F1 score (0.84). Both ML and DL models based on thrombus texture features from CTPA images demonstrated higher predictive efficacy for short-term adverse outcomes in patients with APE, especially the random forest and Vgg 19 models, potentially assisting clinical management in timely interventions to improve patient prognosis.
PMID:39342072 | DOI:10.1007/s11239-024-03044-4
Enhanced gastric cancer classification and quantification interpretable framework using digital histopathology images
Sci Rep. 2024 Sep 28;14(1):22533. doi: 10.1038/s41598-024-73823-9.
ABSTRACT
Recent developments have highlighted the critical role that computer-aided diagnosis (CAD) systems play in analyzing whole-slide digital histopathology images for detecting gastric cancer (GC). We present a novel framework for gastric histology classification and segmentation (GHCS) that offers modest yet meaningful improvements over existing CAD models for GC classification and segmentation. Our methodology achieves marginal improvements over conventional deep learning (DL) and machine learning (ML) models by adaptively focusing on pertinent characteristics of images. This contributes significantly to our study, highlighting that the proposed model, which performs well on normalized images, is robust in certain respects, particularly in handling variability and generalizing to different datasets. We anticipate that this robustness will lead to better results across various datasets. An expectation-maximizing Naïve Bayes classifier that uses an updated Gaussian Mixture Model is at the heart of the suggested GHCS framework. The effectiveness of our classifier is demonstrated by experimental validation on two publicly available datasets, which produced exceptional classification accuracies of 98.87% and 97.28% on validation sets and 98.47% and 97.31% on test sets. Our framework shows a slight but consistent improvement over previously existing techniques in gastric histopathology image classification tasks, as demonstrated by comparative analysis. This may be attributed to its ability to capture critical features of gastric histopathology images better. Furthermore, using an improved Fuzzy c-means method, our study produces good results in GC histopathology picture segmentation, outperforming state-of-the-art segmentation models with a Dice coefficient of 65.21% and a Jaccard index of 60.24%. The model's interpretability is complemented by Grad-CAM visualizations, which help understand the decision-making process and increase the model's trustworthiness for end-users, especially clinicians.
PMID:39342030 | DOI:10.1038/s41598-024-73823-9
NA-segformer: A multi-level transformer model based on neighborhood attention for colonoscopic polyp segmentation
Sci Rep. 2024 Sep 28;14(1):22527. doi: 10.1038/s41598-024-74123-y.
ABSTRACT
In various countries worldwide, the incidence of colon cancer-related deaths has been on the rise in recent years. Early detection of symptoms and identification of intestinal polyps are crucial for improving the cure rate of colon cancer patients. Automated computer-aided diagnosis (CAD) has emerged as a solution to the low efficiency of traditional methods relying on manual diagnosis by physicians. Deep learning is the latest direction of CAD development and has shown promise for colonoscopic polyp segmentation. In this paper, we present a multi-level encoder-decoder architecture for polyp segmentation based on the Transformer architecture, termed NA-SegFormer. To improve the performance of existing Transformer-based segmentation algorithms for edge segmentation on colon polyps, we propose a patch merging module with a neighbor attention mechanism based on overlap patch merging. Since colon tract polyps vary greatly in size and different datasets have different sample sizes, we used a unified focal loss to solve the problem of category imbalance in colon tract polyp data. To assess the effectiveness of our proposed method, we utilized video capsule endoscopy and typical colonoscopy polyp datasets, as well as a dataset containing surgical equipment. On the datasets Kvasir-SEG, Kvasir-Instrument and KvasirCapsule-SEG, the Dice score of our proposed model reached 94.30%, 94.59% and 82.73%, with an accuracy of 98.26%, 99.02% and 81.84% respectively. The proposed method achieved inference speed with an Frame-per-second (FPS) of 125.01. The results demonstrated that our suggested model effectively segmented polyps better than several well-known and latest models. In addition, the proposed method has advantages in trade-off between inference speed and accuracy, and it will be of great significance to real-time colonoscopic polyp segmentation. The code is available at https://github.com/promisedong/NAFormer .
PMID:39342011 | DOI:10.1038/s41598-024-74123-y
Adaptive haze pixel intensity perception transformer structure for image dehazing networks
Sci Rep. 2024 Sep 28;14(1):22435. doi: 10.1038/s41598-024-73866-y.
ABSTRACT
In the realm of deep learning-based networks for dehazing using paired clean-hazy image datasets to address complex real-world haze scenarios in daytime environments and cross-dataset challenges remains a significant concern due to algorithmic inefficiencies and color distortion. To tackle these issues, we propose SwinTieredHazymers (STH), a dehazing network designed to adaptively discern pixel intensities in hazy images and compute haze residue for clarity restoration. Through a unique three-branch design, we hierarchically modulate haze residuals by leveraging the global features brought by Transformer and the local features brought by Convolutional Neural Network (CNN) which has led to the algorithm's widespread applicability. Experimental results demonstrate that our approach surpasses advanced single-image dehazing methods in both quantitative metrics and visual fidelity for real-world hazy image dehazing, while also exhibiting strong performance in cross-dataset dehazing scenarios.
PMID:39341993 | DOI:10.1038/s41598-024-73866-y
Deep learning for identifying personal and family history of suicidal thoughts and behaviors from EHRs
NPJ Digit Med. 2024 Sep 28;7(1):260. doi: 10.1038/s41746-024-01266-7.
ABSTRACT
Personal and family history of suicidal thoughts and behaviors (PSH and FSH, respectively) are significant risk factors associated with suicides. Research is limited in automatic identification of such data from clinical notes in Electronic Health Records. This study developed deep learning (DL) tools utilizing transformer models (Bio_ClinicalBERT and GatorTron) to detect PSH and FSH in clinical notes derived from three academic medical centers, and compared their performance with a rule-based natural language processing tool. For detecting PSH, the rule-based approach obtained an F1-score of 0.75 ± 0.07, while the Bio_ClinicalBERT and GatorTron DL tools scored 0.83 ± 0.09 and 0.84 ± 0.07, respectively. For detecting FSH, the rule-based approach achieved an F1-score of 0.69 ± 0.11, compared to 0.89 ± 0.10 for Bio_ClinicalBERT and 0.92 ± 0.07 for GatorTron. Across sites, the DL tools identified more than 80% of patients at elevated risk for suicide who remain undiagnosed and untreated.
PMID:39341983 | DOI:10.1038/s41746-024-01266-7
Combining 2.5D deep learning and conventional features in a joint model for the early detection of sICH expansion
Sci Rep. 2024 Sep 28;14(1):22467. doi: 10.1038/s41598-024-73415-7.
ABSTRACT
The study aims to investigate the potential of training efficient deep learning models by using 2.5D (2.5-Dimension) masks of sICH. Furthermore, it intends to evaluate and compare the predictive performance of a joint model incorporating four types of features with standalone 2.5D deep learning, radiomics, radiology, and clinical models for early expansion in sICH. A total of 254 sICH patients were enrolled retrospectively and divided into two groups according to whether the hematoma was enlarged or not. The 2.5D mask of sICH is constructed with the maximum axial, coronal and sagittal planes of the hematoma, which is used to train the deep learning model and extract deep learning features. Predictive models were built on clinic, radiology, radiomics and deep learning features separately and four type features jointly. The diagnostic performance of each model was measured using the receiver operating characteristic curve (AUC), Accuracy, Recall, F1 and decision curve analysis (DCA). The AUCs of the clinic model, radiology model, radiomics model, deep learning model, joint model, and nomogram model on the train set (training and Cross-validation) were 0.639, 0.682, 0.859, 0.807, 0.939, and 0.942, respectively, while the AUCs on the test set (external validation) were 0.680, 0.758, 0.802, 0.857, 0.929, and 0.926. Decision curve analysis showed that the joint model was superior to the other models and demonstrated good consistency between the predicted probability of early hematoma expansion and the actual occurrence probability. Our study demonstrates that the joint model is a more efficient and robust prediction model, as verified by multicenter data. This finding highlights the potential clinical utility of a multifactorial prediction model that integrates various data sources for prognostication in patients with intracerebral hemorrhage. The Critical Relevance Statement: Combining 2.5D deep learning features with clinic features, radiology markers, and radiomics signatures to establish a joint model enabling physicians to conduct better-individualized assessments the risk of early expansion of sICH.
PMID:39341957 | DOI:10.1038/s41598-024-73415-7
Real-time segmentation of biliary structure in pure laparoscopic donor hepatectomy
Sci Rep. 2024 Sep 28;14(1):22508. doi: 10.1038/s41598-024-73434-4.
ABSTRACT
Pure laparoscopic donor hepatectomy (PLDH) has become a standard practice for living donor liver transplantation in expert centers. Accurate understanding of biliary structures is crucial during PLDH to minimize the risk of complications. This study aims to develop a deep learning-based segmentation model for real-time identification of biliary structures, assisting surgeons in determining the optimal transection site during PLDH. A single-institution retrospective feasibility analysis was conducted on 30 intraoperative videos of PLDH. All videos were selected for their use of the indocyanine green near-infrared fluorescence technique to identify biliary structure. From the analysis, 10 representative frames were extracted from each video specifically during the bile duct division phase, resulting in 300 frames. These frames underwent pixel-wise annotation to identify biliary structures and the transection site. A segmentation task was then performed using a DeepLabV3+ algorithm, equipped with a ResNet50 encoder, focusing on the bile duct (BD) and anterior wall (AW) for transection. The model's performance was evaluated using the dice similarity coefficient (DSC). The model predicted biliary structures with a mean DSC of 0.728 ± 0.01 for BD and 0.429 ± 0.06 for AW. Inference was performed at a speed of 15.3 frames per second, demonstrating the feasibility of real-time recognition of anatomical structures during surgery. The deep learning-based semantic segmentation model exhibited promising performance in identifying biliary structures during PLDH. Future studies should focus on validating the clinical utility and generalizability of the model and comparing its efficacy with current gold standard practices to better evaluate its potential clinical applications.
PMID:39341910 | DOI:10.1038/s41598-024-73434-4
Effectiveness of data-augmentation on deep learning in evaluating rapid on-site cytopathology at endoscopic ultrasound-guided fine needle aspiration
Sci Rep. 2024 Sep 28;14(1):22441. doi: 10.1038/s41598-024-72312-3.
ABSTRACT
Rapid on-site cytopathology evaluation (ROSE) has been considered an effective method to increase the diagnostic ability of endoscopic ultrasound-guided fine needle aspiration (EUS-FNA); however, ROSE is unavailable in most institutes worldwide due to the shortage of cytopathologists. To overcome this situation, we created an artificial intelligence (AI)-based system (the ROSE-AI system), which was trained with the augmented data to evaluate the slide images acquired by EUS-FNA. This study aimed to clarify the effects of such data-augmentation on establishing an effective ROSE-AI system by comparing the efficacy of various data-augmentation techniques. The ROSE-AI system was trained with increased data obtained by the various data-augmentation techniques, including geometric transformation, color space transformation, and kernel filtering. By performing five-fold cross-validation, we compared the efficacy of each data-augmentation technique on the increasing diagnostic abilities of the ROSE-AI system. We collected 4059 divided EUS-FNA slide images from 36 patients with pancreatic cancer and nine patients with non-pancreatic cancer. The diagnostic ability of the ROSE-AI system without data augmentation had a sensitivity, specificity, and accuracy of 87.5%, 79.7%, and 83.7%, respectively. While, some data-augmentation techniques decreased diagnostic ability, the ROSE-AI system trained only with the augmented data using the geometric transformation technique had the highest diagnostic accuracy (88.2%). We successfully developed a prototype ROSE-AI system with high diagnostic ability. Each data-augmentation technique may have various compatibilities with AI-mediated diagnostics, and the geometric transformation was the most effective for the ROSE-AI system.
PMID:39341885 | DOI:10.1038/s41598-024-72312-3
Enhancing brain tumor classification through ensemble attention mechanism
Sci Rep. 2024 Sep 27;14(1):22260. doi: 10.1038/s41598-024-73803-z.
ABSTRACT
Brain tumors pose a serious threat to public health, impacting thousands of individuals directly or indirectly worldwide. Timely and accurate detection of these tumors is crucial for effective treatment and enhancing the quality of patients' lives. The widely used brain imaging technique is magnetic resonance imaging, the precise identification of brain tumors in MRI images is challenging due to the diverse anatomical structures. This paper introduces an innovative approach known as the ensemble attention mechanism to address this challenge. Initially, the approach uses two networks to extract intermediate- and final-level feature maps from MobileNetV3 and EfficientNetB7. This assists in gathering the relevant feature maps from the different models at different levels. Then, the technique incorporates a co-attention mechanism into the intermediate and final feature map levels on both networks and ensembles them. This directs attention to certain regions to extract global-level features at different levels. Ensemble of attentive feature maps enabling the precise detection of various feature patterns within brain tumor images at both model, local, and global levels. This leads to an improvement in the classification process. The proposed system was evaluated on the Figshare dataset and achieved an accuracy of 98.94%, and 98.48% for the BraTS 2019 dataset which is superior to other methods. Thus, it is robust and suitable for brain tumor detection in healthcare systems.
PMID:39333699 | DOI:10.1038/s41598-024-73803-z
Explainability of CNN-based Alzheimer's disease detection from online handwriting
Sci Rep. 2024 Sep 27;14(1):22108. doi: 10.1038/s41598-024-72650-2.
ABSTRACT
With over 55 million people globally affected by dementia and nearly 10 million new cases reported annually, Alzheimer's disease is a prevalent and challenging neurodegenerative disorder. Despite significant advancements in machine learning techniques for Alzheimer's disease detection, the widespread adoption of deep learning models raises concerns about their explainability. The lack of explainability in deep learning models for online handwriting analysis is a critical gap in the literature in the context of Alzheimer's disease detection. This paper addresses this challenge by interpreting predictions from a Convolutional Neural Network applied to multivariate time series data, generated by online handwriting data associated with continuous loop series handwritten on a graphical tablet. Our explainability methods reveal distinct motor behavior characteristics for healthy individuals and those diagnosed with Alzheimer's. Healthy subjects exhibited consistent, smooth movements, while Alzheimer's patients demonstrated erratic patterns marked by abrupt stops and direction changes. This emphasizes the critical role of explainability in translating complex models into clinically relevant insights. Our research contributes to the enhancement of early diagnosis, providing significant and reliable insights to stakeholders involved in patient care and intervention strategies. Our work bridges the gap between machine learning predictions and clinical insights, fostering a more effective and understandable application of advanced models for Alzheimer's disease assessment.
PMID:39333681 | DOI:10.1038/s41598-024-72650-2
Harnessing probabilistic neural network with triple tree seed algorithm-based smart enterprise quantitative risk management framework
Sci Rep. 2024 Sep 27;14(1):22293. doi: 10.1038/s41598-024-73876-w.
ABSTRACT
Enterprise risk management (ERM) frameworks convey vital principles that help create a consistent risk management culture, irrespective of employee turnover or industry standards. Enterprise Management System (EMS) are becoming a popular research area for assuring a company's long-term success. Statistical pattern recognition, federated learning, database administration, visualization technology, and social networking are all used in this field, which includes artificial intelligence (AI), data science, and statistics. Risk assessment in EMS is critical for enterprise decision-making to be effective. Recent advancements in AI, machine learning (ML), and deep learning (DL) concepts have enabled the development of effective risk assessment models for EMS. This special issue seeks groundbreaking research articles that showcase the application of applied probability and statistics to interdisciplinary studies. This study offers Improved Metaheuristics with a Deep Learning Enabled Risk Assessment Model (IMDLRA-SES) for Smart Enterprise Systems. Using feature selection (FS) and DL models, the provided IMDLRA-SES technique estimates business risks. Preprocessing is used in the IMDLRA-SES technique to change the original financial data into a usable format. In addition, an FS technique based on oppositional lion swarm optimization (OLSO) is utilized to find the best subset of features. In addition, the presence or absence of financial hazards in firms is classified using the triple tree seed algorithm (TTSA) with a probabilistic neural network (PNN) model. The TTSA is used as a hyperparameter optimizer to improve the efficiency of the PNN-based categorization. An extensive set of experimental evaluations is performed on German and Australian credit datasets to illustrate the IMDLRA-SES model's improved performance. The performance validation of the IMDLRA-SES model portrayed a superior accuracy value of 95.70% and 96.09% over existing techniques.
PMID:39333638 | DOI:10.1038/s41598-024-73876-w
Length-scale study in deep learning prediction for non-small cell lung cancer brain metastasis
Sci Rep. 2024 Sep 27;14(1):22328. doi: 10.1038/s41598-024-73428-2.
ABSTRACT
Deep learning-assisted digital pathology has demonstrated the potential to profoundly impact clinical practice, even surpassing human pathologists in performance. However, as deep neural network (DNN) architectures grow in size and complexity, their explainability decreases, posing challenges in interpreting pathology features for broader clinical insights into physiological diseases. To better assess the interpretability of digital microscopic images and guide future microscopic system design, we developed a novel method to study the predictive feature length-scale that underpins a DNN's predictive power. We applied this method to analyze a DNN's capability in predicting brain metastasis from early-stage non-small-cell lung cancer biopsy slides. This study quantifies DNN's attention for brain metastasis prediction, targeting features at both the cellular scale and tissue scale in H&E-stained histological whole slide images. At the cellular scale, the predictive power of DNNs progressively increases with higher resolution and significantly decreases when the resolvable feature length exceeds 5 microns. Additionally, DNN uses more macro-scale features associated with tissue architecture and is optimized when assessing visual fields greater than 41 microns. Our study computes the length-scale requirements for optimal DNN learning on digital whole-slide microscopic images, holding the promise to guide future optical microscope designs in pathology applications and facilitating downstream deep learning analysis.
PMID:39333630 | DOI:10.1038/s41598-024-73428-2
Deep-learning model accurately classifies multi-label lung ultrasound findings, enhancing diagnostic accuracy and inter-reader agreement
Sci Rep. 2024 Sep 27;14(1):22228. doi: 10.1038/s41598-024-72484-y.
ABSTRACT
Despite the increasing use of lung ultrasound (LUS) in the evaluation of respiratory disease, operators' competence constrains its effectiveness. We developed a deep-learning (DL) model for multi-label classification using LUS and validated its performance and efficacy on inter-reader variability. We retrospectively collected LUS and labeled as normal, B-line, consolidation, and effusion from patients undergoing thoracentesis at a tertiary institution between January 2018 and January 2022. The development and internal testing involved 7580 images from January 2018 and December 2020, and the model's performance was validated on a temporally separated test set (n = 985 images collected after January 2021) and two external test sets (n = 319 and 54 images). Two radiologists interpreted LUS with and without DL assistance and compared diagnostic performance and agreement. The model demonstrated robust performance with AUCs: 0.93 (95% CI 0.92-0.94) for normal, 0.87 (95% CI 0.84-0.89) for B-line, 0.82 (95% CI 0.78-0.86) for consolidation, and 0.94 (95% CI 0.93-0.95) for effusion. The model improved reader accuracy for binary discrimination (normal vs. abnormal; reader 1: 87.5-95.6%, p = 0.004; reader 2: 95.0-97.5%, p = 0.19), and agreement (k = 0.73-0.83, p = 0.01). In conclusion, the DL-based model may assist interpretation, improving accuracy and overcoming operator competence limitations in LUS.
PMID:39333570 | DOI:10.1038/s41598-024-72484-y
Prediction of android ransomware with deep learning model using hybrid cryptography
Sci Rep. 2024 Sep 27;14(1):22351. doi: 10.1038/s41598-024-70544-x.
ABSTRACT
In recent times, the number of malware on Android mobile phones has been growing, and a new kind of malware is Android ransomware. This research aims to address the emerging concerns about Android ransomware in the mobile sector. Previous studies highlight that the number of new Android ransomware is increasing annually, which poses a huge threat to the privacy of mobile phone users for sensitive data. Various existing techniques are active to detect ransomware and secure the data in the mobile cloud. However, these approaches lack accuracy and detection performance with insecure storage. To resolve this and enhance the security level, the proposed model is presented. This manuscript provides both recognition algorithms based on the deep learning model and secured storage of detected data in the cloud with a secret key to safeguard sensitive user information using the hybrid cryptographic model. Initially, the input APK files and data are preprocessed to extract features. The collection of optimal features is carried out using the Squirrel search optimization process. After that, the Deep Learning-based model, adaptive deep saliency The AlexNet classifier is presented to detect and classify data as malicious or normal. The detected data, which is not malicious, is stored on a cloud server. For secured storage of data in the cloud, a hybrid cryptographic model such as hybrid homomorphic Elliptic Curve Cryptography and Blowfish is employed, which includes key computation and key generation processes. The cryptographic scheme includes encryption and decryption of data, after which the application response is found to attain a decrypted result upon user request. The performance is carried out for both the Deep Learning-based model and the hybrid cryptography-based security model, and the results obtained are 99.89% accuracy in detecting malware compared with traditional models. The effectiveness of the proposed system over other models such as GNN is 94.76%, CNN is 95.76%, and Random Forest is 96%.
PMID:39333540 | DOI:10.1038/s41598-024-70544-x