Deep learning

UPicker: a semi-supervised particle picking transformer method for cryo-EM micrographs

Tue, 2024-12-10 06:00

Brief Bioinform. 2024 Nov 22;26(1):bbae636. doi: 10.1093/bib/bbae636.

ABSTRACT

Automatic single particle picking is a critical step in the data processing pipeline of cryo-electron microscopy structure reconstruction. In recent years, several deep learning-based algorithms have been developed, demonstrating their potential to solve this challenge. However, current methods highly depend on manually labeled training data, which is labor-intensive and prone to biases especially for high-noise and low-contrast micrographs, resulting in suboptimal precision and recall. To address these problems, we propose UPicker, a semi-supervised transformer-based particle-picking method with a two-stage training process: unsupervised pretraining and supervised fine-tuning. During the unsupervised pretraining, an Adaptive Laplacian of Gaussian region proposal generator is proposed to obtain pseudo-labels from unlabeled data for initial feature learning. For the supervised fine-tuning, UPicker only needs a small amount of labeled data to achieve high accuracy in particle picking. To further enhance model performance, UPicker employs a contrastive denoising training strategy to reduce redundant detections and accelerate convergence, along with a hybrid data augmentation strategy to deal with limited labeled data. Comprehensive experiments on both simulated and experimental datasets demonstrate that UPicker outperforms state-of-the-art particle-picking methods in terms of accuracy and robustness while requiring fewer labeled data than other transformer-based models. Furthermore, ablation studies demonstrate the effectiveness and necessity of each component of UPicker. The source code and data are available at https://github.com/JachyLikeCoding/UPicker.

PMID:39658205 | DOI:10.1093/bib/bbae636

Categories: Literature Watch

Deep learning reveals pathology-confirmed neuroimaging signatures in Alzheimer's, vascular and Lewy body dementias

Tue, 2024-12-10 06:00

Brain. 2024 Dec 9:awae388. doi: 10.1093/brain/awae388. Online ahead of print.

ABSTRACT

Concurrent neurodegenerative and vascular pathologies pose a diagnostic challenge in the clinical setting, with histopathology remaining the definitive modality for dementia-type diagnosis. To address this clinical challenge, we introduce a neuropathology-based, data-driven, multi-label deep learning framework to identify and quantify in-vivo biomarkers for Alzheimer's disease (AD), vascular dementia (VD), and Lewy body dementia (LBD) using antemortem T1-weighted MRI scans of 423 demented and 361 control participants from NACC and ADNI datasets. Based on the best-performing deep learning model, explainable heatmaps are extracted to visualize disease patterns, and the novel Deep Signature of Pathology Atrophy REcognition (DeepSPARE) indices are developed, where a higher DeepSPARE score indicates more brain alterations associated with that specific pathology. A substantial discrepancy in clinical and neuropathology diagnosis was observed in the demented patients: 71% of them had more than one pathology, but 67% of them were clinically diagnosed as AD only. Based on these neuropathology diagnoses and leveraging cross-validation principles, the deep learning model achieved the best performance with a balanced accuracy of 0.844, 0.839, and 0.623 for AD, VD, and LBD, respectively, and was used to generate the explainable deep-learning heatmaps and DeepSPARE indices. The explainable deep-learning heatmaps revealed distinct neuroimaging brain alteration patterns for each pathology: the AD heatmap highlighted bilateral hippocampal regions, the VD heatmap emphasized white matter regions, and the LBD heatmap exposed occipital alterations. The DeepSPARE indices were validated by examining their associations with cognitive testing, neuropathological, and neuroimaging measures using linear mixed-effects models. The DeepSPARE-AD index was associated with MMSE, Trail B, memory, PFDR-adjustedhippocampal volume, Braak stages, CERAD scores, and Thal phases (PFDR-adjusted < 0.05). The DeepSPARE-VD index was associated with white matter hyperintensity volume and cerebral amyloid angiopathy (PFDR-adjusted < 0.001). The DeepSPARE-LBD index was associated with Lewy body stages (PFDR-adjusted < 0.05). The findings were replicated in an out-of-sample ADNI dataset by testing associations with cognitive, imaging, plasma, and CSF measures. CSF and plasma pTau181 were significantly associated with DeepSPARE-AD in the AD/MCIΑβ+ group (PFDR-adjusted < 0.001), and CSF α-synuclein was associated solely with DeepSPARE-LBD (PFDR-adjusted = 0.036). Overall, these findings demonstrate the advantages of our innovative deep-learning framework in detecting antemortem neuroimaging signatures linked to different pathologies. The newly deep learning-derived DeepSPARE indices are precise, pathology-sensitive, and single-valued noninvasive neuroimaging metrics, bridging the traditional widely available in-vivo T1 imaging with histopathology.

PMID:39657969 | DOI:10.1093/brain/awae388

Categories: Literature Watch

End-to-end deep learning patient level classification of affected territory of ischemic stroke patients in DW-MRI

Tue, 2024-12-10 06:00

Neuroradiology. 2024 Dec 10. doi: 10.1007/s00234-024-03520-x. Online ahead of print.

ABSTRACT

PURPOSE: To develop an end-to-end DL model for automated classification of affected territory in DWI of stroke patients.

MATERIALS AND METHODS: In this retrospective multicenter study, brain DWI studies from January 2017 to April 2020 from Center 1, from June 2020 to December 2020 from Center 2, and from November 2019 to April 2020 from Center 3 were included. Four radiologists labeled images into five classes: anterior cerebral artery (ACA), middle cerebral artery (MCA), posterior circulation (PC), and watershed (WS) regions, as well as normal images. Additionally, for Center 1, clinical information was encoded as a domain knowledge vector to incorporate into image embeddings. 3D convolutional neural network (CNN) and attention gate integrated versions for direct 3D encoding, long short-term memory (LSTM-CNN), and time-distributed layer for slice-based encoding were employed. Balanced classification accuracy, macro averaged f1 score, AUC, and interrater Cohen's kappa were calculated.

RESULTS: Overall, 624 DWI MRIs from 3 centers were utilized (mean age, interval: 66.89 years, 29-95 years; 345 male) with 439 patients in the training, 103 in the validation, and 82 in the test sets. The best model was a slice-based parallel encoding model with 0.88 balanced accuracy, 0.80 macro-f1 score, and an AUC of 0.98. Clinical domain knowledge integration improved the performance with 0.93 best overall accuracy with parallel stream model embeddings and support vector machine classifiers. The mean kappa value for interrater agreement was 0.87.

CONCLUSION: Developed end-to-end deep learning models performed well in classifying affected regions from stroke in DWI.

CLINICAL RELEVANCE STATEMENT: The end-to-end deep learning model with a parallel stream encoding strategy for classifying stroke regions in DWI has performed comparably with radiologists.

PMID:39656236 | DOI:10.1007/s00234-024-03520-x

Categories: Literature Watch

Utilizing deep learning-based causal inference to explore vancomycin's impact on continuous kidney replacement therapy necessity in blood culture-positive intensive care unit patients

Tue, 2024-12-10 06:00

Microbiol Spectr. 2024 Dec 10:e0266224. doi: 10.1128/spectrum.02662-24. Online ahead of print.

ABSTRACT

Patients with positive blood cultures in the intensive care unit (ICU) are at high risk for septic acute kidney injury requiring continuous kidney replacement therapy (CKRT), especially when treated with vancomycin. This study developed a machine learning model to predict CKRT and examined vancomycin's impact using deep learning-based causal inference. We analyzed ICU patients with positive blood cultures, utilizing the Medical Information Mart for Intensive Care III data set. The primary outcome was defined as the initiation of CKRT during the ICU stay. The machine learning models were developed to predict the outcome. The deep learning-based causal inference model was utilized to quantitatively demonstrate the impact of vancomycin on the probability of CKRT initiation. Logistic regression was performed to analyze the relationship between the variables and the susceptibility of vancomycin. A total of 1,318 patients were included in the analysis, with 41 requiring CKRT. The Random Forest and Light Gradient Boosting Machine exhibited the best performance, with Area Under Curve of Receiver Operating Characteristic Curve values of 0.905 and 0.886, respectively. The deep learning-based causal inference model demonstrated an average 7.7% increase in the probability of CKRT occurrence when administrating vancomycin in total data set. Additionally, that younger age, lower diastolic blood pressure, higher heart rate, higher baseline creatinine, and lower bicarbonate levels sensitized the probability of CKRT application in response to vancomycin treatment. Deep learning-based causal inference models showed that vancomycin administration increases CKRT risk, identifying specific patient characteristics associated with higher susceptibility.IMPORTANCEThis study assesses the impact of vancomycin on the risk of continuous kidney replacement therapy (CKRT) in intensive care unit (ICU) patients with blood culture-positive infections. Utilizing deep learning-based causal inference and machine learning models, the research quantifies how vancomycin administration increases CKRT risk by an average of 7.7%. Key variables influencing susceptibility include baseline creatinine, diastolic blood pressure, heart rate, and bicarbonate levels. These findings offer insights into managing vancomycin-induced kidney risk and may inform patient-specific treatment strategies in ICU settings.

PMID:39656005 | DOI:10.1128/spectrum.02662-24

Categories: Literature Watch

Rapid diagnosis of bacterial vaginosis using machine-learning-assisted surface-enhanced Raman spectroscopy of human vaginal fluids

Tue, 2024-12-10 06:00

mSystems. 2024 Dec 10:e0105824. doi: 10.1128/msystems.01058-24. Online ahead of print.

ABSTRACT

Bacterial vaginosis (BV) is an abnormal gynecological condition caused by the overgrowth of specific bacteria in the vagina. This study aims to develop a novel method for BV detection by integrating surface-enhanced Raman scattering (SERS) with machine learning (ML) algorithms. Vaginal fluid samples were classified as BV positive or BV negative using the BVBlue Test and clinical microscopy, followed by SERS spectral acquisition to construct the data set. Preliminary SERS spectral analysis revealed notable disparities in characteristic peak features. Multiple ML models were constructed and optimized, with the convolutional neural network (CNN) model achieving the highest prediction accuracy at 99%. Gradient-weighted class activation mapping (Grad-CAM) was used to highlight important regions in the images for prediction. Moreover, the CNN model was blindly tested on SERS spectra of vaginal fluid samples collected from 40 participants with unknown BV infection status, achieving a prediction accuracy of 90.75% compared with the results of the BVBlue Test combined with clinical microscopy. This novel technique is simple, cheap, and rapid in accurately diagnosing bacterial vaginosis, potentially complementing current diagnostic methods in clinical laboratories.

IMPORTANCE: The accurate and rapid diagnosis of bacterial vaginosis (BV) is crucial due to its high prevalence and association with serious health complications, including increased risk of sexually transmitted infections and adverse pregnancy outcomes. Although widely used, traditional diagnostic methods have significant limitations in subjectivity, complexity, and cost. The development of a novel diagnostic approach that integrates SERS with ML offers a promising solution. The CNN model's high prediction accuracy, cost-effectiveness, and extraordinary rapidity underscore its significant potential to enhance the diagnosis of BV in clinical settings. This method not only addresses the limitations of current diagnostic tools but also provides a more accessible and reliable option for healthcare providers, ultimately enhancing patient care and health outcomes.

PMID:39655908 | DOI:10.1128/msystems.01058-24

Categories: Literature Watch

Systematic review of experimental paradigms and deep neural networks for electroencephalography-based cognitive workload detection

Tue, 2024-12-10 06:00

Prog Biomed Eng (Bristol). 2024 Oct 21;6(4). doi: 10.1088/2516-1091/ad8530.

ABSTRACT

This article summarizes a systematic literature review of deep neural network-based cognitive workload (CWL) estimation from electroencephalographic (EEG) signals. The focus of this article can be delineated into two main elements: first is the identification of experimental paradigms prevalently employed for CWL induction, and second, is an inquiry about the data structure and input formulations commonly utilized in deep neural networks (DNN)-based CWL detection. The survey revealed several experimental paradigms that can reliably induce either graded levels of CWL or a desired cognitive state due to sustained induction of CWL. This article has characterized them with respect to the number of distinct CWL levels, cognitive states, experimental environment, and agents in focus. Further, this literature analysis found that DNNs can successfully detect distinct levels of CWL despite the inter-subject and inter-session variability typically observed in EEG signals. Several methodologies were found using EEG signals in its native representation of a two-dimensional matrix as input to the classification algorithm, bypassing traditional feature selection steps. More often than not, researchers used DNNs as black-box type models, and only a few studies employed interpretable or explainable DNNs for CWL detection. However, these algorithms were mostly post hoc data analysis and classification schemes, and only a few studies adopted real-time CWL estimation methodologies. Further, it has been suggested that using interpretable deep learning methodologies may shed light on EEG correlates of CWL, but this remains mostly an unexplored area. This systematic review suggests using networks sensitive to temporal dependencies and appropriate input formulations for each type of DNN architecture to achieve robust classification performance. An additional suggestion is to utilize transfer learning methods to achieve high generalizability across tasks (task-independent classifiers), while simple cross-subject data pooling may achieve the same for subject-independent classifiers.

PMID:39655862 | DOI:10.1088/2516-1091/ad8530

Categories: Literature Watch

Ultrasound imaging based recognition of prenatal anomalies: a systematic clinical engineering review

Tue, 2024-12-10 06:00

Prog Biomed Eng (Bristol). 2024 May 7;6(2). doi: 10.1088/2516-1091/ad3a4b.

ABSTRACT

For prenatal screening, ultrasound (US) imaging allows for real-time observation of developing fetal anatomy. Understanding normal and aberrant forms through extensive fetal structural assessment enables for early detection and intervention. However, the reliability of anomaly diagnosis varies depending on operator expertise and device limits. First trimester scans in conjunction with circulating biochemical markers are critical in identifying high-risk pregnancies, but they also pose technical challenges. Recent engineering advancements in automated diagnosis, such as artificial intelligence (AI)-based US image processing and multimodal data fusion, are developing to improve screening efficiency, accuracy, and consistency. Still, creating trust in these data-driven solutions is necessary for integration and acceptability in clinical settings. Transparency can be promoted by explainable AI (XAI) technologies that provide visual interpretations and illustrate the underlying diagnostic decision making process. An explanatory framework based on deep learning is suggested to construct charts depicting anomaly screening results from US video feeds. AI modelling can then be applied to these charts to connect defects with probable deformations. Overall, engineering approaches that increase imaging, automation, and interpretability hold enormous promise for altering traditional workflows and expanding diagnostic capabilities for better prenatal care.

PMID:39655845 | DOI:10.1088/2516-1091/ad3a4b

Categories: Literature Watch

Machine Learning-Based Prediction for In-Hospital Mortality After Acute Intracerebral Hemorrhage Using Real-World Clinical and Image Data

Tue, 2024-12-10 06:00

J Am Heart Assoc. 2024 Dec 10:e036447. doi: 10.1161/JAHA.124.036447. Online ahead of print.

ABSTRACT

BACKGROUND: Machine learning (ML) techniques are widely employed across various domains to achieve accurate predictions. This study assessed the effectiveness of ML in predicting early mortality risk among patients with acute intracerebral hemorrhage (ICH) in real-world settings.

METHODS AND RESULTS: ML-based models were developed to predict in-hospital mortality in 527 patients with ICH using raw brain imaging data from brain computed tomography and clinical data. The models' performances were evaluated using the area under the receiver operating characteristic curves and calibration plots, comparing them with traditional risk scores such as the ICH score and ICH grading scale. Kaplan-Meier curves were used to examine the post-ICH survival rates, stratified by ML-based risk assessment. The net benefit of ML-based models was evaluated using decision curve analysis. The area under the receiver operating characteristic curves were 0.91 (95% CI, 0.86-0.95) for the ICH score, 0.93 (95% CI, 0.89-0.97) for the ICH grading scale, 0.83 (95% CI, 0.71-0.91) for the ML-based model fitted with raw image data only, and 0.87 (95% CI, 0.76-0.93) for the ML-based model fitted using clinical data without specialist expertise. The area under the receiver operating characteristic curve increased significantly to 0.97 (95% CI, 0.94-0.99) when the ML model was fitted using clinical and image data assessed by specialists. All ML-based models demonstrated good calibration, and the survival rates showed significant differences between risk groups. Decision curve analysis indicated the highest net benefit when utilizing the findings assessed by specialists.

CONCLUSIONS: ML-based prediction models exhibit satisfactory performance in predicting post-ICH in-hospital mortality when utilizing raw imaging data or nonspecialist input. Nevertheless, incorporating specialist expertise notably improves performance.

PMID:39655759 | DOI:10.1161/JAHA.124.036447

Categories: Literature Watch

Assessment of image quality on the diagnostic performance of clinicians and deep learning models: Cross-sectional comparative reader study

Tue, 2024-12-10 06:00

J Eur Acad Dermatol Venereol. 2024 Dec 10. doi: 10.1111/jdv.20462. Online ahead of print.

ABSTRACT

BACKGROUND: Skin cancer is a prevalent and clinically significant condition, with early and accurate diagnosis being crucial for improved patient outcomes. Dermoscopy and artificial intelligence (AI) hold promise in enhancing diagnostic accuracy. However, the impact of image quality, particularly high dynamic range (HDR) conversion in smartphone images, on diagnostic performance remains poorly understood.

OBJECTIVE: This study aimed to investigate the effect of varying image qualities, including HDR-enhanced dermoscopic images, on the diagnostic capabilities of clinicians and a convolutional neural network (CNN) model.

METHODS: Eighteen dermatology clinicians assessed 303 images of 101 skin lesions that were categorized into three image quality groups: low quality (LQ), high quality (HQ) and enhanced quality (EQ) produced using HDR-style conversion. Clinicians participated in a two part reader study that required their diagnosis, management and confidence level for each image assessed.

RESULTS: In the binary classification of lesions, clinicians had the greatest diagnostic performance with HQ images, with sensitivity (77.3%; CI 69.1-85.5), specificity (63.1%; CI 53.7-72.5) and accuracy (70.2%; CI 61.3-79.1). For the multiclass classification, the overall performance was also best with HQ images, attaining the greatest specificity (91.9%; CI 83.2-95.0) and accuracy (51.5%; CI 48.4-54.7). Clinicians had a superior performance (median correct diagnoses) to the CNN model for the binary classification of LQ and EQ images, but their performance was comparable on the HQ images. However, in the multiclass classification, the CNN model significantly outperformed the clinicians on HQ images (p < 0.01).

CONCLUSION: This study highlights the importance of image quality on the diagnostic performance of clinicians and deep learning models. This has significant implications for telehealth reporting and triage.

PMID:39655640 | DOI:10.1111/jdv.20462

Categories: Literature Watch

Automatic classification of fungal-fungal interactions using deep leaning models

Tue, 2024-12-10 06:00

Comput Struct Biotechnol J. 2024 Nov 14;23:4222-4231. doi: 10.1016/j.csbj.2024.11.027. eCollection 2024 Dec.

ABSTRACT

Fungi provide valuable solutions for diverse biotechnological applications, such as enzymes in the food industry, bioactive metabolites for healthcare, and biocontrol organisms in agriculture. Current workflows for identifying new biocontrol fungi often rely on subjective visual observations of strains' performance in microbe-microbe interaction studies, making the process time-consuming and difficult to reproduce. To overcome these challenges, we developed an AI-automated image classification approach using machine learning algorithm based on deep neural network. Our method focuses on analyzing standardized images of 96-well microtiter plates with solid medium for fungal-fungal challenge experiments. We used our model to categorize the outcome of interactions between the plant pathogen Fusarium graminearum and individual isolates from a collection of 38,400 fungal strains. The authors trained multiple deep learning architectures and evaluated their performance. The results strongly support our approach, achieving a peak accuracy of 95.0 % with the DenseNet121 model and a maximum macro-averaged F1-Score of 93.1 across five folds. To the best of our knowledge, this paper introduces the first automated method for classifying fungal-fungal interactions using deep learning, which can easily be adapted for other fungal species.

PMID:39655263 | PMC:PMC11626056 | DOI:10.1016/j.csbj.2024.11.027

Categories: Literature Watch

Focused review on artificial intelligence for disease detection in infants

Tue, 2024-12-10 06:00

Front Digit Health. 2024 Nov 25;6:1459640. doi: 10.3389/fdgth.2024.1459640. eCollection 2024.

ABSTRACT

Over the last years, studies using artificial intelligence (AI) for the detection and prediction of diseases have increased and also concentrated more and more on vulnerable groups of individuals, such as infants. The release of ChatGPT demonstrated the potential of large language models (LLMs) and heralded a new era of AI with manifold application possibilities. However, the impact of this new technology on medical research cannot be fully estimated yet. In this work, we therefore aimed to summarise the most recent pre-ChatGPT developments in the field of automated detection and prediction of diseases and disease status in infants, i.e., within the first 12 months of life. For this, we systematically searched the scientific databases PubMed and IEEE Xplore for original articles published within the last five years preceding the release of ChatGPT (2018-2022). The search revealed 927 articles; a final number of 154 articles was included for review. First of all, we examined research activity over time. Then, we analysed the articles from 2022 for medical conditions, data types, tasks, AI approaches, and reported model performance. A clear trend of increasing research activity over time could be observed. The most recently published articles focused on medical conditions of twelve different ICD-11 categories; "certain conditions originating in the perinatal period" was the most frequently addressed disease category. AI models were trained with a variety of data types, among which clinical and demographic information and laboratory data were most frequently exploited. The most frequently performed tasks aimed to detect present diseases, followed by the prediction of diseases and disease status at a later point in development. Deep neural networks turned out as the most popular AI approach, even though traditional methods, such as random forests and support vector machines, still play a role-presumably due to their explainability or better suitability when the amount of data is limited. Finally, the reported performances in many of the reviewed articles suggest that AI has the potential to assist in diagnostic procedures for infants in the near future. LLMs will boost developments in this field in the upcoming years.

PMID:39654981 | PMC:PMC11625793 | DOI:10.3389/fdgth.2024.1459640

Categories: Literature Watch

LesionScanNet: dual-path convolutional neural network for acute appendicitis diagnosis

Tue, 2024-12-10 06:00

Health Inf Sci Syst. 2024 Dec 7;13(1):3. doi: 10.1007/s13755-024-00321-7. eCollection 2025 Dec.

ABSTRACT

Acute appendicitis is an abrupt inflammation of the appendix, which causes symptoms such as abdominal pain, vomiting, and fever. Computed tomography (CT) is a useful tool in accurate diagnosis of acute appendicitis; however, it causes challenges due to factors such as the anatomical structure of the colon and localization of the appendix in CT images. In this paper, a novel Convolutional Neural Network model, namely, LesionScanNet for the computer-aided detection of acute appendicitis has been proposed. For this purpose, a dataset of 2400 CT scan images was collected by the Department of General Surgery at Kanuni Sultan Süleyman Research and Training Hospital, Istanbul, Turkey. LesionScanNet is a lightweight model with 765 K parameters and includes multiple DualKernel blocks, where each block contains a convolution, expansion, separable convolution layers, and skip connections. The DualKernel blocks work with two paths of input image processing, one of which uses 3 × 3 filters, and the other path encompasses 1 × 1 filters. It has been demonstrated that the LesionScanNet model has an accuracy score of 99% on the test set, a value that is greater than the performance of the benchmark deep learning models. In addition, the generalization ability of the LesionScanNet model has been demonstrated on a chest X-ray image dataset for pneumonia and COVID-19 detection. In conclusion, LesionScanNet is a lightweight and robust network achieving superior performance with smaller number of parameters and its usage can be extended to other medical application domains.

PMID:39654693 | PMC:PMC11625030 | DOI:10.1007/s13755-024-00321-7

Categories: Literature Watch

Integrating artificial intelligence in strabismus management: current research landscape and future directions

Tue, 2024-12-10 06:00

Exp Biol Med (Maywood). 2024 Nov 25;249:10320. doi: 10.3389/ebm.2024.10320. eCollection 2024.

ABSTRACT

Advancements in artificial intelligence (AI) are transforming strabismus management through improved screening, diagnosis, and surgical planning. Deep learning has notably enhanced diagnostic accuracy and optimized surgical outcomes. Despite these advancements, challenges such as the underrepresentation of diverse strabismus types and reliance on single-source data remain prevalent. Emphasizing the need for inclusive AI systems, future research should focus on expanding AI capabilities with large model technologies, integrating multimodal data to bridge existing gaps, and developing integrated management platforms to better accommodate diverse patient demographics and clinical scenarios.

PMID:39654660 | PMC:PMC11625544 | DOI:10.3389/ebm.2024.10320

Categories: Literature Watch

The role of deep learning in myocardial perfusion imaging for diagnosis and prognosis: A systematic review

Tue, 2024-12-10 06:00

iScience. 2024 Nov 12;27(12):111374. doi: 10.1016/j.isci.2024.111374. eCollection 2024 Dec 20.

ABSTRACT

The development of state-of-the-art algorithms for computer visualization has led to a growing interest in applying deep learning (DL) techniques to the field of medical imaging. DL-based algorithms have been extensively utilized in various aspects of cardiovascular imaging, and one notable area of focus is single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI), which is regarded as the gold standard for non-invasive diagnosis of myocardial ischemia. However, due to the complex decision-making process of DL based on convolutional neural networks (CNNs), the explainability of DL results has become a significant area of research, particularly in the field of medical imaging. To better harness the potential of DL and to be well prepared for the ongoing DL revolution in nuclear imaging, this review aims to summarize the recent applications of DL in MPI, with a specific emphasis on the methods in explainable DL for the diagnosis and prognosis of MPI. Furthermore, the challenges and potential directions for future research are also discussed.

PMID:39654634 | PMC:PMC11626733 | DOI:10.1016/j.isci.2024.111374

Categories: Literature Watch

Toward explainable deep learning in healthcare through transition matrix and user-friendly features

Tue, 2024-12-10 06:00

Front Artif Intell. 2024 Nov 25;7:1482141. doi: 10.3389/frai.2024.1482141. eCollection 2024.

ABSTRACT

Modern artificial intelligence (AI) solutions often face challenges due to the "black box" nature of deep learning (DL) models, which limits their transparency and trustworthiness in critical medical applications. In this study, we propose and evaluate a scalable approach based on a transition matrix to enhance the interpretability of DL models in medical signal and image processing by translating complex model decisions into user-friendly and justifiable features for healthcare professionals. The criteria for choosing interpretable features were clearly defined, incorporating clinical guidelines and expert rules to align model outputs with established medical standards. The proposed approach was tested on two medical datasets: electrocardiography (ECG) for arrhythmia detection and magnetic resonance imaging (MRI) for heart disease classification. The performance of the DL models was compared with expert annotations using Cohen's Kappa coefficient to assess agreement, achieving coefficients of 0.89 for the ECG dataset and 0.80 for the MRI dataset. These results demonstrate strong agreement, underscoring the reliability of the approach in providing accurate, understandable, and justifiable explanations of DL model decisions. The scalability of the approach suggests its potential applicability across various medical domains, enhancing the generalizability and utility of DL models in healthcare while addressing practical challenges and ethical considerations.

PMID:39654544 | PMC:PMC11625760 | DOI:10.3389/frai.2024.1482141

Categories: Literature Watch

DTMP-prime: A deep transformer-based model for predicting prime editing efficiency and PegRNA activity

Tue, 2024-12-10 06:00

Mol Ther Nucleic Acids. 2024 Oct 28;35(4):102370. doi: 10.1016/j.omtn.2024.102370. eCollection 2024 Dec 10.

ABSTRACT

Prime editors are CRISPR-based genome engineering tools with significant potential for rectifying patient mutations. However, their usage requires experimental optimization of the prime editing guide RNA (PegRNA) to achieve high editing efficiency. This paper introduces the deep transformer-based model for predicting prime editing efficiency (DTMP-Prime), a tool specifically designed to predict PegRNA activity and prime editing (PE) efficiency. DTMP-Prime facilitates the design of appropriate PegRNA and ngRNA. A transformer-based model was constructed to scrutinize a wide-ranging set of PE data, enabling the extraction of effective features of PegRNAs and target DNA sequences. The integration of these features with the proposed encoding strategy and DNABERT-based embedding has notably improved the predictive capabilities of DTMP-Prime for off-target sites. Moreover, DTMP-Prime is a promising tool for precisely predicting off-target sites in CRISPR experiments. The integration of a multi-head attention framework has additionally improved the precision and generalizability of DTMP-Prime across various PE models and cell lines. Evaluation results based on the Pearson and Spearman correlation coefficient demonstrate that DTMP-Prime outperforms other state-of-the-art models in predicting the efficiency and outcomes of PE experiments.

PMID:39654539 | PMC:PMC11626815 | DOI:10.1016/j.omtn.2024.102370

Categories: Literature Watch

Artificial intelligence applications in smile design dentistry: A scoping review

Tue, 2024-12-10 06:00

J Prosthodont. 2024 Dec 9. doi: 10.1111/jopr.14000. Online ahead of print.

ABSTRACT

PURPOSE: Artificial intelligence (AI) applications are growing in smile design and aesthetic procedures. The current expansion and performance of AI models in digital smile design applications have not yet been systematically documented and analyzed. The purpose of this review was to assess the performance of AI models in smile design, assess the criteria of points of reference using AI analysis, and assess different AI software performance.

METHODS: An electronic review was completed in five databases: MEDLINE/PubMed, EMBASE, World of Science, Cochrane, and Scopus. Studies that developed AI models for smile design were included. The search strategy included articles published until November 1, 2024. Two investigators independently evaluated the quality of the studies by applying the Joanna Briggs Institute (JBI) Critical Appraisal Checklist for Quasi-Experimental Studies and Textual Evidence: Expert Opinion Results.

RESULTS: The search resulted in 2653 articles. A total of 2649 were excluded according to the exclusion criteria after reading the title, abstract, and/or full-text review. Four articles published between 2023 and 2024 were included in the present investigation. Two articles compared 2D and 3D points while one article compared the outcome of satisfaction between dentists and patients, and the last article emphasized the ethical components of using AI.

CONCLUSION: The results of the studies reviewed in this paper suggest that AI-generated smile designs are not significantly different from manually created designs in terms of esthetic perception. 3D designs are more accurate than 2D designs and offer more advantages. More articles are needed in the field of AI and smile design.

PMID:39654301 | DOI:10.1111/jopr.14000

Categories: Literature Watch

Hierarchical Graph Attention Network with Positive and Negative Attentions for Improved Interpretability: ISA-PN

Mon, 2024-12-09 06:00

J Chem Inf Model. 2024 Dec 9. doi: 10.1021/acs.jcim.4c01035. Online ahead of print.

ABSTRACT

With the advancement of deep learning (DL) methods in chemistry and materials science, the interpretability of DL models has become a critical issue in elucidating quantitative (molecular) structure-property relationships. Although attention mechanisms have been generally employed to explain the importance of molecular substructures that contribute to molecular properties, their interpretability remains limited. In this work, we introduce a versatile segmentation method and develop an interpretable subgraph attention (ISA) network with positive and negative streams (ISA-PN) to enhance the understanding of molecular structure-property relationships. The predictive performance of the ISA models was validated using data sets for aqueous solubility, lipophilicity, and melting temperature, with a particular focus on evaluating interpretability for the aqueous solubility data set. The ISA-PN model enables the quantification of the contributions of molecular substructures through positive and negative attention scores. Comparative analyses of the ISA, ISA-PN, and GC-Net (group contribution network) models demonstrate that the ISA-PN model significantly improves interpretability while maintaining similar accuracy levels. This study highlights the efficacy of the ISA-PN model in providing meaningful insights into the contributions of molecular substructures to molecular properties, thereby enhancing the interpretability of DL models in chemical applications.

PMID:39654089 | DOI:10.1021/acs.jcim.4c01035

Categories: Literature Watch

Prediction of Nursing Need Proxies Using Vital Signs and Biomarkers Data: Application of Deep Learning Models

Mon, 2024-12-09 06:00

J Clin Nurs. 2024 Dec 9. doi: 10.1111/jocn.17612. Online ahead of print.

ABSTRACT

AIM: To develop deep learning models to predict nursing need proxies among hospitalised patients and compare their predictive efficacy to that of a traditional regression model.

DESIGN: This methodological study employed a cross-sectional secondary data analysis.

METHODS: This study used de-identified electronic health records data from 20,855 adult patients aged 20 years or older, admitted to the general wards at a tertiary hospital. The models utilised patient information covering the preceding 2 days, comprising vital signs, biomarkers and demographic data. To create nursing need proxies, we identified the six highest-workload nursing tasks. We structured the collected data sequentially to facilitate processing via recurrent neural network (RNN) and long short-term memory (LSTM) algorithms. The STROBE checklist for cross-sectional studies was used for reporting.

RESULTS: Both the RNN and LSTM predicted nursing need proxies more effectively than the traditional regression model. However, upon testing the models using a sample case dataset, we observed a notable reduction in prediction accuracy during periods marked by rapid change.

CONCLUSIONS: The RNN and LSTM, which enhanced predictive performance for nursing needs, were developed using iterative learning processes. The RNN and LSTM demonstrated predictive capabilities superior to the traditional multiple regression model for nursing need proxies.

IMPLICATIONS FOR THE PROFESSION: Applying these predictive models in clinical settings where medical care complexity and diversity are increasing could substantially mitigate the uncertainties inherent in decision-making processes.

PATIENT OR PUBLIC CONTRIBUTION: We used de-identified electronic health record data of 20,855 adult patients about vital signs, biomarkers and nursing activities.

REPORTING METHOD: The authors state that they have adhered to relevant EQUATOR guidelines: STROBE statement for cross-sectional studies.

IMPACT: Despite widespread adoption of deep learning algorithms in various industries, their application in nursing administration for workload distribution and staffing adequacy remains limited. This study amalgamated deep learning technology to develop a predictive model to proactively forecast nursing need proxies. Our study demonstrates that both the RNN and LSTM models outperform a traditional regression model in predicting nursing need proxies. The proactive application of deep learning methods for nursing need prediction could help facilitate timely detection of changes in patient nursing demands, enabling the effective and safe nursing services.

PMID:39654010 | DOI:10.1111/jocn.17612

Categories: Literature Watch

Fusion Learning from Non-contrast CT Scans for the Detection of Hemorrhagic Transformation in Stroke Patients

Mon, 2024-12-09 06:00

J Imaging Inform Med. 2024 Dec 9. doi: 10.1007/s10278-024-01350-0. Online ahead of print.

ABSTRACT

Hemorrhagic transformation (HT) is a potentially catastrophic complication after acute ischemic stroke. Prevention of HT risk is crucial because it worsens prognosis and increases mortality. This study aimed at developing and validating a computer-aided diagnosis system using pretreatment non-contrast computed tomography (CT) scans for HT prediction in stroke patients undergoing revascularization. This retrospective study included all acute ischemic stroke patients with non-contrast CT before reperfusion therapy who also underwent follow-up MRI from January 2018 to December 2022. Among the 188 evaluated patients, any degree of HT at follow-up imaging was observed in 103 patients. HT diagnosis via MRI was defined as the reference standard for neuroradiologists. Using a database of 2076 serial non-contrast CT images of the brain, pretrained deep learning architectures such as convolutional neural networks and vision transformers (ViTs) were used for feature extraction. The performance of the predictive HT risk model was evaluated via tenfold cross-validation in machine learning classifiers. The accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC) were evaluated. Using an individual deep learning architecture, DenseNet201 features achieved the highest accuracy of 87% and an AUC of 0.8863 in the classifier of the subspace ensemble k-nearest neighbor. By combining the DenseNet201 and ViT features, the accuracy and AUC can be improved to 88% and 0.8987, respectively, which are significantly better than those of using ViT alone. Detecting HT in stroke patients is a meaningful but challenging issue. On the basis of the model approach, HT diagnosis would be more automatic, efficient, and consistent, which would be helpful in clinic use.

PMID:39653876 | DOI:10.1007/s10278-024-01350-0

Categories: Literature Watch

Pages