Deep learning
Convolutional neural networks for automatic MR classification of myocardial iron overload in thalassemia major patients
Eur Radiol. 2024 Dec 10. doi: 10.1007/s00330-024-11245-x. Online ahead of print.
ABSTRACT
OBJECTIVES: To develop a deep-learning model for supervised classification of myocardial iron overload (MIO) from magnitude T2* multi-echo MR images.
MATERIALS AND METHODS: Eight hundred twenty-three cardiac magnitude T2* multi-slice, multi-echo MR images from 496 thalassemia major patients (285 females, 57%), labeled for MIO level (normal: T2* > 20 ms, moderate: 10 ≤ T2* ≤ 20 ms, severe: T2* < 10 ms), were retrospectively studied. Two 2D convolutional neural networks (CNN) developed for multi-slice (MS-HippoNet) and single-slice (SS-HippoNet) analysis were trained using 5-fold cross-validation. Performance was assessed using micro-average, multi-class accuracy, and single-class accuracy, sensitivity, and specificity. CNN performance was compared with inter-observer agreement between radiologists on 20% of the patients. The agreement between patients' classifications was assessed by the inter-agreement Kappa test.
RESULTS: Among the 165 images in the test set, a multi-class accuracy of 0.885 and 0.836 was obtained for MS- and SS-Hippo-Net, respectively. Network performances were confirmed on external test set analysis (0.827 and 0.793 multi-class accuracy, 29 patients from the CHMMOTv1 database). The agreement between automatic and ground truth classification was good (MS: κ = 0.771; SS: κ = 0.614), comparable with the inter-observer agreement (MS: κ = 0.872, SS: κ = 0.907) evaluated on the test set.
CONCLUSION: The developed networks performed classification of MIO level from multiecho, bright-blood, and T2* images with good performances.
KEY POINTS: Question MRI T2* represents the established clinical tool for MIO assessment. Quality control of the image analysis is a problem in small centers. Findings Deep learning models can perform MIO staging with good accuracy, comparable to inter-observer variability of the standard procedure. Clinical relevance CNN can perform automated staging of cardiac iron overload from multiecho MR sequences facilitating non-invasive evaluation of patients with various hematologic disorders.
PMID:39658686 | DOI:10.1007/s00330-024-11245-x
The role of deep learning in diagnostic imaging of spondyloarthropathies: a systematic review
Eur Radiol. 2024 Dec 10. doi: 10.1007/s00330-024-11261-x. Online ahead of print.
ABSTRACT
AIM: Diagnostic imaging is an integral part of identifying spondyloarthropathies (SpA), yet the interpretation of these images can be challenging. This review evaluated the use of deep learning models to enhance the diagnostic accuracy of SpA imaging.
METHODS: Following PRISMA guidelines, we systematically searched major databases up to February 2024, focusing on studies that applied deep learning to SpA imaging. Performance metrics, model types, and diagnostic tasks were extracted and analyzed. Study quality was assessed using QUADAS-2.
RESULTS: We analyzed 21 studies employing deep learning in SpA imaging diagnosis across MRI, CT, and X-ray modalities. These models, particularly advanced CNNs and U-Nets, demonstrated high accuracy in diagnosing SpA, differentiating arthritis forms, and assessing disease progression. Performance metrics frequently surpassed traditional methods, with some models achieving AUCs up to 0.98 and matching expert radiologist performance.
CONCLUSION: This systematic review underscores the effectiveness of deep learning in SpA imaging diagnostics across MRI, CT, and X-ray modalities. The studies reviewed demonstrated high diagnostic accuracy. However, the presence of small sample sizes in some studies highlights the need for more extensive datasets and further prospective and external validation to enhance the generalizability of these AI models.
KEY POINTS: Question How can deep learning models improve diagnostic accuracy in imaging for spondyloarthropathies (SpA), addressing challenges in early detection and differentiation from other forms of arthritis? Findings Deep learning models, especially CNNs and U-Nets, showed high accuracy in SpA imaging across MRI, CT, and X-ray, often matching or surpassing expert radiologists. Clinical relevance Deep learning models can enhance diagnostic precision in SpA imaging, potentially reducing diagnostic delays and improving treatment decisions, but further validation on larger datasets is required for clinical integration.
PMID:39658683 | DOI:10.1007/s00330-024-11261-x
Age-dependent changes in CT vertebral attenuation values in opportunistic screening for osteoporosis: a nationwide multi-center study
Eur Radiol. 2024 Dec 10. doi: 10.1007/s00330-024-11263-9. Online ahead of print.
ABSTRACT
OBJECTIVES: To examine how vertebral attenuation changes with aging, and to establish age-adjusted CT attenuation value cutoffs for diagnosing osteoporosis.
MATERIALS AND METHODS: This multi-center retrospective study included 11,246 patients (mean age ± standard deviation, 50 ± 13 years; 7139 men) who underwent CT and dual-energy X-ray absorptiometry (DXA) in six health-screening centers between 2022 and 2023. Using deep-learning-based software, attenuation values of L1 vertebral bodies were measured. Segmented linear regression in women and simple linear regression in men were used to assess how attenuation values change with aging. A multivariable linear regression analysis was performed to determine whether age is associated with CT attenuation values independently of the DXA T-score. Age-adjusted cutoffs targeting either 90% sensitivity or 90% specificity were derived using quantile regression. Performance of both age-adjusted and age-unadjusted cutoffs was measured, where the target sensitivity or specificity was considered achieved if a 95% confidence interval encompassed 90%.
RESULTS: While attenuation values declined consistently with age in men, they declined abruptly in women aged > 42 years. Such decline occurred independently of the DXA T-score (p < 0.001). Age adjustment seemed critical for age ≥ 65 years, where the age-adjusted cutoffs achieved the target (sensitivity of 91.5% (86.3-95.2%) when targeting 90% sensitivity and specificity of 90.0% (88.3-91.6%) when targeting 90% specificity), but age-unadjusted cutoffs did not (95.5% (91.2-98.0%) and 73.8% (71.4-76.1%), respectively).
CONCLUSION: Age-adjusted cutoffs provided a more reliable diagnosis of osteoporosis than age-unadjusted cutoffs since vertebral attenuation values decrease with age, regardless of DXA T-scores.
KEY POINTS: Question How does vertebral CT attenuation change with age? Findings Independent of dual-energy X-ray absorptiometry T-score, vertebral attenuation values on CT declined at a constant rate in men and abruptly in women over 42 years of age. Clinical relevance Age adjustments are needed in opportunistic osteoporosis screening, especially among the elderly.
PMID:39658682 | DOI:10.1007/s00330-024-11263-9
UPicker: a semi-supervised particle picking transformer method for cryo-EM micrographs
Brief Bioinform. 2024 Nov 22;26(1):bbae636. doi: 10.1093/bib/bbae636.
ABSTRACT
Automatic single particle picking is a critical step in the data processing pipeline of cryo-electron microscopy structure reconstruction. In recent years, several deep learning-based algorithms have been developed, demonstrating their potential to solve this challenge. However, current methods highly depend on manually labeled training data, which is labor-intensive and prone to biases especially for high-noise and low-contrast micrographs, resulting in suboptimal precision and recall. To address these problems, we propose UPicker, a semi-supervised transformer-based particle-picking method with a two-stage training process: unsupervised pretraining and supervised fine-tuning. During the unsupervised pretraining, an Adaptive Laplacian of Gaussian region proposal generator is proposed to obtain pseudo-labels from unlabeled data for initial feature learning. For the supervised fine-tuning, UPicker only needs a small amount of labeled data to achieve high accuracy in particle picking. To further enhance model performance, UPicker employs a contrastive denoising training strategy to reduce redundant detections and accelerate convergence, along with a hybrid data augmentation strategy to deal with limited labeled data. Comprehensive experiments on both simulated and experimental datasets demonstrate that UPicker outperforms state-of-the-art particle-picking methods in terms of accuracy and robustness while requiring fewer labeled data than other transformer-based models. Furthermore, ablation studies demonstrate the effectiveness and necessity of each component of UPicker. The source code and data are available at https://github.com/JachyLikeCoding/UPicker.
PMID:39658205 | DOI:10.1093/bib/bbae636
Deep learning reveals pathology-confirmed neuroimaging signatures in Alzheimer's, vascular and Lewy body dementias
Brain. 2024 Dec 9:awae388. doi: 10.1093/brain/awae388. Online ahead of print.
ABSTRACT
Concurrent neurodegenerative and vascular pathologies pose a diagnostic challenge in the clinical setting, with histopathology remaining the definitive modality for dementia-type diagnosis. To address this clinical challenge, we introduce a neuropathology-based, data-driven, multi-label deep learning framework to identify and quantify in-vivo biomarkers for Alzheimer's disease (AD), vascular dementia (VD), and Lewy body dementia (LBD) using antemortem T1-weighted MRI scans of 423 demented and 361 control participants from NACC and ADNI datasets. Based on the best-performing deep learning model, explainable heatmaps are extracted to visualize disease patterns, and the novel Deep Signature of Pathology Atrophy REcognition (DeepSPARE) indices are developed, where a higher DeepSPARE score indicates more brain alterations associated with that specific pathology. A substantial discrepancy in clinical and neuropathology diagnosis was observed in the demented patients: 71% of them had more than one pathology, but 67% of them were clinically diagnosed as AD only. Based on these neuropathology diagnoses and leveraging cross-validation principles, the deep learning model achieved the best performance with a balanced accuracy of 0.844, 0.839, and 0.623 for AD, VD, and LBD, respectively, and was used to generate the explainable deep-learning heatmaps and DeepSPARE indices. The explainable deep-learning heatmaps revealed distinct neuroimaging brain alteration patterns for each pathology: the AD heatmap highlighted bilateral hippocampal regions, the VD heatmap emphasized white matter regions, and the LBD heatmap exposed occipital alterations. The DeepSPARE indices were validated by examining their associations with cognitive testing, neuropathological, and neuroimaging measures using linear mixed-effects models. The DeepSPARE-AD index was associated with MMSE, Trail B, memory, PFDR-adjustedhippocampal volume, Braak stages, CERAD scores, and Thal phases (PFDR-adjusted < 0.05). The DeepSPARE-VD index was associated with white matter hyperintensity volume and cerebral amyloid angiopathy (PFDR-adjusted < 0.001). The DeepSPARE-LBD index was associated with Lewy body stages (PFDR-adjusted < 0.05). The findings were replicated in an out-of-sample ADNI dataset by testing associations with cognitive, imaging, plasma, and CSF measures. CSF and plasma pTau181 were significantly associated with DeepSPARE-AD in the AD/MCIΑβ+ group (PFDR-adjusted < 0.001), and CSF α-synuclein was associated solely with DeepSPARE-LBD (PFDR-adjusted = 0.036). Overall, these findings demonstrate the advantages of our innovative deep-learning framework in detecting antemortem neuroimaging signatures linked to different pathologies. The newly deep learning-derived DeepSPARE indices are precise, pathology-sensitive, and single-valued noninvasive neuroimaging metrics, bridging the traditional widely available in-vivo T1 imaging with histopathology.
PMID:39657969 | DOI:10.1093/brain/awae388
End-to-end deep learning patient level classification of affected territory of ischemic stroke patients in DW-MRI
Neuroradiology. 2024 Dec 10. doi: 10.1007/s00234-024-03520-x. Online ahead of print.
ABSTRACT
PURPOSE: To develop an end-to-end DL model for automated classification of affected territory in DWI of stroke patients.
MATERIALS AND METHODS: In this retrospective multicenter study, brain DWI studies from January 2017 to April 2020 from Center 1, from June 2020 to December 2020 from Center 2, and from November 2019 to April 2020 from Center 3 were included. Four radiologists labeled images into five classes: anterior cerebral artery (ACA), middle cerebral artery (MCA), posterior circulation (PC), and watershed (WS) regions, as well as normal images. Additionally, for Center 1, clinical information was encoded as a domain knowledge vector to incorporate into image embeddings. 3D convolutional neural network (CNN) and attention gate integrated versions for direct 3D encoding, long short-term memory (LSTM-CNN), and time-distributed layer for slice-based encoding were employed. Balanced classification accuracy, macro averaged f1 score, AUC, and interrater Cohen's kappa were calculated.
RESULTS: Overall, 624 DWI MRIs from 3 centers were utilized (mean age, interval: 66.89 years, 29-95 years; 345 male) with 439 patients in the training, 103 in the validation, and 82 in the test sets. The best model was a slice-based parallel encoding model with 0.88 balanced accuracy, 0.80 macro-f1 score, and an AUC of 0.98. Clinical domain knowledge integration improved the performance with 0.93 best overall accuracy with parallel stream model embeddings and support vector machine classifiers. The mean kappa value for interrater agreement was 0.87.
CONCLUSION: Developed end-to-end deep learning models performed well in classifying affected regions from stroke in DWI.
CLINICAL RELEVANCE STATEMENT: The end-to-end deep learning model with a parallel stream encoding strategy for classifying stroke regions in DWI has performed comparably with radiologists.
PMID:39656236 | DOI:10.1007/s00234-024-03520-x
Utilizing deep learning-based causal inference to explore vancomycin's impact on continuous kidney replacement therapy necessity in blood culture-positive intensive care unit patients
Microbiol Spectr. 2024 Dec 10:e0266224. doi: 10.1128/spectrum.02662-24. Online ahead of print.
ABSTRACT
Patients with positive blood cultures in the intensive care unit (ICU) are at high risk for septic acute kidney injury requiring continuous kidney replacement therapy (CKRT), especially when treated with vancomycin. This study developed a machine learning model to predict CKRT and examined vancomycin's impact using deep learning-based causal inference. We analyzed ICU patients with positive blood cultures, utilizing the Medical Information Mart for Intensive Care III data set. The primary outcome was defined as the initiation of CKRT during the ICU stay. The machine learning models were developed to predict the outcome. The deep learning-based causal inference model was utilized to quantitatively demonstrate the impact of vancomycin on the probability of CKRT initiation. Logistic regression was performed to analyze the relationship between the variables and the susceptibility of vancomycin. A total of 1,318 patients were included in the analysis, with 41 requiring CKRT. The Random Forest and Light Gradient Boosting Machine exhibited the best performance, with Area Under Curve of Receiver Operating Characteristic Curve values of 0.905 and 0.886, respectively. The deep learning-based causal inference model demonstrated an average 7.7% increase in the probability of CKRT occurrence when administrating vancomycin in total data set. Additionally, that younger age, lower diastolic blood pressure, higher heart rate, higher baseline creatinine, and lower bicarbonate levels sensitized the probability of CKRT application in response to vancomycin treatment. Deep learning-based causal inference models showed that vancomycin administration increases CKRT risk, identifying specific patient characteristics associated with higher susceptibility.IMPORTANCEThis study assesses the impact of vancomycin on the risk of continuous kidney replacement therapy (CKRT) in intensive care unit (ICU) patients with blood culture-positive infections. Utilizing deep learning-based causal inference and machine learning models, the research quantifies how vancomycin administration increases CKRT risk by an average of 7.7%. Key variables influencing susceptibility include baseline creatinine, diastolic blood pressure, heart rate, and bicarbonate levels. These findings offer insights into managing vancomycin-induced kidney risk and may inform patient-specific treatment strategies in ICU settings.
PMID:39656005 | DOI:10.1128/spectrum.02662-24
Rapid diagnosis of bacterial vaginosis using machine-learning-assisted surface-enhanced Raman spectroscopy of human vaginal fluids
mSystems. 2024 Dec 10:e0105824. doi: 10.1128/msystems.01058-24. Online ahead of print.
ABSTRACT
Bacterial vaginosis (BV) is an abnormal gynecological condition caused by the overgrowth of specific bacteria in the vagina. This study aims to develop a novel method for BV detection by integrating surface-enhanced Raman scattering (SERS) with machine learning (ML) algorithms. Vaginal fluid samples were classified as BV positive or BV negative using the BVBlue Test and clinical microscopy, followed by SERS spectral acquisition to construct the data set. Preliminary SERS spectral analysis revealed notable disparities in characteristic peak features. Multiple ML models were constructed and optimized, with the convolutional neural network (CNN) model achieving the highest prediction accuracy at 99%. Gradient-weighted class activation mapping (Grad-CAM) was used to highlight important regions in the images for prediction. Moreover, the CNN model was blindly tested on SERS spectra of vaginal fluid samples collected from 40 participants with unknown BV infection status, achieving a prediction accuracy of 90.75% compared with the results of the BVBlue Test combined with clinical microscopy. This novel technique is simple, cheap, and rapid in accurately diagnosing bacterial vaginosis, potentially complementing current diagnostic methods in clinical laboratories.
IMPORTANCE: The accurate and rapid diagnosis of bacterial vaginosis (BV) is crucial due to its high prevalence and association with serious health complications, including increased risk of sexually transmitted infections and adverse pregnancy outcomes. Although widely used, traditional diagnostic methods have significant limitations in subjectivity, complexity, and cost. The development of a novel diagnostic approach that integrates SERS with ML offers a promising solution. The CNN model's high prediction accuracy, cost-effectiveness, and extraordinary rapidity underscore its significant potential to enhance the diagnosis of BV in clinical settings. This method not only addresses the limitations of current diagnostic tools but also provides a more accessible and reliable option for healthcare providers, ultimately enhancing patient care and health outcomes.
PMID:39655908 | DOI:10.1128/msystems.01058-24
Systematic review of experimental paradigms and deep neural networks for electroencephalography-based cognitive workload detection
Prog Biomed Eng (Bristol). 2024 Oct 21;6(4). doi: 10.1088/2516-1091/ad8530.
ABSTRACT
This article summarizes a systematic literature review of deep neural network-based cognitive workload (CWL) estimation from electroencephalographic (EEG) signals. The focus of this article can be delineated into two main elements: first is the identification of experimental paradigms prevalently employed for CWL induction, and second, is an inquiry about the data structure and input formulations commonly utilized in deep neural networks (DNN)-based CWL detection. The survey revealed several experimental paradigms that can reliably induce either graded levels of CWL or a desired cognitive state due to sustained induction of CWL. This article has characterized them with respect to the number of distinct CWL levels, cognitive states, experimental environment, and agents in focus. Further, this literature analysis found that DNNs can successfully detect distinct levels of CWL despite the inter-subject and inter-session variability typically observed in EEG signals. Several methodologies were found using EEG signals in its native representation of a two-dimensional matrix as input to the classification algorithm, bypassing traditional feature selection steps. More often than not, researchers used DNNs as black-box type models, and only a few studies employed interpretable or explainable DNNs for CWL detection. However, these algorithms were mostly post hoc data analysis and classification schemes, and only a few studies adopted real-time CWL estimation methodologies. Further, it has been suggested that using interpretable deep learning methodologies may shed light on EEG correlates of CWL, but this remains mostly an unexplored area. This systematic review suggests using networks sensitive to temporal dependencies and appropriate input formulations for each type of DNN architecture to achieve robust classification performance. An additional suggestion is to utilize transfer learning methods to achieve high generalizability across tasks (task-independent classifiers), while simple cross-subject data pooling may achieve the same for subject-independent classifiers.
PMID:39655862 | DOI:10.1088/2516-1091/ad8530
Ultrasound imaging based recognition of prenatal anomalies: a systematic clinical engineering review
Prog Biomed Eng (Bristol). 2024 May 7;6(2). doi: 10.1088/2516-1091/ad3a4b.
ABSTRACT
For prenatal screening, ultrasound (US) imaging allows for real-time observation of developing fetal anatomy. Understanding normal and aberrant forms through extensive fetal structural assessment enables for early detection and intervention. However, the reliability of anomaly diagnosis varies depending on operator expertise and device limits. First trimester scans in conjunction with circulating biochemical markers are critical in identifying high-risk pregnancies, but they also pose technical challenges. Recent engineering advancements in automated diagnosis, such as artificial intelligence (AI)-based US image processing and multimodal data fusion, are developing to improve screening efficiency, accuracy, and consistency. Still, creating trust in these data-driven solutions is necessary for integration and acceptability in clinical settings. Transparency can be promoted by explainable AI (XAI) technologies that provide visual interpretations and illustrate the underlying diagnostic decision making process. An explanatory framework based on deep learning is suggested to construct charts depicting anomaly screening results from US video feeds. AI modelling can then be applied to these charts to connect defects with probable deformations. Overall, engineering approaches that increase imaging, automation, and interpretability hold enormous promise for altering traditional workflows and expanding diagnostic capabilities for better prenatal care.
PMID:39655845 | DOI:10.1088/2516-1091/ad3a4b
Machine Learning-Based Prediction for In-Hospital Mortality After Acute Intracerebral Hemorrhage Using Real-World Clinical and Image Data
J Am Heart Assoc. 2024 Dec 10:e036447. doi: 10.1161/JAHA.124.036447. Online ahead of print.
ABSTRACT
BACKGROUND: Machine learning (ML) techniques are widely employed across various domains to achieve accurate predictions. This study assessed the effectiveness of ML in predicting early mortality risk among patients with acute intracerebral hemorrhage (ICH) in real-world settings.
METHODS AND RESULTS: ML-based models were developed to predict in-hospital mortality in 527 patients with ICH using raw brain imaging data from brain computed tomography and clinical data. The models' performances were evaluated using the area under the receiver operating characteristic curves and calibration plots, comparing them with traditional risk scores such as the ICH score and ICH grading scale. Kaplan-Meier curves were used to examine the post-ICH survival rates, stratified by ML-based risk assessment. The net benefit of ML-based models was evaluated using decision curve analysis. The area under the receiver operating characteristic curves were 0.91 (95% CI, 0.86-0.95) for the ICH score, 0.93 (95% CI, 0.89-0.97) for the ICH grading scale, 0.83 (95% CI, 0.71-0.91) for the ML-based model fitted with raw image data only, and 0.87 (95% CI, 0.76-0.93) for the ML-based model fitted using clinical data without specialist expertise. The area under the receiver operating characteristic curve increased significantly to 0.97 (95% CI, 0.94-0.99) when the ML model was fitted using clinical and image data assessed by specialists. All ML-based models demonstrated good calibration, and the survival rates showed significant differences between risk groups. Decision curve analysis indicated the highest net benefit when utilizing the findings assessed by specialists.
CONCLUSIONS: ML-based prediction models exhibit satisfactory performance in predicting post-ICH in-hospital mortality when utilizing raw imaging data or nonspecialist input. Nevertheless, incorporating specialist expertise notably improves performance.
PMID:39655759 | DOI:10.1161/JAHA.124.036447
Assessment of image quality on the diagnostic performance of clinicians and deep learning models: Cross-sectional comparative reader study
J Eur Acad Dermatol Venereol. 2024 Dec 10. doi: 10.1111/jdv.20462. Online ahead of print.
ABSTRACT
BACKGROUND: Skin cancer is a prevalent and clinically significant condition, with early and accurate diagnosis being crucial for improved patient outcomes. Dermoscopy and artificial intelligence (AI) hold promise in enhancing diagnostic accuracy. However, the impact of image quality, particularly high dynamic range (HDR) conversion in smartphone images, on diagnostic performance remains poorly understood.
OBJECTIVE: This study aimed to investigate the effect of varying image qualities, including HDR-enhanced dermoscopic images, on the diagnostic capabilities of clinicians and a convolutional neural network (CNN) model.
METHODS: Eighteen dermatology clinicians assessed 303 images of 101 skin lesions that were categorized into three image quality groups: low quality (LQ), high quality (HQ) and enhanced quality (EQ) produced using HDR-style conversion. Clinicians participated in a two part reader study that required their diagnosis, management and confidence level for each image assessed.
RESULTS: In the binary classification of lesions, clinicians had the greatest diagnostic performance with HQ images, with sensitivity (77.3%; CI 69.1-85.5), specificity (63.1%; CI 53.7-72.5) and accuracy (70.2%; CI 61.3-79.1). For the multiclass classification, the overall performance was also best with HQ images, attaining the greatest specificity (91.9%; CI 83.2-95.0) and accuracy (51.5%; CI 48.4-54.7). Clinicians had a superior performance (median correct diagnoses) to the CNN model for the binary classification of LQ and EQ images, but their performance was comparable on the HQ images. However, in the multiclass classification, the CNN model significantly outperformed the clinicians on HQ images (p < 0.01).
CONCLUSION: This study highlights the importance of image quality on the diagnostic performance of clinicians and deep learning models. This has significant implications for telehealth reporting and triage.
PMID:39655640 | DOI:10.1111/jdv.20462
Automatic classification of fungal-fungal interactions using deep leaning models
Comput Struct Biotechnol J. 2024 Nov 14;23:4222-4231. doi: 10.1016/j.csbj.2024.11.027. eCollection 2024 Dec.
ABSTRACT
Fungi provide valuable solutions for diverse biotechnological applications, such as enzymes in the food industry, bioactive metabolites for healthcare, and biocontrol organisms in agriculture. Current workflows for identifying new biocontrol fungi often rely on subjective visual observations of strains' performance in microbe-microbe interaction studies, making the process time-consuming and difficult to reproduce. To overcome these challenges, we developed an AI-automated image classification approach using machine learning algorithm based on deep neural network. Our method focuses on analyzing standardized images of 96-well microtiter plates with solid medium for fungal-fungal challenge experiments. We used our model to categorize the outcome of interactions between the plant pathogen Fusarium graminearum and individual isolates from a collection of 38,400 fungal strains. The authors trained multiple deep learning architectures and evaluated their performance. The results strongly support our approach, achieving a peak accuracy of 95.0 % with the DenseNet121 model and a maximum macro-averaged F1-Score of 93.1 across five folds. To the best of our knowledge, this paper introduces the first automated method for classifying fungal-fungal interactions using deep learning, which can easily be adapted for other fungal species.
PMID:39655263 | PMC:PMC11626056 | DOI:10.1016/j.csbj.2024.11.027
Focused review on artificial intelligence for disease detection in infants
Front Digit Health. 2024 Nov 25;6:1459640. doi: 10.3389/fdgth.2024.1459640. eCollection 2024.
ABSTRACT
Over the last years, studies using artificial intelligence (AI) for the detection and prediction of diseases have increased and also concentrated more and more on vulnerable groups of individuals, such as infants. The release of ChatGPT demonstrated the potential of large language models (LLMs) and heralded a new era of AI with manifold application possibilities. However, the impact of this new technology on medical research cannot be fully estimated yet. In this work, we therefore aimed to summarise the most recent pre-ChatGPT developments in the field of automated detection and prediction of diseases and disease status in infants, i.e., within the first 12 months of life. For this, we systematically searched the scientific databases PubMed and IEEE Xplore for original articles published within the last five years preceding the release of ChatGPT (2018-2022). The search revealed 927 articles; a final number of 154 articles was included for review. First of all, we examined research activity over time. Then, we analysed the articles from 2022 for medical conditions, data types, tasks, AI approaches, and reported model performance. A clear trend of increasing research activity over time could be observed. The most recently published articles focused on medical conditions of twelve different ICD-11 categories; "certain conditions originating in the perinatal period" was the most frequently addressed disease category. AI models were trained with a variety of data types, among which clinical and demographic information and laboratory data were most frequently exploited. The most frequently performed tasks aimed to detect present diseases, followed by the prediction of diseases and disease status at a later point in development. Deep neural networks turned out as the most popular AI approach, even though traditional methods, such as random forests and support vector machines, still play a role-presumably due to their explainability or better suitability when the amount of data is limited. Finally, the reported performances in many of the reviewed articles suggest that AI has the potential to assist in diagnostic procedures for infants in the near future. LLMs will boost developments in this field in the upcoming years.
PMID:39654981 | PMC:PMC11625793 | DOI:10.3389/fdgth.2024.1459640
LesionScanNet: dual-path convolutional neural network for acute appendicitis diagnosis
Health Inf Sci Syst. 2024 Dec 7;13(1):3. doi: 10.1007/s13755-024-00321-7. eCollection 2025 Dec.
ABSTRACT
Acute appendicitis is an abrupt inflammation of the appendix, which causes symptoms such as abdominal pain, vomiting, and fever. Computed tomography (CT) is a useful tool in accurate diagnosis of acute appendicitis; however, it causes challenges due to factors such as the anatomical structure of the colon and localization of the appendix in CT images. In this paper, a novel Convolutional Neural Network model, namely, LesionScanNet for the computer-aided detection of acute appendicitis has been proposed. For this purpose, a dataset of 2400 CT scan images was collected by the Department of General Surgery at Kanuni Sultan Süleyman Research and Training Hospital, Istanbul, Turkey. LesionScanNet is a lightweight model with 765 K parameters and includes multiple DualKernel blocks, where each block contains a convolution, expansion, separable convolution layers, and skip connections. The DualKernel blocks work with two paths of input image processing, one of which uses 3 × 3 filters, and the other path encompasses 1 × 1 filters. It has been demonstrated that the LesionScanNet model has an accuracy score of 99% on the test set, a value that is greater than the performance of the benchmark deep learning models. In addition, the generalization ability of the LesionScanNet model has been demonstrated on a chest X-ray image dataset for pneumonia and COVID-19 detection. In conclusion, LesionScanNet is a lightweight and robust network achieving superior performance with smaller number of parameters and its usage can be extended to other medical application domains.
PMID:39654693 | PMC:PMC11625030 | DOI:10.1007/s13755-024-00321-7
Integrating artificial intelligence in strabismus management: current research landscape and future directions
Exp Biol Med (Maywood). 2024 Nov 25;249:10320. doi: 10.3389/ebm.2024.10320. eCollection 2024.
ABSTRACT
Advancements in artificial intelligence (AI) are transforming strabismus management through improved screening, diagnosis, and surgical planning. Deep learning has notably enhanced diagnostic accuracy and optimized surgical outcomes. Despite these advancements, challenges such as the underrepresentation of diverse strabismus types and reliance on single-source data remain prevalent. Emphasizing the need for inclusive AI systems, future research should focus on expanding AI capabilities with large model technologies, integrating multimodal data to bridge existing gaps, and developing integrated management platforms to better accommodate diverse patient demographics and clinical scenarios.
PMID:39654660 | PMC:PMC11625544 | DOI:10.3389/ebm.2024.10320
The role of deep learning in myocardial perfusion imaging for diagnosis and prognosis: A systematic review
iScience. 2024 Nov 12;27(12):111374. doi: 10.1016/j.isci.2024.111374. eCollection 2024 Dec 20.
ABSTRACT
The development of state-of-the-art algorithms for computer visualization has led to a growing interest in applying deep learning (DL) techniques to the field of medical imaging. DL-based algorithms have been extensively utilized in various aspects of cardiovascular imaging, and one notable area of focus is single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI), which is regarded as the gold standard for non-invasive diagnosis of myocardial ischemia. However, due to the complex decision-making process of DL based on convolutional neural networks (CNNs), the explainability of DL results has become a significant area of research, particularly in the field of medical imaging. To better harness the potential of DL and to be well prepared for the ongoing DL revolution in nuclear imaging, this review aims to summarize the recent applications of DL in MPI, with a specific emphasis on the methods in explainable DL for the diagnosis and prognosis of MPI. Furthermore, the challenges and potential directions for future research are also discussed.
PMID:39654634 | PMC:PMC11626733 | DOI:10.1016/j.isci.2024.111374
Toward explainable deep learning in healthcare through transition matrix and user-friendly features
Front Artif Intell. 2024 Nov 25;7:1482141. doi: 10.3389/frai.2024.1482141. eCollection 2024.
ABSTRACT
Modern artificial intelligence (AI) solutions often face challenges due to the "black box" nature of deep learning (DL) models, which limits their transparency and trustworthiness in critical medical applications. In this study, we propose and evaluate a scalable approach based on a transition matrix to enhance the interpretability of DL models in medical signal and image processing by translating complex model decisions into user-friendly and justifiable features for healthcare professionals. The criteria for choosing interpretable features were clearly defined, incorporating clinical guidelines and expert rules to align model outputs with established medical standards. The proposed approach was tested on two medical datasets: electrocardiography (ECG) for arrhythmia detection and magnetic resonance imaging (MRI) for heart disease classification. The performance of the DL models was compared with expert annotations using Cohen's Kappa coefficient to assess agreement, achieving coefficients of 0.89 for the ECG dataset and 0.80 for the MRI dataset. These results demonstrate strong agreement, underscoring the reliability of the approach in providing accurate, understandable, and justifiable explanations of DL model decisions. The scalability of the approach suggests its potential applicability across various medical domains, enhancing the generalizability and utility of DL models in healthcare while addressing practical challenges and ethical considerations.
PMID:39654544 | PMC:PMC11625760 | DOI:10.3389/frai.2024.1482141
DTMP-prime: A deep transformer-based model for predicting prime editing efficiency and PegRNA activity
Mol Ther Nucleic Acids. 2024 Oct 28;35(4):102370. doi: 10.1016/j.omtn.2024.102370. eCollection 2024 Dec 10.
ABSTRACT
Prime editors are CRISPR-based genome engineering tools with significant potential for rectifying patient mutations. However, their usage requires experimental optimization of the prime editing guide RNA (PegRNA) to achieve high editing efficiency. This paper introduces the deep transformer-based model for predicting prime editing efficiency (DTMP-Prime), a tool specifically designed to predict PegRNA activity and prime editing (PE) efficiency. DTMP-Prime facilitates the design of appropriate PegRNA and ngRNA. A transformer-based model was constructed to scrutinize a wide-ranging set of PE data, enabling the extraction of effective features of PegRNAs and target DNA sequences. The integration of these features with the proposed encoding strategy and DNABERT-based embedding has notably improved the predictive capabilities of DTMP-Prime for off-target sites. Moreover, DTMP-Prime is a promising tool for precisely predicting off-target sites in CRISPR experiments. The integration of a multi-head attention framework has additionally improved the precision and generalizability of DTMP-Prime across various PE models and cell lines. Evaluation results based on the Pearson and Spearman correlation coefficient demonstrate that DTMP-Prime outperforms other state-of-the-art models in predicting the efficiency and outcomes of PE experiments.
PMID:39654539 | PMC:PMC11626815 | DOI:10.1016/j.omtn.2024.102370
Artificial intelligence applications in smile design dentistry: A scoping review
J Prosthodont. 2024 Dec 9. doi: 10.1111/jopr.14000. Online ahead of print.
ABSTRACT
PURPOSE: Artificial intelligence (AI) applications are growing in smile design and aesthetic procedures. The current expansion and performance of AI models in digital smile design applications have not yet been systematically documented and analyzed. The purpose of this review was to assess the performance of AI models in smile design, assess the criteria of points of reference using AI analysis, and assess different AI software performance.
METHODS: An electronic review was completed in five databases: MEDLINE/PubMed, EMBASE, World of Science, Cochrane, and Scopus. Studies that developed AI models for smile design were included. The search strategy included articles published until November 1, 2024. Two investigators independently evaluated the quality of the studies by applying the Joanna Briggs Institute (JBI) Critical Appraisal Checklist for Quasi-Experimental Studies and Textual Evidence: Expert Opinion Results.
RESULTS: The search resulted in 2653 articles. A total of 2649 were excluded according to the exclusion criteria after reading the title, abstract, and/or full-text review. Four articles published between 2023 and 2024 were included in the present investigation. Two articles compared 2D and 3D points while one article compared the outcome of satisfaction between dentists and patients, and the last article emphasized the ethical components of using AI.
CONCLUSION: The results of the studies reviewed in this paper suggest that AI-generated smile designs are not significantly different from manually created designs in terms of esthetic perception. 3D designs are more accurate than 2D designs and offer more advantages. More articles are needed in the field of AI and smile design.
PMID:39654301 | DOI:10.1111/jopr.14000