Deep learning

Development and Validation of a Deep Learning System to Differentiate HER2-Zero, HER2-Low, and HER2-Positive Breast Cancer Based on Dynamic Contrast-Enhanced MRI

Fri, 2024-12-06 06:00

J Magn Reson Imaging. 2024 Dec 6. doi: 10.1002/jmri.29670. Online ahead of print.

ABSTRACT

BACKGROUND: Previous studies explored MRI-based radiomic features for differentiating between human epidermal growth factor receptor 2 (HER2)-zero, HER2-low, and HER2-positive breast cancer, but deep learning's effectiveness is uncertain.

PURPOSE: This study aims to develop and validate a deep learning system using dynamic contrast-enhanced MRI (DCE-MRI) for automated tumor segmentation and classification of HER2-zero, HER2-low, and HER2-positive statuses.

STUDY TYPE: Retrospective.

POPULATION: One thousand two hundred ninety-four breast cancer patients from three centers who underwent DCE-MRI before surgery were included in the study (52 ± 11 years, 811/204/279 for training/internal testing/external testing).

FIELD STRENGTH/SEQUENCE: 3 T scanners, using T1-weighted 3D fast spoiled gradient-echo sequence, T1-weighted 3D enhanced fast gradient-echo sequence and T1-weighted turbo field echo sequence.

ASSESSMENT: An automated model segmented tumors utilizing DCE-MRI data, followed by a deep learning models (ResNetGN) trained to classify HER2 statuses. Three models were developed to distinguish HER2-zero, HER2-low, and HER2-positive from their respective non-HER2 categories.

STATISTICAL TESTS: Dice similarity coefficient (DSC) was used to evaluate the segmentation performance of the model. Evaluation of the model performances for HER2 statuses involved receiver operating characteristic (ROC) curve analysis and the area under the curve (AUC), accuracy, sensitivity, and specificity. The P-values <0.05 were considered statistically significant.

RESULTS: The automatic segmentation network achieved DSC values of 0.85 to 0.90 compared to the manual segmentation across different sets. The deep learning models using ResNetGN achieved AUCs of 0.782, 0.776, and 0.768 in differentiating HER2-zero from others in the training, internal test, and external test sets, respectively. Similarly, AUCs of 0.820, 0.813, and 0.787 were achieved for HER2-low vs. others, and 0.792, 0.745, and 0.781 for HER2-positive vs. others, respectively.

DATA CONCLUSION: The proposed DCE-MRI-based deep learning system may have the potential to preoperatively distinct HER2 expressions of breast cancers with therapeutic implications.

EVIDENCE LEVEL: 4 TECHNICAL EFFICACY: Stage 3.

PMID:39643475 | DOI:10.1002/jmri.29670

Categories: Literature Watch

Deep Learning for Contrast Enhanced Mammography - A Systematic Review

Fri, 2024-12-06 06:00

Acad Radiol. 2024 Dec 5:S1076-6332(24)00881-X. doi: 10.1016/j.acra.2024.11.035. Online ahead of print.

ABSTRACT

BACKGROUND/AIM: Contrast-enhanced mammography (CEM) is a relatively novel imaging technique that enables both anatomical and functional breast imaging, with improved diagnostic performance compared to standard 2D mammography. The aim of this study is to systematically review the literature on deep learning (DL) applications for CEM, exploring how these models can further enhance CEM diagnostic potential.

METHODS: This systematic review was reported according to the PRISMA guidelines. We searched for studies published up to April 2024. MEDLINE, Scopus and Google Scholar were used as search databases. Two reviewers independently implemented the search strategy. We included all types of original studies published in English that evaluated DL algorithms for automatic analysis of contrast-enhanced mammography CEM images. The quality of the studies was independently evaluated by two reviewers based on the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria.

RESULTS: Sixteen relevant studies published between 2018 and 2024 were identified. All but one used convolutional neural network models (CNN) models. All studies evaluated DL algorithms for lesion classification, while six studies also assessed lesion detection or segmentation. Segmentation was performed manually in three studies, both manually and automatically in two studies and automatically in ten studies. For lesion classification on retrospective datasets, CNN models reported varied areas under the curve (AUCs) ranging from 0.53 to 0.99. Models incorporating attention mechanism achieved accuracies of 88.1% and 89.1%. Prospective studies reported AUC values of 0.89 and 0.91. Some studies demonstrated that combining DL models with radiomics featured improved classification. Integrating DL algorithms with radiologists' assessments enhanced diagnostic performance.

CONCLUSION: While still at an early research stage, DL can improve CEM diagnostic precision. However, there is a relatively small number of studies evaluating different DL algorithms, and most studies are retrospective. Further prospective testing to assess performance of applications at actual clinical setting is warranted.

PMID:39643464 | DOI:10.1016/j.acra.2024.11.035

Categories: Literature Watch

Predictive utility of artificial intelligence on schizophrenia treatment outcomes: A systematic review and meta-analysis

Fri, 2024-12-06 06:00

Neurosci Biobehav Rev. 2024 Dec 4:105968. doi: 10.1016/j.neubiorev.2024.105968. Online ahead of print.

ABSTRACT

Identifying optimal treatment approaches for schizophrenia is challenging due to varying symptomatology and treatment responses. Artificial intelligence (AI) shows promise in predicting outcomes, prompting this systematic review and meta-analysis to evaluate various AI models' predictive utilities in schizophrenia treatment. A systematic search was conducted, and the risk of bias was evaluated. The pooled sensitivity, specificity, and diagnostic odds ratio with 95% confidence intervals between AI models and the reference standard for response to treatment were assessed. Diagnostic accuracy measures were calculated, and subgroup analysis was performed based on the input data of AI models. Out of the 21 included studies, AI models achieved a pooled sensitivity of 70% and specificity of 76% in predicting schizophrenia treatment response with substantial predictive capacity and a near-to-high level of test accuracy. Subgroup analysis revealed EEG-based models to have the highest sensitivity (89%) and specificity (94%), followed by imaging-based models (76% and 80%, respectively). However, significant heterogeneity was observed across studies in treatment response definitions, participant characteristics, and therapeutic interventions. Despite methodological variations and small sample sizes in some modalities, this study underscores AI's predictive utility in schizophrenia treatment, offering insights for tailored approaches, improving adherence, and reducing relapse risk.

PMID:39643220 | DOI:10.1016/j.neubiorev.2024.105968

Categories: Literature Watch

Using Machine Learning for Personalized Prediction of Longitudinal COVID-19 Vaccine Responses in Transplant Recipients

Fri, 2024-12-06 06:00

Am J Transplant. 2024 Dec 4:S1600-6135(24)00746-9. doi: 10.1016/j.ajt.2024.11.033. Online ahead of print.

ABSTRACT

The COVID-19 pandemic has underscored the importance of vaccines, especially for immunocompromised populations like solid organ transplant (SOT) recipients, who often have weaker immune responses. The purpose of this study was to compare deep learning architectures for predicting SARS-CoV-2 vaccine responses 12 months post-vaccination in this high-risk group. Utilizing data from 303 SOT recipients from a Canadian multicenter cohort, models were developed to forecast anti-receptor binding domain (RBD) antibody levels. The study compared traditional machine learning models-logistic regression, epsilon-support vector regression, random forest regressor, and gradient boosting regressor-and deep learning architectures, including long short-term memory (LSTM), recurrent neural networks, and a novel model, routed LSTM. This new model combines capsule networks with LSTM to reduce the need for large datasets. Demographic, clinical, and transplant-specific data, along with longitudinal antibody measurements, were incorporated into the models. The routed LSTM performed best, achieving a mean square error (MSE) of 0.02±0.02 and a Pearson correlation coefficient (PCC) of 0.79±0.24, outperforming all other models. Key factors influencing vaccine response included age, immunosuppression, breakthrough infection, BMI, sex, and transplant type. These findings suggest that AI could be a valuable tool in tailoring vaccine strategies, improving health outcomes for vulnerable transplant recipients.

PMID:39643006 | DOI:10.1016/j.ajt.2024.11.033

Categories: Literature Watch

Toward automated detection of microbleeds with anatomical scale localization using deep learning

Fri, 2024-12-06 06:00

Med Image Anal. 2024 Nov 30;101:103415. doi: 10.1016/j.media.2024.103415. Online ahead of print.

ABSTRACT

Cerebral Microbleeds (CMBs) are chronic deposits of small blood products in the brain tissues, which have explicit relation to various cerebrovascular diseases depending on their anatomical location, including cognitive decline, intracerebral hemorrhage, and cerebral infarction. However, manual detection of CMBs is a time consuming and error-prone process because of their sparse and tiny structural properties. The detection of CMBs is commonly affected by the presence of many CMB mimics that cause a high false-positive rate (FPR), such as calcifications and pial vessels. This paper proposes a novel 3D deep learning framework that not only detects CMBs but also identifies their anatomical location in the brain (i.e., lobar, deep, and infratentorial regions). For the CMBs detection task, we propose a single end-to-end model by leveraging the 3D U-Net as a backbone with Region Proposal Network (RPN). To significantly reduce the false positives within the same single model, we develop a new scheme, containing Feature Fusion Module (FFM) that detects small candidates utilizing contextual information and Hard Sample Prototype Learning (HSPL) that mines CMB mimics and generates additional loss term called concentration loss using Convolutional Prototype Learning (CPL). For the anatomical localization task, we exploit the 3D U-Net segmentation network to segment anatomical structures of the brain. This task not only identifies to which region the CMBs belong but also eliminates some false positives from the detection task by leveraging anatomical information. We utilize Susceptibility-Weighted Imaging (SWI) and phase images as 3D input to efficiently capture 3D information. The results show that the proposed RPN that utilizes the FFM and HSPL outperforms the baseline RPN and achieves a sensitivity of 94.66 % vs. 93.33 % and an average number of false positives per subject (FPavg) of 0.86 vs. 14.73. Furthermore, the anatomical localization task enhances the detection performance by reducing the FPavg to 0.56 while maintaining the sensitivity of 94.66 %.

PMID:39642804 | DOI:10.1016/j.media.2024.103415

Categories: Literature Watch

Revolutionizing cesium monitoring in seawater through electrochemical voltammetry and machine learning

Fri, 2024-12-06 06:00

J Hazard Mater. 2024 Nov 28;484:136558. doi: 10.1016/j.jhazmat.2024.136558. Online ahead of print.

ABSTRACT

Monitoring radioactive cesium ions (Cs+) in seawater is vital for environmental safety but remains challenging due to limitations in the accessibility, stability, and selectivity of traditional methods. This study presents an innovative approach that combines electrochemical voltammetry using nickel hexacyanoferrate (NiHCF) thin-film electrode with machine learning (ML) to enable accurate and portable detection of Cs+. Optimizing the fabrication of NiHCF thin-film electrodes enabled the development of a robust sensor that generates cyclic voltammograms (CVs) sensitive to Cs⁺ concentrations as low as 1 ppb in synthetic seawater and 10 ppb in real seawater, with subtle changes in CV patterns caused by trace Cs⁺ effectively identified and analyzed using ML. Using 2D convolutional neural networks (CNNs), we classified Cs+ concentrations across eight logarithmic classes (0 - 106 ppb) with 100 % accuracy and an F1-score of 1 in synthetic seawater datasets, outperforming the 1D CNN and deep neural networks. Validation using real seawater datasets confirmed the applicability of our model, achieving high performance. Moreover, gradient-weighted class activation mapping (Grad-CAM) identified critical CV regions that were overlooked during manual inspection, validating model reliability. This integrated method offers sensitive and practical solutions for monitoring Cs+ in seawater, helping to prevent its accumulation in ecosystems.

PMID:39642734 | DOI:10.1016/j.jhazmat.2024.136558

Categories: Literature Watch

Challenges and solutions of deep learning-based automated liver segmentation: A systematic review

Fri, 2024-12-06 06:00

Comput Biol Med. 2024 Dec 5;185:109459. doi: 10.1016/j.compbiomed.2024.109459. Online ahead of print.

ABSTRACT

The liver is one of the vital organs in the body. Precise liver segmentation in medical images is essential for liver disease treatment. The deep learning-based liver segmentation process faces several challenges. This research aims to analyze the challenges of liver segmentation in prior studies and identify the modifications made to network models and other enhancements implemented by researchers to tackle each challenge. In total, 88 articles from Scopus and ScienceDirect databases published between January 2016 and January 2022 have been studied. The liver segmentation challenges are classified into five main categories, each containing some subcategories. For each challenge, the proposed technique to overcome the challenge is investigated. The provided report details the authors, publication years, dataset types, imaging technologies, and evaluation metrics of all references for comparison. Additionally, a summary table outlines the challenges and solutions.

PMID:39642700 | DOI:10.1016/j.compbiomed.2024.109459

Categories: Literature Watch

Learning soft tissue deformation from incremental simulations

Fri, 2024-12-06 06:00

Med Phys. 2024 Dec 6. doi: 10.1002/mp.17554. Online ahead of print.

ABSTRACT

BACKGROUND: Surgical planning for orthognathic procedures demands swift and accurate biomechanical modeling of facial soft tissues. Efficient simulations are vital in the clinical pipeline, as surgeons may iterate through multiple plans. Biomechanical simulations typically use the finite element method (FEM). Prior works divide FEM simulations into increments to enhance convergence and accuracy. However, this practice elongates simulation time, thereby impeding clinical integration. To accelerate simulations, deep learning (DL) models have been explored. Yet, previous efforts either perform simulations in a single step or neglect the temporal aspects in incremental simulations.

PURPOSE: This study investigates the use of spatiotemporal incremental modeling for biomechanics simulations of facial soft tissue.

METHODS: We implement the method using a graph neural network. Our method synergizes spatial features with temporal aggregation using DL networks trained on incremental FEM simulations from 17 subjects that underwent orthognathic surgery.

RESULTS: Our proposed spatiotemporal incremental method achieved a mean accuracy of 0.37 mm with a mean computation time of 1.52 s. In comparison, a spatial-only incremental method yielded a mean accuracy of 0.44 mm and a mean computation time of 1.60 s, while a spatial-only single-step method yielded a mean accuracy of 0.41 mm and a mean computation time of 0.05 s.

CONCLUSIONS: Statistical analysis demonstrated that the spatiotemporal incremental method reduced mean errors compared to the spatial-only incremental method, emphasizing the importance of incorporating temporal information in incremental simulations. Overall, we successfully implemented spatiotemporal incremental learning tailored to simulate soft tissue deformation while substantially reducing simulation time compared to FEM.

PMID:39642013 | DOI:10.1002/mp.17554

Categories: Literature Watch

Magnetic resonance image denoising for Rician noise using a novel hybrid transformer-CNN network (HTC-net) and self-supervised pretraining

Fri, 2024-12-06 06:00

Med Phys. 2024 Dec 6. doi: 10.1002/mp.17562. Online ahead of print.

ABSTRACT

BACKGROUND: Magnetic resonance imaging (MRI) is a crucial technique for both scientific research and clinical diagnosis. However, noise generated during MR data acquisition degrades image quality, particularly in hyperpolarized (HP) gas MRI. While deep learning (DL) methods have shown promise for MR image denoising, most of them fail to adequately utilize the long-range information which is important to improve denoising performance. Furthermore, the sample size of paired noisy and noise-free MR images also limits denoising performance.

PURPOSE: To develop an effective DL method that enhances denoising performance and reduces the requirement of paired MR images by utilizing the long-range information and pretraining.

METHODS: In this work, a hybrid Transformer-convolutional neural network (CNN) network (HTC-net) and a self-supervised pretraining strategy are proposed, which effectively enhance the denoising performance. In HTC-net, a CNN branch is exploited to extract the local features. Then a Transformer-CNN branch with two parallel encoders is designed to capture the long-range information. Within this branch, a residual fusion block (RFB) with a residual feature processing module and a feature fusion module is proposed to aggregate features at different resolutions extracted by two parallel encoders. After that, HTC-net exploits the comprehensive features from the CNN branch and the Transformer-CNN branch to accurately predict noise-free MR images through a reconstruction module. To further enhance the performance on limited MRI datasets, a self-supervised pretraining strategy is proposed. This strategy employs self-supervised denoising to equip the HTC-net with denoising capabilities during pretraining, and then the pre-trained parameters are transferred to facilitate subsequent supervised training.

RESULTS: Experimental results on the pulmonary HP 129Xe MRI dataset (1059 images) and IXI dataset (5000 images) all demonstrate the proposed method outperforms the state-of-the-art methods, exhibiting superior preservation of edges and structures. Quantitatively, on the pulmonary HP 129Xe MRI dataset, the proposed method outperforms the state-of-the-art methods by 0.254-0.597 dB in PSNR and 0.007-0.013 in SSIM. On the IXI dataset, the proposed method outperforms the state-of-the-art methods by 0.3-0.927 dB in PSNR and 0.003-0.016 in SSIM.

CONCLUSIONS: The proposed method can effectively enhance the quality of MR images, which helps improve the diagnosis accuracy in clinical.

PMID:39641989 | DOI:10.1002/mp.17562

Categories: Literature Watch

Estimation of fatty acid composition in mammary adipose tissue using deep neural network with unsupervised training

Fri, 2024-12-06 06:00

Magn Reson Med. 2024 Dec 6. doi: 10.1002/mrm.30401. Online ahead of print.

ABSTRACT

PURPOSE: To develop a deep learning-based method for robust and rapid estimation of the fatty acid composition (FAC) in mammary adipose tissue.

METHODS: A physics-based unsupervised deep learning network for estimation of fatty acid composition-network (FAC-Net) is proposed to estimate the number of double bonds and number of methylene-interrupted double bonds from multi-echo bipolar gradient-echo data, which are subsequently converted to saturated, mono-unsaturated, and poly-unsaturated fatty acids. The loss function was based on a 10 fat peak signal model. The proposed network was tested with a phantom containing eight oils with different FAC and on post-menopausal women scanned using a whole-body 3T MRI system between February 2022 and January 2024. The post-menopausal women included a control group (n = 8) with average risk for breast cancer and a cancer group (n = 7) with biopsy-proven breast cancer.

RESULTS: The FAC values of eight oils in the phantom showed strong correlations between the measured and reference values (R2 > 0.9 except chain length). The FAC values measured from scan and rescan data of the control group showed no significant difference between the two scans. The FAC measurements of the cancer group conducted before contrast and after contrast showed a significant difference in saturated fatty acid and mono-unsaturated fatty acid. The cancer group has higher saturated fatty acid than the control group, although not statistically significant.

CONCLUSION: The results in this study suggest that the proposed FAC-Net can be used to measure the FAC of mammary adipose tissue from gradient-echo MRI data of the breast.

PMID:39641987 | DOI:10.1002/mrm.30401

Categories: Literature Watch

[PSI]-CIC: A Deep-Learning Pipeline for the Annotation of Sectored Saccharomyces cerevisiae Colonies

Fri, 2024-12-06 06:00

Bull Math Biol. 2024 Dec 6;87(1):12. doi: 10.1007/s11538-024-01379-w.

ABSTRACT

The [ P S I + ] prion phenotype in yeast manifests as a white, pink, or red color pigment. Experimental manipulations destabilize prion phenotypes, and allow colonies to exhibit [ p s i - ] (red) sectored phenotypes within otherwise completely white colonies. Further investigation of the size and frequency of sectors that emerge as a result of experimental manipulation is capable of providing critical information on mechanisms of prion curing, but we lack a way to reliably extract this information. Images of experimental colonies exhibiting sectored phenotypes offer an abundance of data to help uncover molecular mechanisms of sectoring, yet the structure of sectored colonies is ignored in traditional biological pipelines. In this study, we present [PSI]-CIC, the first computational pipeline designed to identify and characterize features of sectored yeast colonies. To overcome the barrier of a lack of manually annotated data of colonies, we develop a neural network architecture that we train on synthetic images of colonies and apply to real images of [ P S I + ] , [ p s i - ] , and sectored colonies. In hand-annotated experimental images, our pipeline correctly predicts the state of approximately 95% of colonies detected and frequency of sectors in approximately 89.5% of colonies detected. The scope of our pipeline could be extended to categorizing colonies grown under different experimental conditions, allowing for more meaningful and detailed comparisons between experiments. Our approach streamlines the analysis of sectored yeast colonies providing a rich set of quantitative metrics and provides insight into mechanisms driving the curing of prion phenotypes.

PMID:39641894 | DOI:10.1007/s11538-024-01379-w

Categories: Literature Watch

Leveraging a Vision Transformer Model to Improve Diagnostic Accuracy of Cardiac Amyloidosis With Cardiac Magnetic Resonance

Fri, 2024-12-06 06:00

JACC Cardiovasc Imaging. 2024 Nov 22:S1936-878X(24)00417-0. doi: 10.1016/j.jcmg.2024.09.010. Online ahead of print.

ABSTRACT

BACKGROUND: Cardiac magnetic resonance (CMR) imaging is an important diagnostic tool for diagnosis of cardiac amyloidosis (CA). However, discrimination of CA from other etiologies of myocardial disease can be challenging.

OBJECTIVES: The aim of this study was to develop and rigorously validate a deep learning (DL) algorithm to aid in the discrimination of CA using cine and late gadolinium enhancement CMR imaging.

METHODS: A DL model using a retrospective cohort of 807 patients who were referred for CMR for suspicion of infiltrative disease or hypertrophic cardiomyopathy (HCM) was developed. Confirmed definitive diagnosis was as follows: 252 patients with CA, 290 patients with HCM, and 265 with neither CA or HCM (other). This cohort was split 70/30 into training and test sets. A vision transformer (ViT) model was trained primarily to identify CA. The model was validated in an external cohort of 157 patients also referred for CMR for suspicion of infiltrative disease or HCM (51 CA, 49 HCM, 57 other).

RESULTS: The ViT model achieved a diagnostic accuracy (84.1%) and an area under the curve of 0.954 in the internal testing data set. The ViT model further demonstrated an accuracy of 82.8% and an area under the curve of 0.957 in the external testing set. The ViT model achieved an accuracy of 90% (n = 55 of 61), among studies with clinical reports with moderate/high confidence diagnosis of CA, and 61.1% (n = 22 of 36) among studies with reported uncertain, missing, or incorrect diagnosis of CA in the internal cohort. DL accuracy of this cohort increased to 79.1% when studies with poor image quality, dual pathologies, or ambiguity of clinically significant CA diagnosis were removed.

CONCLUSIONS: A ViT model using only cine and late gadolinium enhancement CMR images can achieve high accuracy in differentiating CA from other underlying etiologies of suspected cardiomyopathy, especially in cases when reported human diagnostic confidence was uncertain in both a large single state health system and in an external CA cohort.

PMID:39641685 | DOI:10.1016/j.jcmg.2024.09.010

Categories: Literature Watch

Reveal the potent antidepressant effects of Zhi-Zi-Hou-Pu Decoction based on integrated network pharmacology and DDI analysis by deep learning

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 3;10(22):e38726. doi: 10.1016/j.heliyon.2024.e38726. eCollection 2024 Nov 30.

ABSTRACT

BACKGROUND AND OBJECTIVE: The multi-targets and multi-components of Traditional Chinese medicine (TCM) coincide with the complex pathogenesis of depression. Zhi-Zi-Hou-Pu Decoction (ZZHPD) has been approved in clinical medication with good antidepression effects for centuries, while the mechanisms under the iceberg haven't been addressed systematically. This study explored its inner active ingredients - potent pharmacological mechanism - DDI to explore more comprehensively and deeply understanding of the complicated TCM in treatment.

METHODS: This research utilized network pharmacology combined with molecular docking to identify pharmacological targets and molecular interactions between ZZHPD and depression. Verification of major active compounds was conducted through UPLC-Q-TOF-MS/MS and assays on LPS-induced neuroblastoma cells. Additionally, the DDIMDL model, a deep learning-based approach, was used to predict DDIs, focusing on serum concentration, metabolism, effectiveness, and adverse reactions.

RESULTS: The antidepressant mechanisms of ZZHPD involve the serotonergic synapse, neuroactive ligand-receptor interaction, and dopaminergic synapse signaling pathways. Eighteen active compounds were identified, with honokiol and eriocitrin significantly modulating neuronal inflammation and promoting differentiation of neuroimmune cells through genes like COMT, PI3KCA, PTPN11, and MAPK1. DDI predictions indicated that eriocitrin's serum concentration increases when combined with hesperidin, while hesperetin's metabolism decreases with certain flavonoids. These findings provide crucial insights into the nervous system's effectiveness and potential cardiovascular or nervous system adverse reactions from core compound combinations.

CONCLUSIONS: This study provides insights into the TCM interpretation, drug compatibility or combined medication for further clinical application or potential drug pairs with a cost-effective method of integrated network pharmacology and deep learning.

PMID:39641032 | PMC:PMC11617927 | DOI:10.1016/j.heliyon.2024.e38726

Categories: Literature Watch

Development and validation of a deep learning pipeline to diagnose ovarian masses using ultrasound screening: a retrospective multicenter study

Fri, 2024-12-06 06:00

EClinicalMedicine. 2024 Nov 19;78:102923. doi: 10.1016/j.eclinm.2024.102923. eCollection 2024 Dec.

ABSTRACT

BACKGROUND: Ovarian cancer has the highest mortality rate among gynaecological malignancies and is initially screened using ultrasound. Owing to the high complexity of ultrasound images of ovarian masses and the anatomical characteristics of the deep pelvic cavity, subjective assessment requires extensive experience and skill. Therefore, detecting the ovaries and ovarian masses and diagnose ovarian cancer are challenging. In the present study, we aimed to develop an automated deep learning framework, the Ovarian Multi-Task Attention Network (OvaMTA), for ovary and ovarian mass detection, segmentation, and classification, as well as further diagnosis of ovarian masses based on ultrasound screening.

METHODS: Between June 2020 and May 2022, the OvaMTA model was trained, validated and tested on a training and validation cohort including 6938 images and an internal testing cohort including 1584 images which were recruited from 21 hospitals involving women who underwent ultrasound examinations for ovarian masses. Subsequently, we recruited two external test cohorts from another two hospitals. We obtained 1896 images between February 2024 and April 2024 as image-based external test dataset, and further obtained 159 videos for the video-based external test dataset between April 2024 and May 2024. We developed an artificial intelligence (AI) system (termed OvaMTA) to diagnose ovarian masses using ultrasound screening. It includes two models: an entire image-based segmentation model, OvaMTA-Seg, for ovary detection and a diagnosis model, OvaMTA-Diagnosis, for predicting the pathological type of ovarian mass using image patches cropped by OvaMTA-Seg. The performance of the system was evaluated in one internal and two external validation cohorts, and compared with doctors' assessments in real-world testing. We recruited eight physicians to assess the real-world data. The value of the system in assisting doctors with diagnosis was also evaluated.

FINDINGS: In terms of segmentation, OvaMTA-Seg achieved an average Dice score of 0.887 on the internal test set and 0.819 on the image-based external test set. OvaMTA-Seg also performed well in ovarian mass detection from test images, including healthy ovaries and masses (internal test area under the curve [AUC]: 0.970; external test AUC: 0.877). In terms of classification diagnosis prediction, OvaMTA-Diagnosis demonstrated high performance on image-based internal (AUC: 0.941) and external test sets (AUC: 0.941). In video-based external testing, OvaMTA recognised 159 videos with ovarian masses with AUC of 0.911, and is comparable to the performance of senior radiologists (ACC: 86.2 vs. 88.1, p = 0.50; SEN: 81.8 vs. 88.6, p = 0.16; SPE: 89.2 vs. 87.6, p = 0.68). There was a significant improvement in junior and intermediate radiologists who were assisted by AI compared to those who were not assisted by AI (ACC: 80.8 vs. 75.3, p = 0.00015; SEN: 79.5 vs. 74.6, p = 0.029; SPE: 81.7 vs. 75.8, p = 0.0032). General practitioners assisted by AI achieved an average performance of radiologists (ACC: 82.7 vs. 81.8, p = 0.80; SEN: 84.8 vs. 82.6, p = 0.72; SPE: 81.2 vs. 81.2, p > 0.99).

INTERPRETATION: The OvaMTA system based on ultrasound imaging is a simple and practical auxiliary tool for screening for ovarian cancer, with a diagnostic performance comparable to that of senior radiologists. This provides a potential tool for screening ovarian cancer.

FUNDING: This work was supported by the National Natural Science Foundation of China (Grant Nos. 12090020, 82071929, and 12090025) and the R&D project of the Pazhou Lab (Huangpu) (Grant No. 2023K0605).

PMID:39640935 | PMC:PMC11617315 | DOI:10.1016/j.eclinm.2024.102923

Categories: Literature Watch

Development and validation of the MRI-based deep learning classifier for distinguishing perianal fistulizing Crohn's disease from cryptoglandular fistula: a multicenter cohort study

Fri, 2024-12-06 06:00

EClinicalMedicine. 2024 Nov 22;78:102940. doi: 10.1016/j.eclinm.2024.102940. eCollection 2024 Dec.

ABSTRACT

BACKGROUND: A singular reliable modality for early distinguishing perianal fistulizing Crohn's disease (PFCD) from cryptoglandular fistula (CGF) is currently lacking. We aimed to develop and validate an MRI-based deep learning classifier to effectively discriminate between them.

METHODS: The present study retrospectively enrolled 1054 patients with PFCD or CGF from three Chinese tertiary referral hospitals between January 1, 2015, and December 31, 2021. The patients were divided into four cohorts: training cohort (n = 800), validation cohort (n = 100), internal test cohort (n = 100) and external test cohort (n = 54). Two deep convolutional neural networks (DCNN), namely MobileNetV2 and ResNet50, were respectively trained using the transfer learning strategy on a dataset consisting of 44871 MR images. The performance of the DCNN models was compared to that of radiologists using various metrics, including receiver operating characteristic curve (ROC) analysis, accuracy, sensitivity, and specificity. Delong testing was employed for comparing the area under curves (AUCs). Univariate and multivariate analyses were conducted to explore potential factors associated with classifier performance.

FINDINGS: A total of 532 PFCD and 522 CGF patients were included. Both pre-trained DCNN classifiers achieved encouraging performances in the internal test cohort (MobileNetV2 AUC: 0.962, 95% CI 0.903-0.990; ResNet50 AUC: 0.963, 95% CI 0.905-0.990), as well as external test cohort (MobileNetV2 AUC: 0.885, 95% CI 0.769-0.956; ResNet50 AUC: 0.874, 95% CI 0.756-0.949). They had greater AUCs than the radiologists (all p ≤ 0.001), while had comparable AUCs to each other (p = 0.83 and p = 0.60 in the two test cohorts). None of the potential characteristics had a significant impact on the performance of pre-trained MobileNetV2 classifier in etiologic diagnosis. Previous fistula surgery influenced the performance of the pre-trained ResNet50 classifier in the internal test cohort (OR 0.157, 95% CI 0.025-0.997, p = 0.05).

INTERPRETATION: The developed DCNN classifiers exhibited superior robustness in distinguishing PFCD from CGF compared to artificial visual assessment, showing their potential for assisting in early detection of PFCD. Our findings highlight the promising generalized performance of MobileNetV2 over ResNet50, rendering it suitable for deployment on mobile terminals.

FUNDING: National Natural Science Foundation of China.

PMID:39640934 | PMC:PMC11618046 | DOI:10.1016/j.eclinm.2024.102940

Categories: Literature Watch

Towards multi-agent system for learning object recommendation

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 11;10(20):e39088. doi: 10.1016/j.heliyon.2024.e39088. eCollection 2024 Oct 30.

ABSTRACT

The rapid increase of online educational content has made it harder for students to find specific information. E-learning recommender systems help students easily find the learning objects they require, improving the learning experience. The effectiveness of these systems is further improved by integrating deep learning with multi-agent systems. Multi-agent systems facilitate adaptable interactions within the system's various parts, and deep learning processes extensive data to understand learners' preferences. This collaboration results in custom-made suggestions that cater to individual learners. Our research introduces a multi-agent system tailored for suggesting learning objects in line with learners' knowledge levels and learning styles. This system uniquely comprises four agents: the learner agent, the tutor agent, the learning object agent, and the recommendation agent. It applies the Felder and Silverman model to pinpoint various student learning styles and organizes educational content based on the newest IEEE Learning Object Metadata standard. The system uses advanced techniques, such as Convolutional Neural Networks (CNN) and Multilayer Perceptrons (MLP), to propose learning objects. In terms of creating personalized learning experiences, this system is a considerable step forward. It effectively suggests learning objects that closely match each learner's personal profile, greatly enhancing student engagement and making the learning process more efficient.

PMID:39640789 | PMC:PMC11620102 | DOI:10.1016/j.heliyon.2024.e39088

Categories: Literature Watch

Comprehensive analysis of computational approaches in plant transcription factors binding regions discovery

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 10;10(20):e39140. doi: 10.1016/j.heliyon.2024.e39140. eCollection 2024 Oct 30.

ABSTRACT

Transcription factors (TFs) are regulatory proteins which bind to a specific DNA region known as the transcription factor binding regions (TFBRs) to regulate the rate of transcription process. The identification of TFBRs has been made possible by a number of experimental and computational techniques established during the past few years. The process of TFBR identification involves peak identification in the binding data, followed by the identification of motif characteristics. Using the same binding data attempts have been made to raise computational models to identify such binding regions which could save time and resources spent for binding experiments. These computational approaches depend a lot on what way they learn and how. These existing computational approaches are skewed heavily around human TFBRs discovery, while plants have drastically different genomic setup for regulation which these approaches have grossly ignored. Here, we provide a comprehensive study of the current state of the matters in plant specific TF discovery algorithms. While doing so, we encountered several software tools' issues rendering the tools not useable to researches. We fixed them and have also provided the corrected scripts for such tools. We expect this study to serve as a guide for better understanding of software tools' approaches for plant specific TFBRs discovery and the care to be taken while applying them, especially during cross-species applications. The corrected scripts of these software tools are made available at https://github.com/SCBB-LAB/Comparative-analysis-of-plant-TFBS-software.

PMID:39640721 | PMC:PMC11620080 | DOI:10.1016/j.heliyon.2024.e39140

Categories: Literature Watch

State-of-health estimation and classification of series-connected batteries by using deep learning based hybrid decision approach

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 9;10(20):e39121. doi: 10.1016/j.heliyon.2024.e39121. eCollection 2024 Oct 30.

ABSTRACT

In rechargeable battery control and operation, one of the primary obstacles is safety concerns where the battery degradation poses a significant factor. Therefore, in recent years, state-of-health assessment of lithium-ion batteries has become a noteworthy issue. On the other hand, it is challenging to ensure robustness and generalization because most state-of-health assessment techniques are implemented for a specific characteristic, operating situation, and battery material system. In most studies, health status of single cell batteries is assessed by using analytical or computer-aided deep learning methods. But, the state-of-health characteristics of series-connected battery systems should be also focused with advances of technology and usage, especially electric vehicles. This study presents a data-driven, deep learning-based hybrid decision approach for predicting the state-of-health of series-connected lithium-ion batteries with different characteristics. The paper consists of generating series-connected battery degradation dataset by using of some mostly used datasets. Also, by employing deep learning-based networks along with hybrid-classification aided by performance metrics, it is shown that estimating and predicting the state-of-health can be achieved not only by using sole deep-learning algorithms but also hybrid-classification techniques. The results demonstrate the high accuracy and simplicity of the proposed novel approach on datasets from Oxford University and Calce battery group. The best estimated mean squared error, root mean square error and mean-absolute percentage error values are not more than 0.0500, 0.2236 and 0.7065, respectively which shows the efficiency not only by accuracy but also error indicators. The results show that the proposed approach can be implemented in offline or online systems with best average accuracy of 98.33 % and classification time of 58 ms per sample.

PMID:39640714 | PMC:PMC11620055 | DOI:10.1016/j.heliyon.2024.e39121

Categories: Literature Watch

Deep learning neural network-assisted badminton movement recognition and physical fitness training optimization

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 2;10(20):e38865. doi: 10.1016/j.heliyon.2024.e38865. eCollection 2024 Oct 30.

ABSTRACT

This work aims to solve the problem of low accuracy in recognizing the trajectory of badminton movement. This work focuses on the visual system in badminton robots and conducts side detection and tracking of flying badminton in two-dimensional image plane video streams. Then, the cropped video images are input into a convolutional neural network frame by frame. By adding an attention mechanism, it helps identify the badminton movement trajectory. Finally, to address the detection challenge of flying badminton as a small target in video streams, the deep learning one-stage detection network, Tiny YOLOv2, is improved from both the loss function and network structure perspectives. Moreover, it is combined with the Unscented Kalman Filter algorithm to predict the trajectory of badminton movement. Simulation results show that the improved algorithm performs excellently in tracking and predicting badminton trajectories compared with the existing algorithms. The average accuracy of the proposed method for tracking badminton trajectories is 91.40 %, and the recall rate is 84.60 %. The average precision, recall, and frame rate of the measured trajectories in four simple and complex scenarios of badminton flight video streams are 96.7 %, 95.7 %, and 29.2 frames/second, respectively. They are all superior to other classic algorithms. It is evident that the proposed method can provide powerful support for badminton trajectory recognition and help improve the accuracy of badminton movement recognition.

PMID:39640697 | PMC:PMC11620146 | DOI:10.1016/j.heliyon.2024.e38865

Categories: Literature Watch

Introducing a novel dataset for facial emotion recognition and demonstrating significant enhancements in deep learning performance through pre-processing techniques

Fri, 2024-12-06 06:00

Heliyon. 2024 Oct 4;10(20):e38913. doi: 10.1016/j.heliyon.2024.e38913. eCollection 2024 Oct 30.

ABSTRACT

Facial expression recognition (FER) plays a pivotal role in various applications, ranging from human-computer interaction to psychoanalysis. To improve the accuracy of facial emotion recognition (FER) models, this study focuses on enhancing and augmenting FER datasets. It comprehensively analyzes the Facial Emotion Recognition dataset (FER13) to identify defects and correct misclassifications. The FER13 dataset represents a crucial resource for researchers developing Deep Learning (DL) models aimed at recognizing emotions based on facial features. Subsequently, this article develops a new facial dataset by expanding upon the original FER13 dataset. Similar to the FER + dataset, the expanded dataset incorporates a wider range of emotions while maintaining data accuracy. To further improve the dataset, it will be integrated with the extended Cohn-Kanade (CK+) dataset. This paper investigates the application of modern DL models to enhance emotion recognition in human faces. By training a new dataset, the study demonstrates significant performance gains compared with its counterparts. Furthermore, the article examines recent advances in FER technology and identifies critical requirements for DL models to overcome the inherent challenges of this task effectively. The study explores several DL architectures for emotion recognition in facial image datasets, with a particular focus on convolutional neural networks (CNNs). Our findings indicate that complex architecture, such as EfficientNetB7, outperforms other DL architectures, achieving a test accuracy of 78.9 %. Notably, the model surpassed the EfficientNet-XGBoost model, especially when used with the new dataset. Our approach leverages EfficientNetB7 as a backbone to build a model capable of efficiently recognizing emotions from facial images. Our proposed model, EfficientNetB7-CNN, achieved a peak accuracy of 81 % on the test set despite facing challenges such as GPU memory limitations. This demonstrates the model's robustness in handling complex facial expressions. Furthermore, to enhance feature extraction and attention mechanisms, we propose a new hybrid model, CBAM-4CNN, which integrates the convolutional block attention module (CBAM) with a custom 4-layer CNN architecture. The results showed that the CBAM-4CNN model outperformed existing models, achieving higher accuracy, precision, and recall metrics across multiple emotion classes. The results highlight the critical role of comprehensive and diverse data in enhancing model performance for facial emotion recognition.

PMID:39640693 | PMC:PMC11620061 | DOI:10.1016/j.heliyon.2024.e38913

Categories: Literature Watch

Pages