Deep learning
Image-based classification of wheat spikes by glume pubescence using convolutional neural networks
Front Plant Sci. 2024 Jan 12;14:1336192. doi: 10.3389/fpls.2023.1336192. eCollection 2023.
ABSTRACT
INTRODUCTION: Pubescence is an important phenotypic trait observed in both vegetative and generative plant organs. Pubescent plants demonstrate increased resistance to various environmental stresses such as drought, low temperatures, and pests. It serves as a significant morphological marker and aids in selecting stress-resistant cultivars, particularly in wheat. In wheat, pubescence is visible on leaves, leaf sheath, glumes and nodes. Regarding glumes, the presence of pubescence plays a pivotal role in its classification. It supplements other spike characteristics, aiding in distinguishing between different varieties within the wheat species. The determination of pubescence typically involves visual analysis by an expert. However, methods without the use of binocular loupe tend to be subjective, while employing additional equipment is labor-intensive. This paper proposes an integrated approach to determine glume pubescence presence in spike images captured under laboratory conditions using a digital camera and convolutional neural networks.
METHODS: Initially, image segmentation is conducted to extract the contour of the spike body, followed by cropping of the spike images to an equal size. These images are then classified based on glume pubescence (pubescent/glabrous) using various convolutional neural network architectures (Resnet-18, EfficientNet-B0, and EfficientNet-B1). The networks were trained and tested on a dataset comprising 9,719 spike images.
RESULTS: For segmentation, the U-Net model with EfficientNet-B1 encoder was chosen, achieving the segmentation accuracy IoU = 0.947 for the spike body and 0.777 for awns. The classification model for glume pubescence with the highest performance utilized the EfficientNet-B1 architecture. On the test sample, the model exhibited prediction accuracy parameters of F1 = 0.85 and AUC = 0.96, while on the holdout sample it showed F1 = 0.84 and AUC = 0.89. Additionally, the study investigated the relationship between image scale, artificial distortions, and model prediction performance, revealing that higher magnification and smaller distortions yielded a more accurate prediction of glume pubescence.
PMID:38283969 | PMC:PMC10811101 | DOI:10.3389/fpls.2023.1336192
Transcriptome-wide analysis of PIP reductase gene family identified a phenylpropene synthase crucial for the biosynthesis of dibenzocyclooctadiene lignans in <em>Schisandra chinensis</em>
Synth Syst Biotechnol. 2023 Dec 7;9(1):78-87. doi: 10.1016/j.synbio.2023.11.011. eCollection 2024 Mar.
ABSTRACT
Phenylpropenes, such as isoeugenol and eugenol, are produced as defend compounds, floral attractants, and flavor constituents by phenylpropene synthases belonging to the PIP reductase family. Moreover, isoeugenol is proposed to be involved in the biosynthesis of dibenzocyclooctadiene lignans, the main active compounds of Schisandra chinensis (Turcz.) Baill. fruits (SCF). S. chinensis, a woody vine plant, is widely used for its medicinal, horticultural, edible, and economic values. In this study, nine ScPIP genes were identified and characterized from the transcriptome datasets of SCF. The expression profiles revealed that ScPIP genes were differentially expressed during different developmental stages of SCF. Three ScPIPs were selected and cloned as candidate genes encoding phenylpropene synthases according to phylogenetic analysis. ScPIP1 was proved to function as isoeugenol synthase (IGS) and designated as ScIGS1 through in vivo functional characterization in Escherichia coli. Subcellular localization analysis demonstrated that ScIGS1 was localized in both the cytoplasm and nucleus. The three-dimensional (3D) model of ScIGS1 was obtained using homology modeling. Site-directed mutagenesis experiments revealed that the substitution of residues at positions 110 and 113 impacted the product specificity of ScIGS1 and the mutation of Lys157 to Ala abolishing catalytic function. Moreover, the kcat values of mutants were lower than that of ScIGS1 using a deep learning approach. In conclusion, this study provides a basis for further research on PIP reductases and the biosynthetic pathway of dibenzocyclooctadiene lignans.
PMID:38283950 | PMC:PMC10819558 | DOI:10.1016/j.synbio.2023.11.011
Quantification of blood flow index in diffuse correlation spectroscopy using a robust deep learning method
J Biomed Opt. 2024 Jan;29(1):015004. doi: 10.1117/1.JBO.29.1.015004. Epub 2024 Jan 27.
ABSTRACT
SIGNIFICANCE: Diffuse correlation spectroscopy (DCS) is a powerful, noninvasive optical technique for measuring blood flow. Traditionally the blood flow index (BFi) is derived through nonlinear least-square fitting the measured intensity autocorrelation function (ACF). However, the fitting process is computationally intensive, susceptible to measurement noise, and easily influenced by optical properties (absorption coefficient μa and reduced scattering coefficient μs') and scalp and skull thicknesses.
AIM: We aim to develop a data-driven method that enables rapid and robust analysis of multiple-scattered light's temporal ACFs. Moreover, the proposed method can be applied to a range of source-detector distances instead of being limited to a specific source-detector distance.
APPROACH: We present a deep learning architecture with one-dimensional convolution neural networks, called DCS neural network (DCS-NET), for BFi and coherent factor (β) estimation. This DCS-NET was performed using simulated DCS data based on a three-layer brain model. We quantified the impact from physiologically relevant optical property variations, layer thicknesses, realistic noise levels, and multiple source-detector distances (5, 10, 15, 20, 25, and 30 mm) on BFi and β estimations among DCS-NET, semi-infinite, and three-layer fitting models.
RESULTS: DCS-NET shows a much faster analysis speed, around 17,000-fold and 32-fold faster than the traditional three-layer and semi-infinite models, respectively. It offers higher intrinsic sensitivity to deep tissues compared with fitting methods. DCS-NET shows excellent anti-noise features and is less sensitive to variations of μa and μs' at a source-detector separation of 30 mm. Also, we have demonstrated that relative BFi (rBFi) can be extracted by DCS-NET with a much lower error of 8.35%. By contrast, the semi-infinite and three-layer fitting models result in significant errors in rBFi of 43.76% and 19.66%, respectively.
CONCLUSIONS: DCS-NET can robustly quantify blood flow measurements at considerable source-detector distances, corresponding to much deeper biological tissues. It has excellent potential for hardware implementation, promising continuous real-time blood flow measurements.
PMID:38283935 | PMC:PMC10821781 | DOI:10.1117/1.JBO.29.1.015004
Explanatory deep learning to predict elevated pulmonary artery pressure in children with ventricular septal defects using standard chest x-rays: a novel approach
Front Cardiovasc Med. 2024 Jan 12;11:1330685. doi: 10.3389/fcvm.2024.1330685. eCollection 2024.
ABSTRACT
OBJECTIVE: Early risk assessment of pulmonary arterial hypertension (PAH) in patients with congenital heart disease (CHD) is crucial to ensure timely treatment. We hypothesize that applying artificial intelligence (AI) to chest x-rays (CXRs) could identify the future risk of PAH in patients with ventricular septal defect (VSD).
METHODS: A total of 831 VSD patients (161 PAH-VSD, 670 nonPAH-VSD) was retrospectively included. A residual neural networks (ResNet) was trained for classify VSD patients with different outcomes based on chest radiographs. The endpoint of this study was the occurrence of PAH in VSD children before or after surgery.
RESULTS: In the validation set, the AI algorithm achieved an area under the curve (AUC) of 0.82. In an independent test set, the AI algorithm significantly outperformed human observers in terms of AUC (0.81 vs. 0.65). Class Activation Mapping (CAM) images demonstrated the model's attention focused on the pulmonary artery segment.
CONCLUSION: The preliminary findings of this study suggest that the application of artificial intelligence to chest x-rays in VSD patients can effectively identify the risk of PAH.
PMID:38283829 | PMC:PMC10811002 | DOI:10.3389/fcvm.2024.1330685
Linking repeated subjective judgments and ConvNets for multimodal assessment of the immediate living environment
MethodsX. 2024 Jan 5;12:102556. doi: 10.1016/j.mex.2024.102556. eCollection 2024 Jun.
ABSTRACT
The integration of alternative data extraction approaches for multimodal data, can significantly reduce modeling difficulties for the automatic location assessment. We develop a method for assessing the quality of the immediate living environment by incorporating human judgments as ground truth into a neural network for generating new synthetic data and testing the effects in surrogate hedonic models. We expect that the quality of the data will be less biased if the annotation is performed by multiple independent persons applying repeated trials which should reduce the overall error variance and lead to more robust results. Experimental results show that linking repeated subjective judgements and Deep Learning can reliably determine the quality scores and thus expand the range of information for the quality assessment. The presented method is not computationally intensive, can be performed repetitively and can also be easily adapted to machine learning approaches in a broader sense or be transferred to other use cases. Following aspects are essential for the implementation of the method:•Sufficient amount of representative data for human assessment.•Repeated assessment trials by individuals.•Confident derivation of the effect of human judgments on property price as an approbation for further generation of synthetic data.
PMID:38283760 | PMC:PMC10820260 | DOI:10.1016/j.mex.2024.102556
DF-dRVFL: A novel deep feature based classifier for breast mass classification
Multimed Tools Appl. 2024;83(5):14393-14422. doi: 10.1007/s11042-023-15864-2. Epub 2023 Jul 11.
ABSTRACT
Amongst all types of cancer, breast cancer has become one of the most common cancers in the UK threatening millions of people's health. Early detection of breast cancer plays a key role in timely treatment for morbidity reduction. Compared to biopsy, which takes tissues from the lesion for further analysis, image-based methods are less time-consuming and pain-free though they are hampered by lower accuracy due to high false positivity rates. Nevertheless, mammography has become a standard screening method due to its high efficiency and low cost with promising performance. Breast mass, as the most palpable symptom of breast cancer, has received wide attention from the community. As a result, the past decades have witnessed the speeding development of computer-aided systems that are aimed at providing radiologists with useful tools for breast mass analysis based on mammograms. However, the main issues of these systems include low accuracy and require enough computational power on a large scale of datasets. To solve these issues, we developed a novel breast mass classification system called DF-dRVFL. On the public dataset DDSM with more than 3500 images, our best model based on deep random vector functional link network showed promising results through five-cross validation with an averaged AUC of 0.93 and an average accuracy of 81.71%. Compared to sole deep learning based methods, average accuracy has increased by 0.38. Compared with the state-of-the-art methods, our method showed better performance considering the number of images for evaluation and the overall accuracy.
PMID:38283725 | PMC:PMC10817886 | DOI:10.1007/s11042-023-15864-2
Automatic detection of mild cognitive impairment based on deep learning and radiomics of MR imaging
Front Med (Lausanne). 2024 Jan 12;11:1305565. doi: 10.3389/fmed.2024.1305565. eCollection 2024.
ABSTRACT
PURPOSE: Early and rapid diagnosis of mild cognitive impairment (MCI) has important clinical value in improving the prognosis of Alzheimer's disease (AD). The hippocampus and parahippocampal gyrus play crucial roles in the occurrence of cognitive function decline. In this study, deep learning and radiomics techniques were used to automatically detect MCI from healthy controls (HCs).
METHOD: This study included 115 MCI patients and 133 normal individuals with 3D-T1 weighted MR structural images from the ADNI database. The identification and segmentation of the hippocampus and parahippocampal gyrus were automatically performed with a VB-net, and radiomics features were extracted. Relief, Minimum Redundancy Maximum Correlation, Recursive Feature Elimination and the minimum absolute shrinkage and selection operator (LASSO) were used to reduce the dimensionality and select the optimal features. Five independent machine learning classifiers including Support Vector Machine (SVM), Random forest (RF), Logistic Regression (LR), Bagging Decision Tree (BDT), and Gaussian Process (GP) were trained on the training set, and validated on the testing set to detect the MCI. The Delong test was used to assess the performance of different models.
RESULT: Our VB-net could automatically identify and segment the bilateral hippocampus and parahippocampal gyrus. After four steps of feature dimensionality reduction, the GP models based on combined features (11 features from the hippocampus, and 4 features from the parahippocampal gyrus) showed the best performance for the MCI and normal control subject discrimination. The AUC of the training set and test set were 0.954 (95% CI: 0.929-0.979) and 0.866 (95% CI: 0.757-0.976), respectively. Decision curve analysis showed that the clinical benefit of the line graph model was high.
CONCLUSION: The GP classifier based on 15 radiomics features of bilateral hippocampal and parahippocampal gyrus could detect MCI from normal controls with high accuracy based on conventional MR images. Our fully automatic model could rapidly process the MRI data and give results in 1 minute, which provided important clinical value in assisted diagnosis.
PMID:38283620 | PMC:PMC10811129 | DOI:10.3389/fmed.2024.1305565
Use of deep learning for the classification of hyperplastic lymph node and common subtypes of canine lymphomas: a preliminary study
Front Vet Sci. 2024 Jan 12;10:1309877. doi: 10.3389/fvets.2023.1309877. eCollection 2023.
ABSTRACT
Artificial Intelligence has observed significant growth in its ability to classify different types of tumors in humans due to advancements in digital pathology technology. Among these tumors, lymphomas are quite common in dogs, despite studies on the application of AI in domestic species are scarce. This research aims to employ deep learning (DL) through convolutional neural networks (CNNs) to distinguish between normal lymph nodes and 3 WHO common subtypes of canine lymphomas. To train and validate the CNN, 1,530 high-resolution microscopic images derived from whole slide scans (WSIs) were used, including those of background areas, hyperplastic lymph nodes (n = 4), and three different lymphoma subtypes: diffuse large B cell lymphoma (DLBCL; n = 5), lymphoblastic (LBL; n = 5), and marginal zone lymphoma (MZL; n = 3). The CNN was able to correctly identify 456 images of the possible 457 test sets, achieving a maximum accuracy of 99.34%. The results of this study have demonstrated the feasibility of using deep learning to differentiate between hyperplastic lymph nodes and lymphomas, as well as to classify common WHO subtypes. Further research is required to explore the implications of these findings and validate the ability of the network to classify a broader range of lymphomas.
PMID:38283371 | PMC:PMC10811236 | DOI:10.3389/fvets.2023.1309877
Neural deformation fields for template-based reconstruction of cortical surfaces from MRI
Med Image Anal. 2024 Jan 26;93:103093. doi: 10.1016/j.media.2024.103093. Online ahead of print.
ABSTRACT
The reconstruction of cortical surfaces is a prerequisite for quantitative analyses of the cerebral cortex in magnetic resonance imaging (MRI). Existing segmentation-based methods separate the surface registration from the surface extraction, which is computationally inefficient and prone to distortions. We introduce Vox2Cortex-Flow (V2C-Flow), a deep mesh-deformation technique that learns a deformation field from a brain template to the cortical surfaces of an MRI scan. To this end, we present a geometric neural network that models the deformation-describing ordinary differential equation in a continuous manner. The network architecture comprises convolutional and graph-convolutional layers, which allows it to work with images and meshes at the same time. V2C-Flow is not only very fast, requiring less than two seconds to infer all four cortical surfaces, but also establishes vertex-wise correspondences to the template during reconstruction. In addition, V2C-Flow is the first approach for cortex reconstruction that models white matter and pial surfaces jointly, therefore avoiding intersections between them. Our comprehensive experiments on internal and external test data demonstrate that V2C-Flow results in cortical surfaces that are state-of-the-art in terms of accuracy. Moreover, we show that the established correspondences are more consistent than in FreeSurfer and that they can directly be utilized for cortex parcellation and group analyses of cortical thickness.
PMID:38281362 | DOI:10.1016/j.media.2024.103093
Transfer learning of pre-treatment quantitative ultrasound multi-parametric images for the prediction of breast cancer response to neoadjuvant chemotherapy
Sci Rep. 2024 Jan 29;14(1):2340. doi: 10.1038/s41598-024-52858-y.
ABSTRACT
Locally advanced breast cancer (LABC) is a severe type of cancer with a poor prognosis, despite advancements in therapy. As the disease is often inoperable, current guidelines suggest upfront aggressive neoadjuvant chemotherapy (NAC). Complete pathological response to chemotherapy is linked to improved survival, but conventional clinical assessments like physical exams, mammography, and imaging are limited in detecting early response. Early detection of tissue response can improve complete pathological response and patient survival while reducing exposure to ineffective and potentially harmful treatments. A rapid, cost-effective modality without the need for exogenous contrast agents would be valuable for evaluating neoadjuvant therapy response. Conventional ultrasound provides information about tissue echogenicity, but image comparisons are difficult due to instrument-dependent settings and imaging parameters. Quantitative ultrasound (QUS) overcomes this by using normalized power spectra to calculate quantitative metrics. This study used a novel transfer learning-based approach to predict LABC response to neoadjuvant chemotherapy using QUS imaging at pre-treatment. Using data from 174 patients, QUS parametric images of breast tumors with margins were generated. The ground truth response to therapy for each patient was based on standard clinical and pathological criteria. The Residual Network (ResNet) deep learning architecture was used to extract features from the parametric QUS maps. This was followed by SelectKBest and Synthetic Minority Oversampling (SMOTE) techniques for feature selection and data balancing, respectively. The Support Vector Machine (SVM) algorithm was employed to classify patients into two distinct categories: nonresponders (NR) and responders (RR). Evaluation results on an unseen test set demonstrate that the transfer learning-based approach using spectral slope parametric maps had the best performance in the identification of nonresponders with precision, recall, F1-score, and balanced accuracy of 100, 71, 83, and 86%, respectively. The transfer learning-based approach has many advantages over conventional deep learning methods since it reduces the need for large image datasets for training and shortens the training time. The results of this study demonstrate the potential of transfer learning in predicting LABC response to neoadjuvant chemotherapy before the start of treatment using quantitative ultrasound imaging. Prediction of NAC response before treatment can aid clinicians in customizing ineffectual treatment regimens for individual patients.
PMID:38282158 | DOI:10.1038/s41598-024-52858-y
Hybrid representation-enhanced sampling for Bayesian active learning in musculoskeletal segmentation of lower extremities
Int J Comput Assist Radiol Surg. 2024 Jan 29. doi: 10.1007/s11548-024-03065-7. Online ahead of print.
ABSTRACT
PURPOSE: Manual annotations for training deep learning models in auto-segmentation are time-intensive. This study introduces a hybrid representation-enhanced sampling strategy that integrates both density and diversity criteria within an uncertainty-based Bayesian active learning (BAL) framework to reduce annotation efforts by selecting the most informative training samples.
METHODS: The experiments are performed on two lower extremity datasets of MRI and CT images, focusing on the segmentation of the femur, pelvis, sacrum, quadriceps femoris, hamstrings, adductors, sartorius, and iliopsoas, utilizing a U-net-based BAL framework. Our method selects uncertain samples with high density and diversity for manual revision, optimizing for maximal similarity to unlabeled instances and minimal similarity to existing training data. We assess the accuracy and efficiency using dice and a proposed metric called reduced annotation cost (RAC), respectively. We further evaluate the impact of various acquisition rules on BAL performance and design an ablation study for effectiveness estimation.
RESULTS: In MRI and CT datasets, our method was superior or comparable to existing ones, achieving a 0.8% dice and 1.0% RAC increase in CT (statistically significant), and a 0.8% dice and 1.1% RAC increase in MRI (not statistically significant) in volume-wise acquisition. Our ablation study indicates that combining density and diversity criteria enhances the efficiency of BAL in musculoskeletal segmentation compared to using either criterion alone.
CONCLUSION: Our sampling method is proven efficient in reducing annotation costs in image segmentation tasks. The combination of the proposed method and our BAL framework provides a semi-automatic way for efficient annotation of medical image datasets.
PMID:38282095 | DOI:10.1007/s11548-024-03065-7
Dual contrastive learning based image-to-image translation of unstained skin tissue into virtually stained H&E images
Sci Rep. 2024 Jan 28;14(1):2335. doi: 10.1038/s41598-024-52833-7.
ABSTRACT
Staining is a crucial step in histopathology that prepares tissue sections for microscopic examination. Hematoxylin and eosin (H&E) staining, also known as basic or routine staining, is used in 80% of histopathology slides worldwide. To enhance the histopathology workflow, recent research has focused on integrating generative artificial intelligence and deep learning models. These models have the potential to improve staining accuracy, reduce staining time, and minimize the use of hazardous chemicals, making histopathology a safer and more efficient field. In this study, we introduce a novel three-stage, dual contrastive learning-based, image-to-image generative (DCLGAN) model for virtually applying an "H&E stain" to unstained skin tissue images. The proposed model utilizes a unique learning setting comprising two pairs of generators and discriminators. By employing contrastive learning, our model maximizes the mutual information between traditional H&E-stained and virtually stained H&E patches. Our dataset consists of pairs of unstained and H&E-stained images, scanned with a brightfield microscope at 20 × magnification, providing a comprehensive set of training and testing images for evaluating the efficacy of our proposed model. Two metrics, Fréchet Inception Distance (FID) and Kernel Inception Distance (KID), were used to quantitatively evaluate virtual stained slides. Our analysis revealed that the average FID score between virtually stained and H&E-stained images (80.47) was considerably lower than that between unstained and virtually stained slides (342.01), and unstained and H&E stained (320.4) indicating a similarity virtual and H&E stains. Similarly, the mean KID score between H&E stained and virtually stained images (0.022) was significantly lower than the mean KID score between unstained and H&E stained (0.28) or unstained and virtually stained (0.31) images. In addition, a group of experienced dermatopathologists evaluated traditional and virtually stained images and demonstrated an average agreement of 78.8% and 90.2% for paired and single virtual stained image evaluations, respectively. Our study demonstrates that the proposed three-stage dual contrastive learning-based image-to-image generative model is effective in generating virtual stained images, as indicated by quantified parameters and grader evaluations. In addition, our findings suggest that GAN models have the potential to replace traditional H&E staining, which can reduce both time and environmental impact. This study highlights the promise of virtual staining as a viable alternative to traditional staining techniques in histopathology.
PMID:38282056 | DOI:10.1038/s41598-024-52833-7
Evaluating a radiotherapy deep learning synthetic CT algorithm for PET-MR attenuation correction in the pelvis
EJNMMI Phys. 2024 Jan 29;11(1):10. doi: 10.1186/s40658-024-00617-3.
ABSTRACT
BACKGROUND: Positron emission tomography-magnetic resonance (PET-MR) attenuation correction is challenging because the MR signal does not represent tissue density and conventional MR sequences cannot image bone. A novel zero echo time (ZTE) MR sequence has been previously developed which generates signal from cortical bone with images acquired in 65 s. This has been combined with a deep learning model to generate a synthetic computed tomography (sCT) for MR-only radiotherapy. This study aimed to evaluate this algorithm for PET-MR attenuation correction in the pelvis.
METHODS: Ten patients being treated with ano-rectal radiotherapy received a [Formula: see text]F-FDG-PET-MR in the radiotherapy position. Attenuation maps were generated from ZTE-based sCT (sCTAC) and the standard vendor-supplied MRAC. The radiotherapy planning CT scan was rigidly registered and cropped to generate a gold standard attenuation map (CTAC). PET images were reconstructed using each attenuation map and compared for standard uptake value (SUV) measurement, automatic thresholded gross tumour volume (GTV) delineation and GTV metabolic parameter measurement. The last was assessed for clinical equivalence to CTAC using two one-sided paired t tests with a significance level corrected for multiple testing of [Formula: see text]. Equivalence margins of [Formula: see text] were used.
RESULTS: Mean whole-image SUV differences were -0.02% (sCTAC) compared to -3.0% (MRAC), with larger differences in the bone regions (-0.5% to -16.3%). There was no difference in thresholded GTVs, with Dice similarity coefficients [Formula: see text]. However, there were larger differences in GTV metabolic parameters. Mean differences to CTAC in [Formula: see text] were [Formula: see text] (± standard error, sCTAC) and [Formula: see text] (MRAC), and [Formula: see text] (sCTAC) and [Formula: see text] (MRAC) in [Formula: see text]. The sCTAC was statistically equivalent to CTAC within a [Formula: see text] equivalence margin for [Formula: see text] and [Formula: see text] ([Formula: see text] and [Formula: see text]), whereas the MRAC was not ([Formula: see text] and [Formula: see text]).
CONCLUSION: Attenuation correction using this radiotherapy ZTE-based sCT algorithm was substantially more accurate than current MRAC methods with only a 40 s increase in MR acquisition time. This did not impact tumour delineation but did significantly improve the accuracy of whole-image and tumour SUV measurements, which were clinically equivalent to CTAC. This suggests PET images reconstructed with sCTAC would enable accurate quantitative PET images to be acquired on a PET-MR scanner.
PMID:38282050 | DOI:10.1186/s40658-024-00617-3
Optimized network based natural language processing approach to reveal disease comorbidities in COVID-19
Sci Rep. 2024 Jan 28;14(1):2325. doi: 10.1038/s41598-024-52819-5.
ABSTRACT
A novel virus emerged from Wuhan, China, at the end of 2019 and quickly evolved into a pandemic, significantly impacting various industries, especially healthcare. One critical lesson from COVID-19 is the importance of understanding and predicting underlying comorbidities to better prioritize care and pharmacological therapies. Factors like age, race, and comorbidity history are crucial in determining disease mortality. While clinical data from hospitals and cohorts have led to the identification of these comorbidities, traditional approaches often lack a mechanistic understanding of the connections between them. In response, we utilized a deep learning approach to integrate COVID-19 data with data from other diseases, aiming to detect comorbidities with mechanistic insights. Our modified algorithm in the mpDisNet package, based on word-embedding deep learning techniques, incorporates miRNA expression profiles from SARS-CoV-2 infected cell lines and their target transcription factors. This approach is aligned with the emerging field of network medicine, which seeks to define diseases based on distinct pathomechanisms rather than just phenotypes. The main aim is discovery of possible unknown comorbidities by connecting the diseases by their miRNA mediated regulatory interactions. The algorithm can predict the majority of COVID-19's known comorbidities, as well as several diseases that have yet to be discovered to be comorbid with COVID-19. These potentially comorbid diseases should be investigated further to raise awareness and prevention, as well as informing the comorbidity research for the next possible outbreak.
PMID:38282038 | DOI:10.1038/s41598-024-52819-5
A computational model of circRNA-associated diseases based on a graph neural network: prediction and case studies for follow-up experimental validation
BMC Biol. 2024 Jan 29;22(1):24. doi: 10.1186/s12915-024-01826-z.
ABSTRACT
BACKGROUND: Circular RNAs (circRNAs) have been confirmed to play a vital role in the occurrence and development of diseases. Exploring the relationship between circRNAs and diseases is of far-reaching significance for studying etiopathogenesis and treating diseases. To this end, based on the graph Markov neural network algorithm (GMNN) constructed in our previous work GMNN2CD, we further considered the multisource biological data that affects the association between circRNA and disease and developed an updated web server CircDA and based on the human hepatocellular carcinoma (HCC) tissue data to verify the prediction results of CircDA.
RESULTS: CircDA is built on a Tumarkov-based deep learning framework. The algorithm regards biomolecules as nodes and the interactions between molecules as edges, reasonably abstracts multiomics data, and models them as a heterogeneous biomolecular association network, which can reflect the complex relationship between different biomolecules. Case studies using literature data from HCC, cervical, and gastric cancers demonstrate that the CircDA predictor can identify missing associations between known circRNAs and diseases, and using the quantitative real-time PCR (RT-qPCR) experiment of HCC in human tissue samples, it was found that five circRNAs were significantly differentially expressed, which proved that CircDA can predict diseases related to new circRNAs.
CONCLUSIONS: This efficient computational prediction and case analysis with sufficient feedback allows us to identify circRNA-associated diseases and disease-associated circRNAs. Our work provides a method to predict circRNA-associated diseases and can provide guidance for the association of diseases with certain circRNAs. For ease of use, an online prediction server ( http://server.malab.cn/CircDA ) is provided, and the code is open-sourced ( https://github.com/nmt315320/CircDA.git ) for the convenience of algorithm improvement.
PMID:38281919 | DOI:10.1186/s12915-024-01826-z
Expert System for Fourier Transform Infrared Spectra Recognition Based on a Convolutional Neural Network With Multiclass Classification
Appl Spectrosc. 2024 Jan 28:37028241226732. doi: 10.1177/00037028241226732. Online ahead of print.
ABSTRACT
Fourier transform infrared spectroscopy (FT-IR) is a widely used spectroscopic method for routine analysis of substances and compounds. Spectral interpretation of spectra is a labor-intensive process that provides important information about functional groups or bonds present in compounds and complex substances. In this paper, based on deep learning methods of convolutional neural networks, models were developed to determine the presence of 17 classes of functional groups or 72 classes of coupling oscillations in the FT-IR spectra. Using web scanning, the spectra of 14 361 FT-IR spectra of organic molecules were obtained. Several different variants of model architectures with different sizes of feature maps have been tested. Based on the Shapley additive explanations (SHAP) and gradient-weighted class activation mapping (GradCAM) methods, visualization tools have been developed for visualizing and highlighting the areas of absorption bands manifestation for corresponding functional groups or bonds in the spectrum. To determine 17 and 72 classes, the F1-weighted metric, which is the harmonic mean of the class' precision and class' recall weighted by class' fraction, reached 93 and 88%, respectively, when using data on the position of absorption maxima in the spectrum as an additional source layer. The resulting model can be used to facilitate the routine analysis of spectra for all areas such as organic chemistry, materials science, and biology, as well as to facilitate the preparation of the obtained experimental data for publication.
PMID:38281905 | DOI:10.1177/00037028241226732
Prediction of Disease-Free Survival in Breast Cancer using Deep Learning with Ultrasound and Mammography: A Multicenter Study
Clin Breast Cancer. 2024 Jan 17:S1526-8209(24)00006-5. doi: 10.1016/j.clbc.2024.01.005. Online ahead of print.
ABSTRACT
BACKGROUND: Breast cancer is a leading cause of cancer morbility and mortality in women. The possibility of overtreatment or inappropriate treatment exists, and methods for evaluating prognosis need to be improved.
MATERIALS AND METHODS: Patients (from January 2013 to December 2018) were recruited and divided into a training group and a testing group. All patients were followed for more than 3 years. Patients were divided into a disease-free group and a recurrence group based on follow up results at 3 years. Ultrasound (US) and mammography (MG) images were collected to establish deep learning models (DLMs) using ResNet50. Clinical data, MG, and US characteristics were collected to select independent prognostic factors using a cox proportional hazards model to establish a clinical model. DLM and independent prognostic factors were combined to establish a combined model.
RESULTS: In total, 1242 patients were included. Independent prognostic factors included age, neoadjuvant chemotherapy, HER2, orientation, blood flow, dubious calcification, and size. We established 5 models: the US DLM, MG DLM, US + MG DLM, clinical and combined model. The combined model using US images, MG images, and pathological, clinical, and radiographic characteristics had the highest predictive performance (AUC = 0.882 in the training group, AUC = 0.739 in the testing group).
CONCLUSION: DLMs based on the combination of US, MG, and clinical data have potential as predictive tools for breast cancer prognosis.
PMID:38281863 | DOI:10.1016/j.clbc.2024.01.005
Development of Efficient Brain Age Estimation Method Based on Regional Brain Volume From Structural Magnetic Resonance Imaging
Psychiatry Investig. 2024 Jan;21(1):37-43. doi: 10.30773/pi.2023.0183. Epub 2024 Jan 22.
ABSTRACT
OBJECTIVE: We aimed to create an efficient and valid predicting model which can estimate individuals' brain age by quantifying their regional brain volumes.
METHODS: A total of 2,560 structural brain magnetic resonance imaging (MRI) scans, along with demographic and clinical data, were obtained. Pretrained deep-learning models were employed to automatically segment the MRI data, which enabled fast calculation of regional brain volumes. Brain age gaps for each subject were estimated using volumetric values from predefined 12 regions of interest (ROIs): bilateral frontal, parietal, occipital, and temporal lobes, as well as bilateral hippocampus and lateral ventricles. A larger weight was given to the ROIs having a larger mean volumetric difference between the cognitively unimpaired (CU) and cognitively impaired group including mild cognitive impairment (MCI), and dementia groups. The brain age was predicted by adding or subtracting the brain age gap to the chronological age according to the presence or absence of the atrophy region.
RESULTS: The study showed significant differences in brain age gaps among CU, MCI, and dementia groups. Furthermore, the brain age gaps exhibited significant correlations with education level and measures of cognitive function, including the clinical dementia rating sum-of-boxes and the Korean version of the Mini-Mental State Examination.
CONCLUSION: The brain age that we developed enabled fast and efficient brain age calculations, and it also reflected individual's cognitive function and cognitive reserve. Thus, our study suggested that the brain age might be an important marker of brain health that can be used effectively in real clinical settings.
PMID:38281737 | DOI:10.30773/pi.2023.0183
ChatGPT in Maternal-Fetal Medicine practice: a primer for clinicians
Am J Obstet Gynecol MFM. 2024 Jan 25:101302. doi: 10.1016/j.ajogmf.2024.101302. Online ahead of print.
ABSTRACT
ChatGPT (Generative Pre-trained Transformer) is a language model developed by OpenAI and launched in November 2022, that generates human-like responses to prompts using deep-learning technology. The integration of large language processing models into healthcare has the potential to improve the accessibility of medical information for both patients and health professionals alike. In this commentary, we demonstrated the ability of ChatGPT to produce patient information sheets. Four board-certified maternal-fetal medicine attending physicians rated the accuracy and humanness of the information according to two predefined scales of accuracy and completeness. The median score for accuracy of information was rated 4.8 on a six-point scale and the median score for completeness of information was 2.2 on a three-point sclale for the five patient information leaflets generated by ChatGPT. Concerns raised included the omission of clinically important information for patient counseling in some patient information leaflets and the inability to verify the source of information as ChatGPT does not provide references. ChatGPT is a powerful tool that has the potential to enhance patient care but such a tool requires extensive validation and is perhaps best considered as an adjunct to clinical practice, rather than a tool to be utilized freely by the public with regards to healthcare information.
PMID:38281582 | DOI:10.1016/j.ajogmf.2024.101302
Automatic data augmentation to improve generalization of deep learning in H&E stained histopathology
Comput Biol Med. 2024 Jan 24;170:108018. doi: 10.1016/j.compbiomed.2024.108018. Online ahead of print.
ABSTRACT
In histopathology practice, scanners, tissue processing, staining, and image acquisition protocols vary from center to center, resulting in subtle variations in images. Vanilla convolutional neural networks are sensitive to such domain shifts. Data augmentation is a popular way to improve domain generalization. Currently, state-of-the-art domain generalization in computational pathology is achieved using a manually curated set of augmentation transforms. However, manual tuning of augmentation parameters is time-consuming and can lead to sub-optimal generalization performance. Meta-learning frameworks can provide efficient ways to find optimal training hyper-parameters, including data augmentation. In this study, we hypothesize that an automated search of augmentation hyper-parameters can provide superior generalization performance and reduce experimental optimization time. We select four state-of-the-art automatic augmentation methods from general computer vision and investigate their capacity to improve domain generalization in histopathology. We analyze their performance on data from 25 centers across two different tasks: tumor metastasis detection in lymph nodes and breast cancer tissue type classification. On tumor metastasis detection, most automatic augmentation methods achieve comparable performance to state-of-the-art manual augmentation. On breast cancer tissue type classification, the leading automatic augmentation method significantly outperforms state-of-the-art manual data augmentation.
PMID:38281317 | DOI:10.1016/j.compbiomed.2024.108018