Deep learning
A novel deep-learning based weighted feature fusion architecture for precise classification of pressure injury
Front Physiol. 2024 Feb 22;15:1304829. doi: 10.3389/fphys.2024.1304829. eCollection 2024.
ABSTRACT
Introduction: Precise classification has an important role in treatment of pressure injury (PI), while current machine-learning or deeplearning based methods of PI classification remain low accuracy. Methods: In this study, we developed a deeplearning based weighted feature fusion architecture for fine-grained classification, which combines a top-down and bottom-up pathway to fuse high-level semantic information and low-level detail representation. We validated it in our established database that consist of 1,519 images from multi-center clinical cohorts. ResNeXt was set as the backbone network. Results: We increased the accuracy of stage 3 PI from 60.3% to 76.2% by adding weighted feature pyramid network (wFPN). The accuracy for stage 1, 2, 4 PI were 0.870, 0.788, and 0.845 respectively. We found the overall accuracy, precision, recall, and F1-score of our network were 0.815, 0.808, 0.816, and 0.811 respectively. The area under the receiver operating characteristic curve was 0.940. Conclusions: Compared with current reported study, our network significantly increased the overall accuracy from 75% to 81.5% and showed great performance in predicting each stage. Upon further validation, our study will pave the path to the clinical application of our network in PI management.
PMID:38455845 | PMC:PMC10917912 | DOI:10.3389/fphys.2024.1304829
Retrieval augmented scientific claim verification
JAMIA Open. 2024 Feb 21;7(1):ooae021. doi: 10.1093/jamiaopen/ooae021. eCollection 2024 Apr.
ABSTRACT
OBJECTIVE: To automate scientific claim verification using PubMed abstracts.
MATERIALS AND METHODS: We developed CliVER, an end-to-end scientific Claim VERification system that leverages retrieval-augmented techniques to automatically retrieve relevant clinical trial abstracts, extract pertinent sentences, and use the PICO framework to support or refute a scientific claim. We also created an ensemble of three state-of-the-art deep learning models to classify rationale of support, refute, and neutral. We then constructed CoVERt, a new COVID VERification dataset comprising 15 PICO-encoded drug claims accompanied by 96 manually selected and labeled clinical trial abstracts that either support or refute each claim. We used CoVERt and SciFact (a public scientific claim verification dataset) to assess CliVER's performance in predicting labels. Finally, we compared CliVER to clinicians in the verification of 19 claims from 6 disease domains, using 189 648 PubMed abstracts extracted from January 2010 to October 2021.
RESULTS: In the evaluation of label prediction accuracy on CoVERt, CliVER achieved a notable F1 score of 0.92, highlighting the efficacy of the retrieval-augmented models. The ensemble model outperforms each individual state-of-the-art model by an absolute increase from 3% to 11% in the F1 score. Moreover, when compared with four clinicians, CliVER achieved a precision of 79.0% for abstract retrieval, 67.4% for sentence selection, and 63.2% for label prediction, respectively.
CONCLUSION: CliVER demonstrates its early potential to automate scientific claim verification using retrieval-augmented strategies to harness the wealth of clinical trial abstracts in PubMed. Future studies are warranted to further test its clinical utility.
PMID:38455840 | PMC:PMC10919922 | DOI:10.1093/jamiaopen/ooae021
Digital health technologies for high-risk pregnancy management: three case studies using Digilego framework
JAMIA Open. 2024 Mar 7;7(1):ooae022. doi: 10.1093/jamiaopen/ooae022. eCollection 2024 Apr.
ABSTRACT
OBJECTIVE: High-risk pregnancy (HRP) conditions such as gestational diabetes mellitus (GDM), hypertension (HTN), and peripartum depression (PPD) affect maternal and neonatal health. Patient engagement is critical for effective HRP management (HRPM). While digital technologies and analytics hold promise, emerging research indicates limited and suboptimal support offered by the highly prevalent pregnancy digital solutions within the commercial marketplace. In this article, we describe our efforts to develop a portfolio of digital products leveraging advances in social computing, data science, and digital health.
METHODS: We describe three studies that leverage core methods from Digilego digital health development framework to (1) conduct large-scale social media analysis (n = 55 301 posts) to understand population-level patterns in women's needs, (2) architect a digital repository to enable women curate HRP related information, and (3) develop a digital platform to support PPD prevention. We applied a combination of qualitative coding, machine learning, theory-mapping, and programmatic implementation of theory-linked digital features. Further, we conducted preliminary testing of the resulting products for acceptance with sample of pregnant women for GDM/HTN information management (n = 10) and PPD prevention (n = 30).
RESULTS: Scalable social computing models using deep learning classifiers with reasonable accuracy have allowed us to capture and examine psychosociobehavioral drivers associated with HRPM. Our work resulted in two digital health solutions, MyPregnancyChart and MomMind are developed. Initial evaluation of both tools indicates positive acceptance from potential end users. Further evaluation with MomMind revealed statistically significant improvements (P < .05) in PPD recognition and knowledge on how to seek PPD information.
DISCUSSION: Digilego framework provides an integrative methodological lens to gain micro-macro perspective on women's needs, theory integration, engagement optimization, as well as subsequent feature and content engineering, which can be organized into core and specialized digital pathways for women engagement in disease management.
CONCLUSION: Future works should focus on implementation and testing of digital solutions that facilitate women to capture, aggregate, preserve, and utilize, otherwise siloed, prenatal information artifacts for enhanced self-management of their high-risk conditions, ultimately leading to improved health outcomes.
PMID:38455839 | PMC:PMC10919928 | DOI:10.1093/jamiaopen/ooae022
Enhancing ASD detection accuracy: a combined approach of machine learning and deep learning models with natural language processing
Health Inf Sci Syst. 2024 Mar 6;12(1):20. doi: 10.1007/s13755-024-00281-y. eCollection 2024 Dec.
ABSTRACT
PURPOSE: The main aim of our study was to explore the utility of artificial intelligence (AI) in diagnosing autism spectrum disorder (ASD). The study primarily focused on using machine learning (ML) and deep learning (DL) models to detect ASD potential cases by analyzing text inputs, especially from social media platforms like Twitter. This is to overcome the ongoing challenges in ASD diagnosis, such as the requirement for specialized professionals and extensive resources. Timely identification, particularly in children, is essential to provide immediate intervention and support, thereby improving the quality of life for affected individuals.
METHODS: We employed natural language processing (NLP) techniques along with ML models like decision trees, extreme gradient boosting (XGB), k-nearest neighbors algorithm (KNN), and DL models such as recurrent neural networks (RNN), long short-term memory (LSTM), bidirectional long short-term memory (Bi-LSTM), bidirectional encoder representations from transformers (BERT and BERTweet). We extracted a dataset of 404,627 tweets from Twitter users using the platform's API and classified them based on whether they were written by individuals claiming to have ASD (ASD users) or by those without ASD (non-ASD users). From this dataset, we used a subset of 90,000 tweets (45,000 from each classification group) for the training and testing of these models.
RESULTS: The application of our AI models yielded promising results, with the predictive model reaching an accuracy of almost 88% when classifying texts that potentially originated from individuals with ASD.
CONCLUSION: Our research demonstrated the potential of using AI, particularly DL models, in enhancing the accuracy of ASD detection and diagnosis. This innovative approach signifies the critical role AI can play in advancing early diagnostic techniques, enabling better patient outcomes and underlining the importance of early identification of ASD, especially in children.
PMID:38455725 | PMC:PMC10917721 | DOI:10.1007/s13755-024-00281-y
Multimodal risk prediction with physiological signals, medical images and clinical notes
Heliyon. 2024 Feb 28;10(5):e26772. doi: 10.1016/j.heliyon.2024.e26772. eCollection 2024 Mar 15.
ABSTRACT
The broad adoption of electronic health record (EHR) systems brings us a tremendous amount of clinical data and thus provides opportunities to conduct data-based healthcare research to solve various clinical problems in the medical domain. Machine learning and deep learning methods are widely used in the medical informatics and healthcare domain due to their power to mine insights from raw data. When adapting deep learning models for EHR data, it is essential to consider its heterogeneous nature: EHR contains patient records from various sources including medical tests (e.g. blood test, microbiology test), medical imaging, diagnosis, medications, procedures, clinical notes, etc. Those modalities together provide a holistic view of patient health status and complement each other. Therefore, combining data from multiple modalities that are intrinsically different is challenging but intuitively promising in deep learning for EHR. To assess the expectations of multimodal data, we introduce a comprehensive fusion framework designed to integrate temporal variables, medical images, and clinical notes in EHR for enhanced performance in clinical risk prediction. Early, joint, and late fusion strategies are employed to combine data from various modalities effectively. We test the model with three predictive tasks: in-hospital mortality, long length of stay, and 30-day readmission. Experimental results show that multimodal models outperform uni-modal models in the tasks involved. Additionally, by training models with different input modality combinations, we calculate the Shapley value for each modality to quantify their contribution to multimodal performance. It is shown that temporal variables tend to be more helpful than CXR images and clinical notes in the three explored predictive tasks.
PMID:38455585 | PMC:PMC10918115 | DOI:10.1016/j.heliyon.2024.e26772
Exploring COVID-related relationship extraction: Contrasting data sources and analyzing misinformation
Heliyon. 2024 Feb 28;10(5):e26973. doi: 10.1016/j.heliyon.2024.e26973. eCollection 2024 Mar 15.
ABSTRACT
The COVID-19 pandemic presented an unparalleled challenge to global healthcare systems. A central issue revolves around the urgent need to swiftly amass critical biological and medical knowledge concerning the disease, its treatment, and containment. Remarkably, text data remains an underutilized resource in this context. In this paper, we delve into the extraction of COVID-related relations using transformer-based language models, including Bidirectional Encoder Representations from Transformers (BERT) and DistilBERT. Our analysis scrutinizes the performance of five language models, comparing information from both PubMed and Reddit, and assessing their ability to make novel predictions, including the detection of "misinformation." Key findings reveal that, despite inherent differences, both PubMed and Reddit data contain remarkably similar information, suggesting that Reddit can serve as a valuable resource for rapidly acquiring information during times of crisis. Furthermore, our results demonstrate that language models can unveil previously unseen entities and relations, a crucial aspect in identifying instances of misinformation.
PMID:38455555 | PMC:PMC10918203 | DOI:10.1016/j.heliyon.2024.e26973
Deep learning-based diffusion tensor cardiac magnetic resonance reconstruction: a comparison study
Sci Rep. 2024 Mar 7;14(1):5658. doi: 10.1038/s41598-024-55880-2.
ABSTRACT
In vivo cardiac diffusion tensor imaging (cDTI) is a promising Magnetic Resonance Imaging (MRI) technique for evaluating the microstructure of myocardial tissue in living hearts, providing insights into cardiac function and enabling the development of innovative therapeutic strategies. However, the integration of cDTI into routine clinical practice poses challenging due to the technical obstacles involved in the acquisition, such as low signal-to-noise ratio and prolonged scanning times. In this study, we investigated and implemented three different types of deep learning-based MRI reconstruction models for cDTI reconstruction. We evaluated the performance of these models based on the reconstruction quality assessment, the diffusion tensor parameter assessment as well as the computational cost assessment. Our results indicate that the models discussed in this study can be applied for clinical use at an acceleration factor (AF) of × 2 and × 4 , with the D5C5 model showing superior fidelity for reconstruction and the SwinMR model providing higher perceptual scores. There is no statistical difference from the reference for all diffusion tensor parameters at AF × 2 or most DT parameters at AF × 4 , and the quality of most diffusion tensor parameter maps is visually acceptable. SwinMR is recommended as the optimal approach for reconstruction at AF × 2 and AF × 4 . However, we believe that the models discussed in this study are not yet ready for clinical use at a higher AF. At AF × 8 , the performance of all models discussed remains limited, with only half of the diffusion tensor parameters being recovered to a level with no statistical difference from the reference. Some diffusion tensor parameter maps even provide wrong and misleading information.
PMID:38454072 | DOI:10.1038/s41598-024-55880-2
SHAP value-based ERP analysis (SHERPA): Increasing the sensitivity of EEG signals with explainable AI methods
Behav Res Methods. 2024 Mar 7. doi: 10.3758/s13428-023-02335-7. Online ahead of print.
ABSTRACT
Conventionally, event-related potential (ERP) analysis relies on the researcher to identify the sensors and time points where an effect is expected. However, this approach is prone to bias and may limit the ability to detect unexpected effects or to investigate the full range of the electroencephalography (EEG) signal. Data-driven approaches circumvent this limitation, however, the multiple comparison problem and the statistical correction thereof affect both the sensitivity and specificity of the analysis. In this study, we present SHERPA - a novel approach based on explainable artificial intelligence (XAI) designed to provide the researcher with a straightforward and objective method to find relevant latency ranges and electrodes. SHERPA is comprised of a convolutional neural network (CNN) for classifying the conditions of the experiment and SHapley Additive exPlanations (SHAP) as a post hoc explainer to identify the important temporal and spatial features. A classical EEG face perception experiment is employed to validate the approach by comparing it to the established researcher- and data-driven approaches. Likewise, SHERPA identified an occipital cluster close to the temporal coordinates for the N170 effect expected. Most importantly, SHERPA allows quantifying the relevance of an ERP for a psychological mechanism by calculating an "importance score". Hence, SHERPA suggests the presence of a negative selection process at the early and later stages of processing. In conclusion, our new method not only offers an analysis approach suitable in situations with limited prior knowledge of the effect in question but also an increased sensitivity capable of distinguishing neural processes with high precision.
PMID:38453828 | DOI:10.3758/s13428-023-02335-7
Application of deep-learning to the automatic segmentation and classification of lateral lymph nodes on ultrasound images of papillary thyroid carcinoma
Asian J Surg. 2024 Mar 6:S1015-9584(24)00401-9. doi: 10.1016/j.asjsur.2024.02.140. Online ahead of print.
ABSTRACT
PURPOSE: It is crucial to preoperatively diagnose lateral cervical lymph node (LN) metastases (LNMs) in papillary thyroid carcinoma (PTC) patients. This study aims to develop deep-learning models for the automatic segmentation and classification of LNM on original ultrasound images.
METHODS: This study included 1000 lateral cervical LN ultrasound images (consisting of 512 benign and 558 metastatic LNs) collected from 728 patients at the Chongqing General Hospital between March 2022 and July 2023. Three instance segmentation models (MaskRCNN, SOLO and Mask2Former) were constructed to segment and classify ultrasound images of lateral cervical LNs by recognizing each object individually and in a pixel-by-pixel manner. The segmentation and classification results of the three models were compared with an experienced sonographer in the test set.
RESULTS: Upon completion of a 200-epoch learning cycle, the loss among the three unique models became negligible. To evaluate the performance of the deep-learning models, the intersection over union threshold was set at 0.75. The mean average precision scores for MaskRCNN, SOLO and Mask2Former were 88.8%, 86.7% and 89.5%, respectively. The segmentation accuracies of the MaskRCNN, SOLO, Mask2Former models and sonographer were 85.6%, 88.0%, 89.5% and 82.3%, respectively. The classification AUCs of the MaskRCNN, SOLO, Mask2Former models and sonographer were 0.886, 0.869, 0.90.2 and 0.852 in the test set, respectively.
CONCLUSIONS: The deep learning models could automatically segment and classify lateral cervical LNs with an AUC of 0.92. This approach may serve as a promising tool to assist sonographers in diagnosing lateral cervical LNMs among patients with PTC.
PMID:38453612 | DOI:10.1016/j.asjsur.2024.02.140
Synthesizing Contrast-Enhanced MR Images from Noncontrast MR Images Using Deep Learning
AJNR Am J Neuroradiol. 2024 Mar 7;45(3):312-319. doi: 10.3174/ajnr.A8107.
ABSTRACT
BACKGROUND AND PURPOSE: Recent developments in deep learning methods offer a potential solution to the need for alternative imaging methods due to concerns about the toxicity of gadolinium-based contrast agents. The purpose of the study was to synthesize virtual gadolinium contrast-enhanced T1-weighted MR images from noncontrast multiparametric MR images in patients with primary brain tumors by using deep learning.
MATERIALS AND METHODS: We trained and validated a deep learning network by using MR images from 335 subjects in the Brain Tumor Segmentation Challenge 2019 training data set. A held out set of 125 subjects from the Brain Tumor Segmentation Challenge 2019 validation data set was used to test the generalization of the model. A residual inception DenseNet network, called T1c-ET, was developed and trained to simultaneously synthesize virtual contrast-enhanced T1-weighted (vT1c) images and segment the enhancing portions of the tumor. Three expert neuroradiologists independently scored the synthesized vT1c images by using a 3-point Likert scale, evaluating image quality and contrast enhancement against ground truth T1c images (1 = poor, 2 = good, 3 = excellent).
RESULTS: The synthesized vT1c images achieved structural similarity index, peak signal-to-noise ratio, and normalized mean square error scores of 0.91, 64.35, and 0.03, respectively. There was moderate interobserver agreement between the 3 raters, regarding the algorithm's performance in predicting contrast enhancement, with a Fleiss kappa value of 0.61. Our model was able to accurately predict contrast enhancement in 88.8% of the cases (scores of 2 to 3 on the 3-point scale).
CONCLUSIONS: We developed a novel deep learning architecture to synthesize virtual postcontrast enhancement by using only conventional noncontrast brain MR images. Our results demonstrate the potential of deep learning methods to reduce the need for gadolinium contrast in the evaluation of primary brain tumors.
PMID:38453408 | DOI:10.3174/ajnr.A8107
Global Research Evolution and Frontier Analysis of Artificial Intelligence in Brain Injury: A Bibliometric Analysis
Brain Res Bull. 2024 Mar 5:110920. doi: 10.1016/j.brainresbull.2024.110920. Online ahead of print.
ABSTRACT
Research on artificial intelligence for brain injury is currently a prominent area of scientific research. A significant amount of related literature has been accumulated in this field. This study aims to identify hotspots and clarify research resources by conducting literature metrology visualization analysis, providing valuable ideas and references for related fields. The research object of this paper consists of 3000 articles cited in the core database of Web of Science from 1998 to 2023. These articles are visualized and analyzed using VOSviewer and CiteSpace. The bibliometric analysis reveals a continuous increase in the number of articles published on this topic, particularly since 2016, indicating significant growth. The United States stands out as the leading country in artificial intelligence for brain injury, followed by China, which tends to catch up. The core research institutions are primarily universities in developed countries, but there is a lack of cooperation and communication between research groups. With the development of computer technology, the research in this field has shown strong wave characteristics, experiencing the early stage of applied research based on expert systems, the middle stage of prediction research based on machine learning, and the current phase of diversified research focused on deep learning. Artificial intelligence has innovative development prospects in brain injury, providing a new orientation for the treatment and auxiliary diagnosis in this field.
PMID:38453035 | DOI:10.1016/j.brainresbull.2024.110920
Development and validation of a prehospital termination of resuscitation (TOR) rule for Out - Of Hospital Cardiac Arrest (OHCA) Cases using general purpose artificial intelligence (AI)
Resuscitation. 2024 Mar 5:110165. doi: 10.1016/j.resuscitation.2024.110165. Online ahead of print.
ABSTRACT
BACKGROUND: Prehospital identification of futile resuscitation efforts (defined as a predicted probability of survival lower than 1%) for out-of-hospital cardiac arrest (OHCA) may reduce unnecessary transport. Reliable prediction variables for OHCA 'termination of resuscitation' (TOR) rules are needed to guide treatment decisions. The Universal TOR rule uses only three variables (Absence of Prehospital ROSC, Event not witnessed by EMS and no shock delivered on the scene) has been externally validated and is used by many EMS systems. Deep learning, an artificial intelligence (AI) platform is an attractive model to guide the development of TOR rule for OHCA. The purpose of this study was to assess the feasibility of developing an AI-TOR rule for neurologically favorable outcomes using general purpose AI and compare its performance to the Universal TOR rule.
METHODS: We identified OHCA cases of presumed cardiac etiology who were 18 years of age or older from 2016 to 2019 in the All-Japan Utstein Registry. We divided the dataset into 2 parts, the first half (2016- 2017) was used as a training dataset for rule development and second half (2018- 2019) for validation. The AI software (Prediction One®) created the model using the training dataset with internal cross-validation. It also evaluated the prediction accuracy and displayed the ranking of influencing variables. We performed validation using the second half cases and calculated the prediction model AUC. The top four of the 11 variables identified in the model were then selected as prognostic factors to be used in an AI-TOR rule, and sensitivity, specificity, positive predictive value, and negative predictive value were calculated from validation cohort. This was then compared to the performance of the Universal TOR rule using same dataset.
RESULTS: There were 504,561 OHCA cases, 18 years of age or older, 302,799 cases were presumed cardiac origin. Of these, 149,425 cases were used for the training dataset and 153,374 cases for the validation dataset. The model developed by AI using 11 variables had an AUC of 0.969, and its AUC for the validation dataset was 0.965. The top four influencing variables for neurologically favorable outcome were Prehospital ROSC, witnessed by EMS, Age (68 years old and younger) and nonasystole. The AUC calculated using the 4 variables for the AI-TOR rule was 0.953, and its AUC for the validation dataset was 0.952 (95%CI 0.949 -0.954). Of 80,198 patients in the validation cohort that satisfied all four criteria for the AI-TOR rule, 58 (0.07%) had a neurologically favorable one-month survival. The specificity of AI-TOR rule was 0.990, and the PPV was 0.999 for predicting lack of neurologically favorable survival, both the specificity and PPV were higher than that achieved with the universal TOR (0.959, 0.998).
CONCLUSIONS: The accuracy of prediction models using AI software to determine outcomes in OHCA was excellent and the AI-TOR rule's variables from prediction model performed better than the Universal TOR rule. External validation of our findings as well as further research into the utility of using AI platforms for TOR prediction in clinical practice is needed.
PMID:38452995 | DOI:10.1016/j.resuscitation.2024.110165
Proteome-scale movements and compartment connectivity during the eukaryotic cell cycle
Cell. 2024 Mar 1:S0092-8674(24)00180-6. doi: 10.1016/j.cell.2024.02.014. Online ahead of print.
ABSTRACT
Cell cycle progression relies on coordinated changes in the composition and subcellular localization of the proteome. By applying two distinct convolutional neural networks on images of millions of live yeast cells, we resolved proteome-level dynamics in both concentration and localization during the cell cycle, with resolution of ∼20 subcellular localization classes. We show that a quarter of the proteome displays cell cycle periodicity, with proteins tending to be controlled either at the level of localization or concentration, but not both. Distinct levels of protein regulation are preferentially utilized for different aspects of the cell cycle, with changes in protein concentration being mostly involved in cell cycle control and changes in protein localization in the biophysical implementation of the cell cycle program. We present a resource for exploring global proteome dynamics during the cell cycle, which will aid in understanding a fundamental biological process at a systems level.
PMID:38452761 | DOI:10.1016/j.cell.2024.02.014
Explainable AI-based Deep-SHAP for mapping the multivariate relationships between regional neuroimaging biomarkers and cognition
Eur J Radiol. 2024 Mar 2;174:111403. doi: 10.1016/j.ejrad.2024.111403. Online ahead of print.
ABSTRACT
BACKGROUND: Mild cognitive impairment (MCI)/Alzheimer's disease (AD) is associated with cognitive decline beyond normal aging and linked to the alterations of brain volume quantified by magnetic resonance imaging (MRI) and amyloid-beta (Aβ) quantified by positron emission tomography (PET). Yet, the complex relationships between these regional imaging measures and cognition in MCI/AD remain unclear. Explainable artificial intelligence (AI) may uncover such relationships.
METHOD: We integrate the AI-based deep learning neural network and Shapley additive explanations (SHAP) approaches and introduce the Deep-SHAP method to investigate the multivariate relationships between regional imaging measures and cognition. After validating this approach on simulated data, we apply it to real experimental data from MCI/AD patients.
RESULTS: Deep-SHAP significantly predicted cognition using simulated regional features and identified the ground-truth simulated regions as the most significant multivariate predictors. When applied to experimental MRI data, Deep-SHAP revealed that the insula, lateral occipital, medial frontal, temporal pole, and occipital fusiform gyrus are the primary contributors to global cognitive decline in MCI/AD. Furthermore, when applied to experimental amyloid Pittsburgh compound B (PiB)-PET data, Deep-SHAP identified the key brain regions for global cognitive decline in MCI/AD as the inferior temporal, parahippocampal, inferior frontal, supratemporal, and lateral frontal gray matter.
CONCLUSION: Deep-SHAP method uncovered the multivariate relationships between regional brain features and cognition, offering insights into the most critical modality-specific brain regions involved in MCI/AD mechanisms.
PMID:38452732 | DOI:10.1016/j.ejrad.2024.111403
3D-MRI super-resolution reconstruction using multi-modality based on multi-resolution CNN
Comput Methods Programs Biomed. 2024 Mar 5;248:108110. doi: 10.1016/j.cmpb.2024.108110. Online ahead of print.
ABSTRACT
BACKGROUND AND OBJECTIVE: High-resolution (HR) MR images provide rich structural detail to assist physicians in clinical diagnosis and treatment plan. However, it is arduous to acquire HR MRI due to equipment limitations, scanning time or patient comfort. Instead, HR MRI could be obtained through a number of computer assisted post-processing methods that have proven to be effective and reliable. This paper aims to develop a convolutional neural network (CNN) based super-resolution reconstruction framework for low-resolution (LR) T2w images.
METHOD: In this paper, we propose a novel multi-modal HR MRI generation framework based on deep learning techniques. Specifically, we construct a CNN based on multi-resolution analysis to learn an end-to-end mapping between LR T2w and HR T2w, where HR T1w is fed into the network to offer detailed a priori information to help generate HR T2w. Furthermore, a low-frequency filtering module is introduced to filter out the interference from HR-T1w during high-frequency information extraction. Based on the idea of multi-resolution analysis, detailed features extracted from HR T1w and LR T2w are fused at two scales in the network and then HR T2w is reconstructed by upsampling and dense connectivity module.
RESULTS: Extensive quantitative and qualitative evaluations demonstrate that the proposed method enhances the recovered HR T2w details and outperforms other state-of-the-art methods. In addition, the experimental results also suggest that our network has a lightweight structure and favorable generalization performance.
CONCLUSION: The results show that the proposed method is capable of reconstructing HR T2w with higher accuracy. Meanwhile, the super-resolution reconstruction results on other dataset illustrate the excellent generalization ability of the method.
PMID:38452685 | DOI:10.1016/j.cmpb.2024.108110
Medical long-tailed learning for imbalanced data: Bibliometric analysis
Comput Methods Programs Biomed. 2024 Feb 29;247:108106. doi: 10.1016/j.cmpb.2024.108106. Online ahead of print.
ABSTRACT
BACKGROUND: In the last decade, long-tail learning has become a popular research focus in deep learning applications in medicine. However, no scientometric reports have provided a systematic overview of this scientific field. We utilized bibliometric techniques to identify and analyze the literature on long-tailed learning in deep learning applications in medicine and investigate research trends, core authors, and core journals. We expanded our understanding of the primary components and principal methodologies of long-tail learning research in the medical field.
METHODS: Web of Science was utilized to collect all articles on long-tailed learning in medicine published until December 2023. The suitability of all retrieved titles and abstracts was evaluated. For bibliometric analysis, all numerical data were extracted. CiteSpace was used to create clustered and visual knowledge graphs based on keywords.
RESULTS: A total of 579 articles met the evaluation criteria. Over the last decade, the annual number of publications and citation frequency both showed significant growth, following a power-law and exponential trend, respectively. Noteworthy contributors to this field include Husanbir Singh Pannu, Fadi Thabtah, and Talha Mahboob Alam, while leading journals such as IEEE ACCESS, COMPUTERS IN BIOLOGY AND MEDICINE, IEEE TRANSACTIONS ON MEDICAL IMAGING, and COMPUTERIZED MEDICAL IMAGING AND GRAPHICS have emerged as pivotal platforms for disseminating research in this area. The core of long-tailed learning research within the medical domain is encapsulated in six principal themes: deep learning for imbalanced data, model optimization, neural networks in image analysis, data imbalance in health records, CNN in diagnostics and risk assessment, and genetic information in disease mechanisms.
CONCLUSION: This study summarizes recent advancements in applying long-tail learning to deep learning in medicine through bibliometric analysis and visual knowledge graphs. It explains new trends, sources, core authors, journals, and research hotspots. Although this field has shown great promise in medical deep learning research, our findings will provide pertinent and valuable insights for future research and clinical practice.
PMID:38452661 | DOI:10.1016/j.cmpb.2024.108106
Peanut origin traceability: A hybrid neural network combining an electronic nose system and a hyperspectral system
Food Chem. 2024 Mar 2;447:138915. doi: 10.1016/j.foodchem.2024.138915. Online ahead of print.
ABSTRACT
Peanuts, sourced from various regions, exhibit noticeable differences in quality owing to the impact of their natural environments. This study proposes a fast and nondestructive detection method to identify peanut quality by combining an electronic nose system with a hyperspectral system. First, the electronic nose and hyperspectral systems are used to gather gas and spectral information from peanuts. Second, a module for extracting gas and spectral information is designed, combining the lightweight multi-head transposed attention mechanism (LMTA) and convolutional computation. The fusion of gas and spectral information is achieved through matrix combination and lightweight convolution. A hybrid neural network, named UnitFormer, is designed based on the information extraction and fusion processes. UnitFormer demonstrates an accuracy of 99.06 %, a precision of 99.12 %, and a recall of 99.05 %. In conclusion, UnitFormer effectively distinguishes quality differences among peanuts from various regions, offering an effective technological solution for quality supervision in the food market.
PMID:38452539 | DOI:10.1016/j.foodchem.2024.138915
Automated entry of paper-based patient-reported outcomes: Applying deep learning to the Japanese orthopaedic association back pain evaluation questionnaire
Comput Biol Med. 2024 Feb 19;172:108197. doi: 10.1016/j.compbiomed.2024.108197. Online ahead of print.
ABSTRACT
BACKGROUND: Health-related patient-reported outcomes (HR-PROs) are crucial for assessing the quality of life among individuals experiencing low back pain. However, manual data entry from paper forms, while convenient for patients, imposes a considerable tallying burden on collectors. In this study, we developed a deep learning (DL) model capable of automatically reading these paper forms.
METHODS: We employed the Japanese Orthopaedic Association Back Pain Evaluation Questionnaire, a globally recognized assessment tool for low back pain. The questionnaire comprised 25 low back pain-related multiple-choice questions and three pain-related visual analog scales (VASs). We collected 1305 forms from an academic medical center as the training set, and 483 forms from a community medical center as the test set. The performance of our DL model for multiple-choice questions was evaluated using accuracy as a categorical classification task. The performance for VASs was evaluated using the correlation coefficient and absolute error as regression tasks.
RESULT: In external validation, the mean accuracy of the categorical questions was 0.997. When outputs for categorical questions with low probability (threshold: 0.9996) were excluded, the accuracy reached 1.000 for the remaining 65 % of questions. Regarding the VASs, the average of the correlation coefficients was 0.989, with the mean absolute error being 0.25.
CONCLUSION: Our DL model demonstrated remarkable accuracy and correlation coefficients when automatic reading paper-based HR-PROs during external validation.
PMID:38452472 | DOI:10.1016/j.compbiomed.2024.108197
Transcriptomic signature of cancer cachexia by integration of machine learning, literature mining and meta-analysis
Comput Biol Med. 2024 Feb 28;172:108233. doi: 10.1016/j.compbiomed.2024.108233. Online ahead of print.
ABSTRACT
BACKGROUND: Cancer cachexia is a severe metabolic syndrome marked by skeletal muscle atrophy. A successful clinical intervention for cancer cachexia is currently lacking. The study of cachexia mechanisms is largely based on preclinical animal models and the availability of high-throughput transcriptomic datasets of cachectic mouse muscles is increasing through the extensive use of next generation sequencing technologies.
METHODS: Cachectic mouse muscle transcriptomic datasets of ten different studies were combined and mined by seven attribute weighting models, which analysed both categorical variables and numerical variables. The transcriptomic signature of cancer cachexia was identified by attribute weighting algorithms and was used to evaluate the performance of eleven pattern discovery models. The signature was employed to find the best combination of drugs (drug repurposing) for developing cancer cachexia treatment strategies, as well as to evaluate currently used cachexia drugs by literature mining.
RESULTS: Attribute weighting algorithms ranked 26 genes as the transcriptomic signature of muscle from mice with cancer cachexia. Deep Learning and Random Forest models performed better in differentiating cancer cachexia cases based on muscle transcriptomic data. Literature mining revealed that a combination of melatonin and infliximab has negative interactions with 2 key genes (Rorc and Fbxo32) upregulated in the transcriptomic signature of cancer cachexia in muscle.
CONCLUSIONS: The integration of machine learning, meta-analysis and literature mining was found to be an efficient approach to identifying a robust transcriptomic signature for cancer cachexia, with implications for improving clinical diagnosis and management of this condition.
PMID:38452471 | DOI:10.1016/j.compbiomed.2024.108233
Bioinspiration from bats and new paradigms for autonomy in natural environments
Bioinspir Biomim. 2024 Mar 7. doi: 10.1088/1748-3190/ad311e. Online ahead of print.
ABSTRACT
Achieving autonomous operation in complex natural environment remains an unsolved challenge. Conventional engineering approaches to this problem have focused on collecting large amounts of sensory data that are used to create detailed digital models of the environment. However, this only postpones solving the challenge of identifying the relevant sensory information and linking it to action control to the domain of the digital world model. Furthermore, it imposes high demands in terms of computing power and introduces large processing latencies that hamper autonomous real-time performance. Certain species of bats that are able to navigate and hunt their prey in dense vegetation could be a biological model system for an alternative approach to addressing the fundamental issues associated with autonomy in complex natural environments. Bats navigating in dense vegetation rely on clutter echoes, i.e., signals that consist of unresolved contributions from many scatters. Yet, the animals are able to extract the relevant information from these input signals with brains that are often less than one gram in mass. Pilot results indicate that information relevant to location identification and passageway finding can be directly obtained from clutter echoes, opening up the possibility that the bats' skill can be replicated in man-made autonomous systems.
.
PMID:38452384 | DOI:10.1088/1748-3190/ad311e