Deep learning
Dyslexia Data Consortium Repository: A Data Sharing and Delivery Platform for Research
Brain Inform (2023). 2023 Aug;13974:167-178. doi: 10.1007/978-3-031-43075-6_15. Epub 2023 Sep 13.
ABSTRACT
Specific learning disability of reading, or dyslexia, affects 5-17% of the population in the United States. Research on the neurobiology of dyslexia has included studies with relatively small sample sizes across research sites, thus limiting inference and the application of novel methods, such as deep learning. To address these issues and facilitate open science, we developed an online platform for data-sharing and advanced research programs to enhance opportunities for replication by providing researchers with secondary data that can be used in their research (https://www.dyslexiadata.org). This platform integrates a set of well-designed machine learning algorithms and tools to generate secondary datasets, such as cortical thickness, as well as regional brain volume metrics that have been consistently associated with dyslexia. Researchers can access shared data to address fundamental questions about dyslexia and development, replicate research findings, apply new methods, and educate the next generation of researchers. The overarching goal of this platform is to advance our understanding of a disorder that has significant academic, social, and economic impacts on children, their families, and society.
PMID:38352916 | PMC:PMC10859776 | DOI:10.1007/978-3-031-43075-6_15
Artificial intelligence in Immuno-genetics
Bioinformation. 2024 Jan 31;20(1):29-35. doi: 10.6026/973206300200029. eCollection 2024.
ABSTRACT
Rapid advancements in the field of artificial intelligence (AI) have opened up unprecedented opportunities to revolutionize various scientific domains, including immunology and genetics. Therefore, it is of interest to explore the emerging applications of AI in immunology and genetics, with the objective of enhancing our understanding of the dynamic intricacies of the immune system, disease etiology, and genetic variations. Hence, the use of AI methodologies in immunological and genetic datasets, thereby facilitating the development of innovative approaches in the realms of diagnosis, treatment, and personalized medicine is reviewed.
PMID:38352901 | PMC:PMC10859949 | DOI:10.6026/973206300200029
A machine learning and deep learning-based integrated multi-omics technique for leukemia prediction
Heliyon. 2024 Feb 1;10(3):e25369. doi: 10.1016/j.heliyon.2024.e25369. eCollection 2024 Feb 15.
ABSTRACT
In recent years, scientific data on cancer has expanded, providing potential for a better understanding of malignancies and improved tailored care. Advances in Artificial Intelligence (AI) processing power and algorithmic development position Machine Learning (ML) and Deep Learning (DL) as crucial players in predicting Leukemia, a blood cancer, using integrated multi-omics technology. However, realizing these goals demands novel approaches to harness this data deluge. This study introduces a novel Leukemia diagnosis approach, analyzing multi-omics data for accuracy using ML and DL algorithms. ML techniques, including Random Forest (RF), Naive Bayes (NB), Decision Tree (DT), Logistic Regression (LR), Gradient Boosting (GB), and DL methods such as Recurrent Neural Networks (RNN) and Feedforward Neural Networks (FNN) are compared. GB achieved 97 % accuracy in ML, while RNN outperformed by achieving 98 % accuracy in DL. This approach filters unclassified data effectively, demonstrating the significance of DL for leukemia prediction. The testing validation was based on 17 different features such as patient age, sex, mutation type, treatment methods, chromosomes, and others. Our study compares ML and DL techniques and chooses the best technique that gives optimum results. The study emphasizes the implications of high-throughput technology in healthcare, offering improved patient care.
PMID:38352790 | PMC:PMC10862685 | DOI:10.1016/j.heliyon.2024.e25369
Performance of convolutional neural networks for the classification of brain tumors using magnetic resonance imaging
Heliyon. 2024 Feb 2;10(3):e25468. doi: 10.1016/j.heliyon.2024.e25468. eCollection 2024 Feb 15.
ABSTRACT
Brain tumors are a diverse group of neoplasms that are challenging to detect and classify due to their varying characteristics. Deep learning techniques have proven to be effective in tumor classification. However, there is a lack of studies that compare these techniques using a common methodology. This work aims to analyze the performance of convolutional neural networks in the classification of brain tumors. We propose a network consisting of a few convolutional layers, batch normalization, and max-pooling. Then, we explore recent deep architectures, such as VGG, ResNet, EfficientNet, or ConvNeXt. The study relies on two magnetic resonance imaging datasets with over 3000 images of three types of tumors -gliomas, meningiomas, and pituitary tumors-, as well as images without tumors. We determine the optimal hyperparameters of the networks using the training and validation sets. The training and test sets are used to assess the performance of the models from different perspectives, including training from scratch, data augmentation, transfer learning, and fine-tuning. The experiments are performed using the TensorFlow and Keras libraries in Python. We compare the accuracy of the models and analyze their complexity based on the capacity of the networks, their training times, and image throughput. Several networks achieve high accuracy rates on both datasets, with the best model achieving 98.7% accuracy, which is on par with state-of-the-art methods. The average precision for each type of tumor is 94.3% for gliomas, 93.8% for meningiomas, 97.9% for pituitary tumors, and 95.3% for images without tumors. VGG is the largest model with over 171 million parameters, whereas MobileNet and EfficientNetB0 are the smallest ones with 3.2 and 5.9 million parameters, respectively. These two neural networks are also the fastest to train with 23.7 and 25.4 seconds per epoch, respectively. On the other hand, ConvNext is the slowest model with 58.2 seconds per epoch. Our custom model obtained the highest image throughput with 234.37 images per second, followed by MobileNet with 226 images per second. ConvNext obtained the smallest throughput with 97.35 images per second. ResNet, MobileNet, and EfficientNet are the most accurate networks, with MobileNet and EfficientNet demonstrating superior performance in terms of complexity. Most models achieve the best accuracy using transfer learning followed by a fine-tuning step. However, data augmentation does not contribute to increasing the accuracy of the models in general.
PMID:38352765 | PMC:PMC10862681 | DOI:10.1016/j.heliyon.2024.e25468
Bidirectional feature pyramid attention-based temporal convolutional network model for motor imagery electroencephalogram classification
Front Neurorobot. 2024 Jan 30;18:1343249. doi: 10.3389/fnbot.2024.1343249. eCollection 2024.
ABSTRACT
INTRODUCTION: As an interactive method gaining popularity, brain-computer interfaces (BCIs) aim to facilitate communication between the brain and external devices. Among the various research topics in BCIs, the classification of motor imagery using electroencephalography (EEG) signals has the potential to greatly improve the quality of life for people with disabilities.
METHODS: This technology assists them in controlling computers or other devices like prosthetic limbs, wheelchairs, and drones. However, the current performance of EEG signal decoding is not sufficient for real-world applications based on Motor Imagery EEG (MI-EEG). To address this issue, this study proposes an attention-based bidirectional feature pyramid temporal convolutional network model for the classification task of MI-EEG. The model incorporates a multi-head self-attention mechanism to weigh significant features in the MI-EEG signals. It also utilizes a temporal convolution network (TCN) to separate high-level temporal features. The signals are enhanced using the sliding-window technique, and channel and time-domain information of the MI-EEG signals is extracted through convolution.
RESULTS: Additionally, a bidirectional feature pyramid structure is employed to implement attention mechanisms across different scales and multiple frequency bands of the MI-EEG signals. The performance of our model is evaluated on the BCI Competition IV-2a dataset and the BCI Competition IV-2b dataset, and the results showed that our model outperformed the state-of-the-art baseline model, with an accuracy of 87.5 and 86.3% for the subject-dependent, respectively.
DISCUSSION: In conclusion, the BFATCNet model offers a novel approach for EEG-based motor imagery classification in BCIs, effectively capturing relevant features through attention mechanisms and temporal convolutional networks. Its superior performance on the BCI Competition IV-2a and IV-2b datasets highlights its potential for real-world applications. However, its performance on other datasets may vary, necessitating further research on data augmentation techniques and integration with multiple modalities to enhance interpretability and generalization. Additionally, reducing computational complexity for real-time applications is an important area for future work.
PMID:38352723 | PMC:PMC10861766 | DOI:10.3389/fnbot.2024.1343249
Derivation, External Validation and Clinical Implications of a deep learning approach for intracranial pressure estimation using non-cranial waveform measurements
medRxiv. 2024 Jan 30:2024.01.30.24301974. doi: 10.1101/2024.01.30.24301974. Preprint.
ABSTRACT
IMPORTANCE: Increased intracranial pressure (ICP) is associated with adverse neurological outcomes, but needs invasive monitoring.
OBJECTIVE: Development and validation of an AI approach for detecting increased ICP (aICP) using only non-invasive extracranial physiological waveform data.
DESIGN: Retrospective diagnostic study of AI-assisted detection of increased ICP. We developed an AI model using exclusively extracranial waveforms, externally validated it and assessed associations with clinical outcomes.
SETTING: MIMIC-III Waveform Database (2000-2013), a database derived from patients admitted to an ICU in an academic Boston hospital, was used for development of the aICP model, and to report association with neurologic outcomes. Data from Mount Sinai Hospital (2020-2022) in New York City was used for external validation.
PARTICIPANTS: Patients were included if they were older than 18 years, and were monitored with electrocardiograms, arterial blood pressure, respiratory impedance plethysmography and pulse oximetry. Patients who additionally had intracranial pressure monitoring were used for development (N=157) and external validation (N=56). Patients without intracranial monitors were used for association with outcomes (N=1694).
EXPOSURES: Extracranial waveforms including electrocardiogram, arterial blood pressure, plethysmography and SpO 2 .
MAIN OUTCOMES AND MEASURES: Intracranial pressure > 15 mmHg. Measures were Area under receiver operating characteristic curves (AUROCs), sensitivity, specificity, and accuracy at threshold of 0.5. We calculated odds ratios and p-values for phenotype association.
RESULTS: The AUROC was 0.91 (95% CI, 0.90-0.91) on testing and 0.80 (95% CI, 0.80-0.80) on external validation. aICP had accuracy, sensitivity, and specificity of 73.8% (95% CI, 72.0%-75.6%), 99.5% (95% CI 99.3%-99.6%), and 76.9% (95% CI, 74.0-79.8%) on external validation. A ten-percentile increment was associated with stroke (OR=2.12; 95% CI, 1.27-3.13), brain malignancy (OR=1.68; 95% CI, 1.09-2.60), subdural hemorrhage (OR=1.66; 95% CI, 1.07-2.57), intracerebral hemorrhage (OR=1.18; 95% CI, 1.07-1.32), and procedures like percutaneous brain biopsy (OR=1.58; 95% CI, 1.15-2.18) and craniotomy (OR = 1.43; 95% CI, 1.12-1.84; P < 0.05 for all).
CONCLUSIONS AND RELEVANCE: aICP provides accurate, non-invasive estimation of increased ICP, and is associated with neurological outcomes and neurosurgical procedures in patients without intracranial monitoring.
PMID:38352556 | PMC:PMC10863000 | DOI:10.1101/2024.01.30.24301974
Piscis: a novel loss estimator of the F1 score enables accurate spot detection in fluorescence microscopy images via deep learning
bioRxiv. 2024 Jan 31:2024.01.31.578123. doi: 10.1101/2024.01.31.578123. Preprint.
ABSTRACT
Single-molecule RNA fluorescence in situ hybridization (RNA FISH)-based spatial transcriptomics methods have enabled the accurate quantification of gene expression at single-cell resolution by visualizing transcripts as diffraction-limited spots. While these methods generally scale to large samples, image analysis remains challenging, often requiring manual parameter tuning. We present Piscis, a fully automatic deep learning algorithm for spot detection trained using a novel loss function, the SmoothF1 loss, that approximates the F1 score to directly penalize false positives and false negatives but remains differentiable and hence usable for training by deep learning approaches. Piscis was trained and tested on a diverse dataset composed of 358 manually annotated experimental RNA FISH images representing multiple cell types and 240 additional synthetic images. Piscis outperforms other state-of-the-art spot detection methods, enabling accurate, high-throughput analysis of RNA FISH-derived imaging data without the need for manual parameter tuning.
PMID:38352551 | PMC:PMC10862914 | DOI:10.1101/2024.01.31.578123
RNA3DB: A dataset for training and benchmarking deep learning models for RNA structure prediction
bioRxiv. 2024 Feb 2:2024.01.30.578025. doi: 10.1101/2024.01.30.578025. Preprint.
ABSTRACT
With advances in protein structure prediction thanks to deep learning models like AlphaFold, RNA structure prediction has recently received increased attention from deep learning researchers. RNAs introduce substantial challenges due to the sparser availability and lower structural diversity of the experimentally resolved RNA structures in comparison to protein structures. These challenges are often poorly addressed by the existing literature, many of which report inflated performance due to using training and testing sets with significant structural overlap. Further, the most recent Critical Assessment of Structure Prediction (CASP15) has shown that deep learning models for RNA structure are currently outperformed by traditional methods. In this paper we present RNA3DB, a dataset of structured RNAs, derived from the Protein Data Bank (PDB), that is designed for training and benchmarking deep learning models. Our dataset clusters RNA 3D chains into distinct groups that are non-redundant both with regard to sequence as well as structure, providing a robust way of dividing training, validation, and testing sets. For the PDB RNA chains as of 2024-01-10, RNA3DB produces 118 independent components with a total of 1,645 distinct RNA sequences with 21,005 reported crystal structures, representing 216 different Rfam structural families. A potential split consists of a training set of 1,152 RNA sequences, with 9,832 experimentally determined structures that belong to 169 distinct RNA structural Rfam families (at an E-value of 10 -3 ), and a test set of 493 RNA sequences with 1,344 structures that belong to 47 structural Rfam families. This split guarantees that all test RNA chains are distinct by sequence and structure from those in the training set. We provide the methodology along with the source-code, with the goal of creating a reproducible and customizable tool for RNA structure prediction.
PMID:38352531 | PMC:PMC10862857 | DOI:10.1101/2024.01.30.578025
A deep-learning strategy to identify cell types across species from high-density extracellular recordings
bioRxiv. 2024 Jan 31:2024.01.30.577845. doi: 10.1101/2024.01.30.577845. Preprint.
ABSTRACT
High-density probes allow electrophysiological recordings from many neurons simultaneously across entire brain circuits but fail to determine each recorded neuron's cell type. Here, we develop a strategy to identify cell types from extracellular recordings in awake animals, opening avenues to unveil the computational roles of neurons with distinct functional, molecular, and anatomical properties. We combine optogenetic activation and pharmacology using the cerebellum as a testbed to generate a curated ground-truth library of electrophysiological properties for Purkinje cells, molecular layer interneurons, Golgi cells, and mossy fibers. We train a semi-supervised deep-learning classifier that predicts cell types with greater than 95% accuracy based on waveform, discharge statistics, and layer of the recorded neuron. The classifier's predictions agree with expert classification on recordings using different probes, in different laboratories, from functionally distinct cerebellar regions, and across animal species. Our approach provides a general blueprint for cell-type identification from extracellular recordings across the brain.
PMID:38352514 | PMC:PMC10862837 | DOI:10.1101/2024.01.30.577845
Evaluating the Diagnostic Potential of Connected Speech for Benign Laryngeal Disease Using Deep Learning Analysis
J Voice. 2024 Feb 12:S0892-1997(24)00018-3. doi: 10.1016/j.jvoice.2024.01.015. Online ahead of print.
ABSTRACT
OBJECTIVES: This study aimed to evaluate the performance of artificial intelligence (AI) models using connected speech and vowel sounds in detecting benign laryngeal diseases.
STUDY DESIGN: Retrospective.
METHODS: Voice samples from 772 patients, including 502 with normal voices and 270 with vocal cord polyps, cysts, or nodules, were analyzed. We employed deep learning architectures, including convolutional neural networks (CNNs) and time series models, to process the speech data. The primary endpoint was the area under the receiver's operating characteristic curve for binary classification.
RESULTS: CNN models analyzing speech segments significantly outperformed those using vowel sounds in distinguishing patients with and without benign laryngeal diseases. The best-performing CNN model achieved areas under the receiver operating characteristic curve of 0.895 and 0.845 for speech and vowel sounds, respectively. Correlations between AI-generated disease probabilities and perceptual assessments were more pronounced in the connected-speech analyses. However, the time series models performed worse than the CNNs.
CONCLUSION: Connected speech analysis is more effective than traditional vowel sound analysis for the diagnosis of laryngeal voice disorders. This study highlights the potential of AI technologies in enhancing the diagnostic capabilities of speech, advocating further exploration, and validation in this field.
PMID:38350806 | DOI:10.1016/j.jvoice.2024.01.015
Recycling of straw-biochar-biogas-electricity for sustainable food production pathways: Toward an integrated modeling approach
Sci Total Environ. 2024 Feb 11:170804. doi: 10.1016/j.scitotenv.2024.170804. Online ahead of print.
ABSTRACT
As global greenhouse gas emissions increase and fossil energy sources decline dramatically, the energy transition is at the heart of many countries' development initiatives. As a biomass resource, straw plays a positive role in energy transformation and environmental improvement. However, there is still a challenge to explore the best options and models for straw production and utilization of green and efficient biomass energy in agricultural systems. This study establishes an economic-environmental-resource synergistic Straw Green recycling optimization model based on straw-electricity-biochar-biogas core (Straw Green recycling optimization model, SGROM). Firstly, we explore the effects of biochar return to the field on crop yield and greenhouse gas emission by Meta-analysis method, and on this basis, we construct SGROM to weigh the three objectives of economic-greenhouse gas emission-resource utilization, and explore the best allocation ratio between four utilization methods of straw: power generation, biochar preparation, biogas and derivatives preparation and sale, so as to obtain a straw recycling and efficient low-carbon utilization model. Exploring the response of straw green utilization patterns to crop market prices with the help of deep learning methods, SGROM has been applied to the main grain producing areas in the Sanjiang Plain of China, and the results of comparison with the traditional straw utilization (TSU) model show that the greenhouse gas emissions per unit of production value of SGROM are 19.66 % lower than that of TSU model, the electricity consumption is saved by 2.00 %, and the optimal ratios of straw for power generation, biogas and biochar production, and sale are 1.00 %, 10.75 %, 62.11 % and 26.14 %. The economic benefits and total greenhouse gas emissions of the integrated straw utilization mode are better than those of the single straw utilization mode, proving the superiority of SGROM in optimizing the straw utilization mode.
PMID:38350576 | DOI:10.1016/j.scitotenv.2024.170804
Towards a diagnostic tool for neurological gait disorders in childhood combining 3D gait kinematics and deep learning
Comput Biol Med. 2024 Feb 3;171:108095. doi: 10.1016/j.compbiomed.2024.108095. Online ahead of print.
ABSTRACT
Gait abnormalities are frequent in children and can be caused by different pathologies, such as cerebral palsy, neuromuscular disease, toe walker syndrome, etc. Analysis of the "gait pattern" (i.e., the way the person walks) using 3D analysis provides highly relevant clinical information. This information is used to guide therapeutic choices; however, it is underused in diagnostic processes, probably because of the lack of standardization of data collection methods. Therefore, 3D gait analysis is currently used as an assessment rather than a diagnostic tool. In this work, we aimed to determine if deep learning could be combined with 3D gait analysis data to diagnose gait disorders in children. We tested the diagnostic accuracy of deep learning methods combined with 3D gait analysis data from 371 children (148 with unilateral cerebral palsy, 60 with neuromuscular disease, 19 toe walkers, 60 with bilateral cerebral palsy, 25 stroke, and 59 typically developing children), with a total of 6400 gait cycles. We evaluated the accuracy, sensitivity, specificity, F1 score, Area Under the Curve (AUC) score, and confusion matrix of the predictions by ResNet, LSTM, and InceptionTime deep learning architectures for time series data. The deep learning-based models had good to excellent diagnostic accuracy (ranging from 0.77 to 0.99) for discrimination between healthy and pathological gait, discrimination between different etiologies of pathological gait (binary and multi-classification); and determining stroke onset time. LSTM performed best overall. This study revealed that the gait pattern contains specific, pathology-related information. These results open the way for an extension of 3D gait analysis from evaluation to diagnosis. Furthermore, the method we propose is a data-driven diagnostic model that can be trained and used without human intervention or expert knowledge. Furthermore, the method could be used to distinguish gait-related pathologies and their onset times beyond those studied in this research.
PMID:38350399 | DOI:10.1016/j.compbiomed.2024.108095
Cross comparison representation learning for semi-supervised segmentation of cellular nuclei in immunofluorescence staining
Comput Biol Med. 2024 Feb 6;171:108102. doi: 10.1016/j.compbiomed.2024.108102. Online ahead of print.
ABSTRACT
The morphological analysis of cells from optical images is vital for interpreting brain function in disease states. Extracting comprehensive cell morphology from intricate backgrounds, common in neural and some medical images, poses a significant challenge. Due to the huge workload of manual recognition, automated neuron cell segmentation using deep learning algorithms with labeled data is integral to neural image analysis tools. To combat the high cost of acquiring labeled data, we propose a novel semi-supervised cell segmentation algorithm for immunofluorescence-stained cell image datasets (ISC), utilizing a mean-teacher semi-supervised learning framework. We include a "cross comparison representation learning block" to enhance the teacher-student model comparison on high-dimensional channels, thereby improving feature compactness and separability, which results in the extraction of higher-dimensional features from unlabeled data. We also suggest a new network, the Multi Pooling Layer Attention Dense Network (MPAD-Net), serving as the backbone of the student model to augment segmentation accuracy. Evaluations on the immunofluorescence staining datasets and the public CRAG dataset illustrate our method surpasses other top semi-supervised learning methods, achieving average Jaccard, Dice and Normalized Surface Dice (NSD) indicators of 83.22%, 90.95% and 81.90% with only 20% labeled data. The datasets and code are available on the website at https://github.com/Brainsmatics/CCRL.
PMID:38350398 | DOI:10.1016/j.compbiomed.2024.108102
Predicting Drug-Protein Interactions through Branch-Chain Mining and multi-dimensional attention network
Comput Biol Med. 2024 Feb 7;171:108127. doi: 10.1016/j.compbiomed.2024.108127. Online ahead of print.
ABSTRACT
Identifying drug-protein interactions (DPIs) is crucial in drug discovery and repurposing. Computational methods for precise DPI identification can expedite development timelines and reduce expenses compared with conventional experimental methods. Lately, deep learning techniques have been employed for predicting DPIs, enhancing these processes. Nevertheless, the limitations observed in prior studies, where many extract features from complete drug and protein entities, overlooking the crucial theoretical foundation that pharmacological responses are often correlated with specific substructures, can lead to poor predictive performance. Furthermore, certain substructure-focused research confines its exploration to a solitary fragment category, such as a functional group. In this study, addressing these constraints, we present an end-to-end framework termed BCMMDA for predicting DPIs. The framework considers various substructure types, including branch chains, common substructures, and specific fragments. We designed a specific feature learning module by combining our proposed multi-dimensional attention mechanism with convolutional neural networks (CNNs). Deep CNNs assist in capturing the synergistic effects among these fragment sets, enabling the extraction of relevant features of drugs and proteins. Meanwhile, the multi-dimensional attention mechanism refines the relationship between drug and protein features by assigning attention vectors to each drug compound and amino acid. This mechanism empowers the model to further concentrate on pivotal substructures and elements, thereby improving its ability to identify essential interactions in DPI prediction. We evaluated the performance of BCMMDA on four well-known benchmark datasets. The results indicated that BCMMDA outperformed state-of-the-art baseline models, demonstrating significant improvement in performance.
PMID:38350397 | DOI:10.1016/j.compbiomed.2024.108127
Deep learning predicts prevalent and incident Parkinson's disease from UK Biobank fundus imaging
Sci Rep. 2024 Feb 13;14(1):3637. doi: 10.1038/s41598-024-54251-1.
ABSTRACT
Parkinson's disease is the world's fastest-growing neurological disorder. Research to elucidate the mechanisms of Parkinson's disease and automate diagnostics would greatly improve the treatment of patients with Parkinson's disease. Current diagnostic methods are expensive and have limited availability. Considering the insidious and preclinical onset and progression of the disease, a desirable screening should be diagnostically accurate even before the onset of symptoms to allow medical interventions. We highlight retinal fundus imaging, often termed a window to the brain, as a diagnostic screening modality for Parkinson's disease. We conducted a systematic evaluation of conventional machine learning and deep learning techniques to classify Parkinson's disease from UK Biobank fundus imaging. Our results suggest Parkinson's disease individuals can be differentiated from age and gender-matched healthy subjects with 68% accuracy. This accuracy is maintained when predicting either prevalent or incident Parkinson's disease. Explainability and trustworthiness are enhanced by visual attribution maps of localized biomarkers and quantified metrics of model robustness to data perturbations.
PMID:38351326 | DOI:10.1038/s41598-024-54251-1
Accurate Liver Fibrosis Detection Through Hybrid MRMR-BiLSTM-CNN Architecture with Histogram Equalization and Optimization
J Imaging Inform Med. 2024 Feb 13. doi: 10.1007/s10278-024-00995-1. Online ahead of print.
ABSTRACT
The early detection and accurate diagnosis of liver fibrosis, a progressive and potentially serious liver condition, are crucial for effective medical intervention. Invasive methods like biopsies for diagnosis can be risky and expensive. This research presents a novel computer-aided diagnosis model for liver fibrosis using a hybrid approach of minimum redundancy maximum relevance (MRMR) feature selection, bidirectional long short-term memory (BiLSTM), and convolutional neural networks (CNN). The proposed model involves multiple stages, including image acquisition, preprocessing, feature representation, fibrous tissue identification, and classification. Notably, histogram equalization is employed to enhance image quality by addressing variations in brightness levels. Performance evaluation encompasses a range of metrics such as accuracy, precision, sensitivity, specificity, F1 score, and error rate. Comparative analyses with established methods like DCNN, ANN-FLI, LungNet22, and SDAE-GAN underscore the efficacy of the proposed model. The innovative integration of hybrid MRMR-BiLSTM-CNN architecture and the horse herd optimization algorithm significantly enhances accuracy and F1 score, even with small datasets. The model tackles the complexities of hyperparameter optimization through the IHO algorithm and reduces training time by leveraging MRMR feature selection. In practical application, the proposed hybrid MRMR-BiLSTM-CNN method demonstrates remarkable performance with a 97.8% accuracy rate in identifying liver fibrosis images. It exhibits high precision, sensitivity, specificity, and minimal error rate, showcasing its potential for accurate and non-invasive diagnosis.
PMID:38351226 | DOI:10.1007/s10278-024-00995-1
A new framework for assessment of park management in smart cities: a study based on social media data and deep learning
Sci Rep. 2024 Feb 13;14(1):3630. doi: 10.1038/s41598-024-53345-0.
ABSTRACT
Urban park management assessment is critical to park operation and service quality. Traditional assessment methods cannot comprehensively assess park use and environmental conditions. Besides, although social media and big data have shown significant advantages in understanding public behavior or preference and park features or values, there has been little relevant research on park management assessment. This study proposes a deep learning-based framework for assessing urban park intelligent management from macro to micro levels with comment data from social media. By taking seven parks in Wuhan City as the objects, this study quantitatively assesses their overall state and performance in facilities, safety, environment, activities, and services, and reveals their main problems in management. The results demonstrate the impacts of various factors, including park type, season, and specific events such as remodeling and refurbishment, on visitor satisfaction and the characteristics of individual parks and their management. Compared with traditional methods, this framework enables real-time intelligent assessment of park management, which can accurately reflect park use and visitor feedback, and improve park service quality and management efficiency. Overall, this study provides important reference for intelligent park management assessment based on big data and artificial intelligence, which can facilitate the future development of smart cities.
PMID:38351201 | DOI:10.1038/s41598-024-53345-0
Deep learning assisted XRF spectra classification
Sci Rep. 2024 Feb 14;14(1):3666. doi: 10.1038/s41598-024-53988-z.
ABSTRACT
EDXRF spectrometry is a well-established and often-used analytical technique in examining materials from which cultural heritage objects are made. The analytical results are traditionally subjected to additional multivariate analysis for archaeometry studies to reduce the initial data's dimensionality based on informative features. Nowadays, artificial intelligence (AI) techniques are used more for this purpose. Different soft computing techniques are used to improve speed and accuracy. Choosing the most suitable AI method can increase the sustainability of the analytical process and postprocessing activities. An autoencoder neural network has been designed and used as a dimension reduction tool of initial [Formula: see text] data collected in the raw EDXRF spectra, containing information about the selected points' elemental composition on the canvas paintings' surface. The autoencoder network design enables the best possible reconstruction of the original EDXRF spectrum and the most informative feature extraction, which has been used for dimension reduction. Such configuration allows for efficient classification algorithms and their performances. The autoencoder neural network approach is more sustainable, especially in processing time consumption and experts' manual work.
PMID:38351176 | DOI:10.1038/s41598-024-53988-z
Automatic enhancement preprocessing for segmentation of low quality cell images
Sci Rep. 2024 Feb 13;14(1):3619. doi: 10.1038/s41598-024-53411-7.
ABSTRACT
We present a novel automatic preprocessing and ensemble learning technique for the segmentation of low-quality cell images. Capturing cells subjected to intense light is challenging due to their vulnerability to light-induced cell death. Consequently, microscopic cell images tend to be of low quality and it causes low accuracy for semantic segmentation. This problem can not be satisfactorily solved by classical image preprocessing methods. Therefore, we propose a novel approach of automatic enhancement preprocessing (AEP), which translates an input image into images that are easy to recognize by deep learning. AEP is composed of two deep neural networks, and the penultimate feature maps of the first network are employed as filters to translate an input image with low quality into images that are easily classified by deep learning. Additionally, we propose an automatic weighted ensemble learning (AWEL), which combines the multiple segmentation results. Since the second network predicts segmentation results corresponding to each translated input image, multiple segmentation results can be aggregated by automatically determining suitable weights. Experiments on two types of cell image segmentation confirmed that AEP can translate low-quality cell images into images that are easy to segment and that segmentation accuracy improves using AWEL.
PMID:38351053 | DOI:10.1038/s41598-024-53411-7
Freezing of gait assessment with inertial measurement units and deep learning: effect of tasks, medication states, and stops
J Neuroeng Rehabil. 2024 Feb 13;21(1):24. doi: 10.1186/s12984-024-01320-1.
ABSTRACT
BACKGROUND: Freezing of gait (FOG) is an episodic and highly disabling symptom of Parkinson's Disease (PD). Traditionally, FOG assessment relies on time-consuming visual inspection of camera footage. Therefore, previous studies have proposed portable and automated solutions to annotate FOG. However, automated FOG assessment is challenging due to gait variability caused by medication effects and varying FOG-provoking tasks. Moreover, whether automated approaches can differentiate FOG from typical everyday movements, such as volitional stops, remains to be determined. To address these questions, we evaluated an automated FOG assessment model with deep learning (DL) based on inertial measurement units (IMUs). We assessed its performance trained on all standardized FOG-provoking tasks and medication states, as well as on specific tasks and medication states. Furthermore, we examined the effect of adding stopping periods on FOG detection performance.
METHODS: Twelve PD patients with self-reported FOG (mean age 69.33 ± 6.02 years) completed a FOG-provoking protocol, including timed-up-and-go and 360-degree turning-in-place tasks in On/Off dopaminergic medication states with/without volitional stopping. IMUs were attached to the pelvis and both sides of the tibia and talus. A temporal convolutional network (TCN) was used to detect FOG episodes. FOG severity was quantified by the percentage of time frozen (%TF) and the number of freezing episodes (#FOG). The agreement between the model-generated outcomes and the gold standard experts' video annotation was assessed by the intra-class correlation coefficient (ICC).
RESULTS: For FOG assessment in trials without stopping, the agreement of our model was strong (ICC (%TF) = 0.92 [0.68, 0.98]; ICC(#FOG) = 0.95 [0.72, 0.99]). Models trained on a specific FOG-provoking task could not generalize to unseen tasks, while models trained on a specific medication state could generalize to unseen states. For assessment in trials with stopping, the agreement of our model was moderately strong (ICC (%TF) = 0.95 [0.73, 0.99]; ICC (#FOG) = 0.79 [0.46, 0.94]), but only when stopping was included in the training data.
CONCLUSION: A TCN trained on IMU signals allows valid FOG assessment in trials with/without stops containing different medication states and FOG-provoking tasks. These results are encouraging and enable future work investigating automated FOG assessment during everyday life.
PMID:38350964 | DOI:10.1186/s12984-024-01320-1