Deep learning

Editorial: Human-centered robot vision and artificial perception

Tue, 2024-07-23 06:00

Front Robot AI. 2024 Jul 8;11:1406280. doi: 10.3389/frobt.2024.1406280. eCollection 2024.

NO ABSTRACT

PMID:39040427 | PMC:PMC11260560 | DOI:10.3389/frobt.2024.1406280

Categories: Literature Watch

Deep learning Assisted Biomarker Development in Patients with Chronic Hepatitis B

Mon, 2024-07-22 06:00

Clin Mol Hepatol. 2024 Jul 23. doi: 10.3350/cmh.2024.0563. Online ahead of print.

NO ABSTRACT

PMID:39038960 | DOI:10.3350/cmh.2024.0563

Categories: Literature Watch

DANTE-CAIPI Accelerated Contrast-Enhanced 3D T1: Deep learning-based image quality improvement for Vessel Wall MR

Mon, 2024-07-22 06:00

AJNR Am J Neuroradiol. 2024 Jul 22:ajnr.A8424. doi: 10.3174/ajnr.A8424. Online ahead of print.

ABSTRACT

BACKGROUND AND PURPOSE: Accelerated and blood-suppressed post-contrast 3D intracranial vessel wall MRI (IVW) enables high-resolution rapid scanning but is associated with low SNR. We hypothesized that a deep-learning (DL) denoising algorithm applied to accelerated, blood-suppressed post-contrast IVW can yield high-quality images with reduced artifacts and higher SNR in shorter scan times.

MATERIALS AND METHODS: Sixty-four consecutive patients underwent IVW, including conventional post-contrast 3D T1-sampling perfection with application-optimized contrasts by using different flip angle evolution (SPACE) and delay-alternating with nutation for tailored excitation (DANTE) blood-suppressed and CAIPIRINHIA-accelerated (CAIPI) 3D T1-weighted TSE post-contrast sequences (DANTE-CAIPI-SPACE). DANTE-CAIPI-SPACE acquisitions were then denoised using an unrolled deep convolutional network (DANTECAIPI-SPACE+DL). SPACE, DANTE-CAIPI-SPACE, and DANTE-CAIPI-SPACE+DL images were compared for overall image quality, SNR, severity of artifacts, arterial and venous suppression, and lesion assessment using 4-point or 5-point Likert scales. Quantitative evaluation of SNR and contrast-to-noise ratio (CNR) was performed.

RESULTS: DANTE-CAIPI-SPACE+DL showed significantly reduced arterial (1 [1-1.75] vs. 3 [3-4], p<0.001) and venous flow artifacts (1 [1-2] vs. 3 [3-4], p<0.001) compared to SPACE. There was no significant difference between DANTE-CAIPI-SPACE+DL and SPACE in terms of image quality, SNR, artifact ratings and lesion assessment. For SNR ratings, DANTE-CAIPI-SPACE+DL was significantly better compared to DANTE-CAIPI-SPACE (2 [1-2], vs. 3 [2-3], p<0.001). No statistically significant differences were found between DANTECAIPI-SPACE and DANTE-CAIPI-SPACE+DL for image quality, artifact, arterial blood and venous blood flow artifacts, and lesion assessment. Quantitative vessel wall SNR and CNR median values were significantly higher for DANTE-CAIPI-SPACE+DL (SNR: 9.71, CNR: 4.24) compared to DANTE-CAIPI-SPACE (SNR: 5.50, CNR: 2.64), (p<0.001 for each), but there was no significant difference between SPACE (SNR: 10.82, CNR: 5.21) and DANTE-CAIPI-SPACE+DL.

CONCLUSIONS: Deep-learning denoised post-contrast T1-weighted DANTE-CAIPI-SPACE accelerated and blood-suppressed IVW showed improved flow suppression with a shorter scan time and equivalent qualitative and quantitative SNR measures relative to conventional post-contrast IVW. It also improved SNR metrics relative to post-contrast DANTE-CAIPI-SPACE IVW. Implementing deep-learning denoised DANTE-CAIPI-SPACE IVW has the potential to shorten protocol time while maintaining or improving the image quality of IVW.

ABBREVIATIONS: DL=deep learning; IVW=Intracranial vessel wall MRI; SPACE=sampling perfection with application-optimized contrasts by using different flip angle evolution; DANTE=delay-alternating with nutation for tailored excitation; CAIPI=controlled aliasing in parallel imaging; CNR=contrast-to-noise ratio.

PMID:39038956 | DOI:10.3174/ajnr.A8424

Categories: Literature Watch

Introducing enzymatic cleavage features and transfer learning realizes accurate peptide half-life prediction across species and organs

Mon, 2024-07-22 06:00

Brief Bioinform. 2024 May 23;25(4):bbae350. doi: 10.1093/bib/bbae350.

ABSTRACT

Peptide drugs are becoming star drug agents with high efficiency and selectivity which open up new therapeutic avenues for various diseases. However, the sensitivity to hydrolase and the relatively short half-life have severely hindered their development. In this study, a new generation artificial intelligence-based system for accurate prediction of peptide half-life was proposed, which realized the half-life prediction of both natural and modified peptides and successfully bridged the evaluation possibility between two important species (human, mouse) and two organs (blood, intestine). To achieve this, enzymatic cleavage descriptors were integrated with traditional peptide descriptors to construct a better representation. Then, robust models with accurate performance were established by comparing traditional machine learning and transfer learning, systematically. Results indicated that enzymatic cleavage features could certainly enhance model performance. The deep learning model integrating transfer learning significantly improved predictive accuracy, achieving remarkable R2 values: 0.84 for natural peptides and 0.90 for modified peptides in human blood, 0.984 for natural peptides and 0.93 for modified peptides in mouse blood, and 0.94 for modified peptides in mouse intestine on the test set, respectively. These models not only successfully composed the above-mentioned system but also improved by approximately 15% in terms of correlation compared to related works. This study is expected to provide powerful solutions for peptide half-life evaluation and boost peptide drug development.

PMID:39038937 | DOI:10.1093/bib/bbae350

Categories: Literature Watch

A large-scale assessment of sequence database search tools for homology-based protein function prediction

Mon, 2024-07-22 06:00

Brief Bioinform. 2024 May 23;25(4):bbae349. doi: 10.1093/bib/bbae349.

ABSTRACT

Sequence database searches followed by homology-based function transfer form one of the oldest and most popular approaches for predicting protein functions, such as Gene Ontology (GO) terms. These searches are also a critical component in most state-of-the-art machine learning and deep learning-based protein function predictors. Although sequence search tools are the basis of homology-based protein function prediction, previous studies have scarcely explored how to select the optimal sequence search tools and configure their parameters to achieve the best function prediction. In this paper, we evaluate the effect of using different options from among popular search tools, as well as the impacts of search parameters, on protein function prediction. When predicting GO terms on a large benchmark dataset, we found that BLASTp and MMseqs2 consistently exceed the performance of other tools, including DIAMOND-one of the most popular tools for function prediction-under default search parameters. However, with the correct parameter settings, DIAMOND can perform comparably to BLASTp and MMseqs2 in function prediction. Additionally, we developed a new scoring function to derive GO prediction from homologous hits that consistently outperform previously proposed scoring functions. These findings enable the improvement of almost all protein function prediction algorithms with a few easily implementable changes in their sequence homolog-based component. This study emphasizes the critical role of search parameter settings in homology-based function transfer and should have an important contribution to the development of future protein function prediction algorithms.

PMID:39038936 | DOI:10.1093/bib/bbae349

Categories: Literature Watch

CELA-MFP: a contrast-enhanced and label-adaptive framework for multi-functional therapeutic peptides prediction

Mon, 2024-07-22 06:00

Brief Bioinform. 2024 May 23;25(4):bbae348. doi: 10.1093/bib/bbae348.

ABSTRACT

Functional peptides play crucial roles in various biological processes and hold significant potential in many fields such as drug discovery and biotechnology. Accurately predicting the functions of peptides is essential for understanding their diverse effects and designing peptide-based therapeutics. Here, we propose CELA-MFP, a deep learning framework that incorporates feature Contrastive Enhancement and Label Adaptation for predicting Multi-Functional therapeutic Peptides. CELA-MFP utilizes a protein language model (pLM) to extract features from peptide sequences, which are then fed into a Transformer decoder for function prediction, effectively modeling correlations between different functions. To enhance the representation of each peptide sequence, contrastive learning is employed during training. Experimental results demonstrate that CELA-MFP outperforms state-of-the-art methods on most evaluation metrics for two widely used datasets, MFBP and MFTP. The interpretability of CELA-MFP is demonstrated by visualizing attention patterns in pLM and Transformer decoder. Finally, a user-friendly online server for predicting multi-functional peptides is established as the implementation of the proposed CELA-MFP and can be freely accessed at http://dreamai.cmii.online/CELA-MFP.

PMID:39038935 | DOI:10.1093/bib/bbae348

Categories: Literature Watch

Discovering predisposing genes for hereditary breast cancer using deep learning

Mon, 2024-07-22 06:00

Brief Bioinform. 2024 May 23;25(4):bbae346. doi: 10.1093/bib/bbae346.

ABSTRACT

Breast cancer (BC) is the most common malignancy affecting Western women today. It is estimated that as many as 10% of BC cases can be attributed to germline variants. However, the genetic basis of the majority of familial BC cases has yet to be identified. Discovering predisposing genes contributing to familial BC is challenging due to their presumed rarity, low penetrance, and complex biological mechanisms. Here, we focused on an analysis of rare missense variants in a cohort of 12 families of Middle Eastern origins characterized by a high incidence of BC cases. We devised a novel, high-throughput, variant analysis pipeline adapted for family studies, which aims to analyze variants at the protein level by employing state-of-the-art machine learning models and three-dimensional protein structural analysis. Using our pipeline, we analyzed 1218 rare missense variants that are shared between affected family members and classified 80 genes as candidate pathogenic. Among these genes, we found significant functional enrichment in peroxisomal and mitochondrial biological pathways which segregated across seven families in the study and covered diverse ethnic groups. We present multiple evidence that peroxisomal and mitochondrial pathways play an important, yet underappreciated, role in both germline BC predisposition and BC survival.

PMID:39038933 | DOI:10.1093/bib/bbae346

Categories: Literature Watch

EnzyACT: A Novel Deep Learning Method to Predict the Impacts of Single and Multiple Mutations on Enzyme Activity

Mon, 2024-07-22 06:00

J Chem Inf Model. 2024 Jul 22. doi: 10.1021/acs.jcim.4c00920. Online ahead of print.

ABSTRACT

Enzyme engineering involves the customization of enzymes by introducing mutations to expand the application scope of natural enzymes. One limitation of that is the complex interaction between two key properties, activity and stability, where the enhancement of one often leads to the reduction of the other, also called the trade-off mechanism. Although dozens of methods that predict the change of protein stability upon mutations have been developed, the prediction of the effect on activity is still in its early stage. Therefore, developing a fast and accurate method to predict the impact of the mutations on enzyme activity is helpful for enzyme design and understanding of the trade-off mechanism. Here, we introduce a novel approach, EnzyACT, a deep learning method that fuses graph technique and protein embedding to predict activity changes upon single or multiple mutations. Our model combines graph-based techniques and language models to predict the activity changes. Moreover, EnzyACT is trained on a new curated data set including both single- and multiple-point mutations. When benchmarked on multiple independent data sets, it shows uniform performance on problems affected by mutations. This work also provides insights into the impact of distant mutations within activity design, which could also be useful for predicting catalytic residues and developing improved enzyme-engineering strategies.

PMID:39038814 | DOI:10.1021/acs.jcim.4c00920

Categories: Literature Watch

An XAI-Enhanced EfficientNetB0 Framework for Precision Brain Tumor Detection in MRI Imaging

Mon, 2024-07-22 06:00

J Neurosci Methods. 2024 Jul 20:110227. doi: 10.1016/j.jneumeth.2024.110227. Online ahead of print.

ABSTRACT

BACKGROUND: Accurately diagnosing brain tumors from MRI scans is crucial for effective treatment planning. While traditional methods heavily rely on radiologist expertise, the integration of AI, particularly Convolutional Neural Networks (CNNs), has shown promise in improving accuracy. However, the lack of transparency in AI decision-making processes presents a challenge for clinical adoption.

METHODS: Recent advancements in deep learning, particularly the utilization of CNNs, have facilitated the development of models for medical image analysis. In this study, we employed the EfficientNetB0 architecture and integrated explainable AI techniques to enhance both accuracy and interpretability. Grad-CAM visualization was utilized to highlight significant areas in MRI scans influencing classification decisions.

RESULTS: Our model achieved a classification accuracy of 98.72% across four categories of brain tumors (Glioma, Meningioma, No Tumor, Pituitary), with precision and recall exceeding 97% for all categories. The incorporation of explainable AI techniques was validated through visual inspection of Grad-CAM heatmaps, which aligned well with established diagnostic markers in MRI scans.

CONCLUSION: The AI-enhanced EfficientNetB0 framework with explainable AI techniques significantly improves brain tumor classification accuracy to 98.72%, offering clear visual insights into the decision-making process. This method enhances diagnostic reliability and trust, demonstrating substantial potential for clinical adoption in medical diagnostics.

PMID:39038716 | DOI:10.1016/j.jneumeth.2024.110227

Categories: Literature Watch

A personal view on the history of toxins: from ancient times to artificial intelligence

Mon, 2024-07-22 06:00

Toxicon. 2024 Jul 20:108034. doi: 10.1016/j.toxicon.2024.108034. Online ahead of print.

ABSTRACT

Bioactive substances found in plants, microorganisms and animals have fascinated mankind since time immemorial. This review will focus on the progress that has been made over the centuries and our growing insights. The developments relate to both the discovery and characterization of novel bioactive substances, as well as the ceaseless implementation of refined techniques, the use of high-end instruments and breakthroughs in artificial intelligence with deep learning-based computational methods. As these approaches possess great translational potential, with many applications in different fields, such as therapeutic, diagnostic and agrochemical use, there is a good rationale to continue investing in toxinology-related research.

PMID:39038662 | DOI:10.1016/j.toxicon.2024.108034

Categories: Literature Watch

Differentiation of granulomatous nodules with lobulation and spiculation signs from solid lung adenocarcinomas using a CT deep learning model

Mon, 2024-07-22 06:00

BMC Cancer. 2024 Jul 22;24(1):875. doi: 10.1186/s12885-024-12611-0.

ABSTRACT

BACKGROUND: The diagnosis of solitary pulmonary nodules has always been a difficult and important point in clinical research, especially granulomatous nodules (GNs) with lobulation and spiculation signs, which are easily misdiagnosed as malignant tumors. Therefore, in this study, we utilised a CT deep learning (DL) model to distinguish GNs with lobulation and spiculation signs from solid lung adenocarcinomas (LADCs), to improve the diagnostic accuracy of preoperative diagnosis.

METHODS: 420 patients with pathologically confirmed GNs and LADCs from three medical institutions were retrospectively enrolled. The regions of interest in non-enhanced CT (NECT) and venous contrast-enhanced CT (VECT) were identified and labeled, and self-supervised labels were constructed. Cases from institution 1 were randomly divided into a training set (TS) and an internal validation set (IVS), and cases from institutions 2 and 3 were treated as an external validation set (EVS). Training and validation were performed using self-supervised transfer learning, and the results were compared with the radiologists' diagnoses.

RESULTS: The DL model achieved good performance in distinguishing GNs and LADCs, with area under curve (AUC) values of 0.917, 0.876, and 0.896 in the IVS and 0.889, 0.879, and 0.881 in the EVS for NECT, VECT, and non-enhanced with venous contrast-enhanced CT (NEVECT) images, respectively. The AUCs of radiologists 1, 2, 3, and 4 were, respectively, 0.739, 0.783, 0.883, and 0.901 in the (IVS) and 0.760, 0.760, 0.841, and 0.844 in the EVS.

CONCLUSIONS: A CT DL model showed great value for preoperative differentiation of GNs with lobulation and spiculation signs from solid LADCs, and its predictive performance was higher than that of radiologists.

PMID:39039511 | DOI:10.1186/s12885-024-12611-0

Categories: Literature Watch

DEL-Thyroid: deep ensemble learning framework for detection of thyroid cancer progression through genomic mutation

Mon, 2024-07-22 06:00

BMC Med Inform Decis Mak. 2024 Jul 22;24(1):198. doi: 10.1186/s12911-024-02604-1.

ABSTRACT

Genes, expressed as sequences of nucleotides, are susceptible to mutations, some of which can lead to cancer. Machine learning and deep learning methods have emerged as vital tools in identifying mutations associated with cancer. Thyroid cancer ranks as the 5th most prevalent cancer in the USA, with thousands diagnosed annually. This paper presents an ensemble learning model leveraging deep learning techniques such as Long Short-Term Memory (LSTM), Gated Recurrent Units (GRUs), and Bi-directional LSTM (Bi-LSTM) to detect thyroid cancer mutations early. The model is trained on a dataset sourced from asia.ensembl.org and IntOGen.org, consisting of 633 samples with 969 mutations across 41 genes, collected from individuals of various demographics. Feature extraction encompasses techniques including Hahn moments, central moments, raw moments, and various matrix-based methods. Evaluation employs three testing methods: self-consistency test (SCT), independent set test (IST), and 10-fold cross-validation test (10-FCVT). The proposed ensemble learning model demonstrates promising performance, achieving 96% accuracy in the independent set test (IST). Statistical measures such as training accuracy, testing accuracy, recall, sensitivity, specificity, Mathew's Correlation Coefficient (MCC), loss, training accuracy, F1 Score, and Cohen's kappa are utilized for comprehensive evaluation.

PMID:39039464 | DOI:10.1186/s12911-024-02604-1

Categories: Literature Watch

Radiograph-based rheumatoid arthritis diagnosis via convolutional neural network

Mon, 2024-07-22 06:00

BMC Med Imaging. 2024 Jul 22;24(1):180. doi: 10.1186/s12880-024-01362-w.

ABSTRACT

OBJECTIVES: Rheumatoid arthritis (RA) is a severe and common autoimmune disease. Conventional diagnostic methods are often subjective, error-prone, and repetitive works. There is an urgent need for a method to detect RA accurately. Therefore, this study aims to develop an automatic diagnostic system based on deep learning for recognizing and staging RA from radiographs to assist physicians in diagnosing RA quickly and accurately.

METHODS: We develop a CNN-based fully automated RA diagnostic model, exploring five popular CNN architectures on two clinical applications. The model is trained on a radiograph dataset containing 240 hand radiographs, of which 39 are normal and 201 are RA with five stages. For evaluation, we use 104 hand radiographs, of which 13 are normal and 91 RA with five stages.

RESULTS: The CNN model achieves good performance in RA diagnosis based on hand radiographs. For the RA recognition, all models achieve an AUC above 90% with a sensitivity over 98%. In particular, the AUC of the GoogLeNet-based model is 97.80%, and the sensitivity is 100.0%. For the RA staging, all models achieve over 77% AUC with a sensitivity over 80%. Specifically, the VGG16-based model achieves 83.36% AUC with 92.67% sensitivity.

CONCLUSION: The presented GoogLeNet-based model and VGG16-based model have the best AUC and sensitivity for RA recognition and staging, respectively. The experimental results demonstrate the feasibility and applicability of CNN in radiograph-based RA diagnosis. Therefore, this model has important clinical significance, especially for resource-limited areas and inexperienced physicians.

PMID:39039460 | DOI:10.1186/s12880-024-01362-w

Categories: Literature Watch

Deep learning classification of pediatric spinal radiographs for use in large scale imaging registries

Mon, 2024-07-22 06:00

Spine Deform. 2024 Jul 22. doi: 10.1007/s43390-024-00933-9. Online ahead of print.

ABSTRACT

PURPOSE: The purpose of this study is to develop and apply an algorithm that automatically classifies spine radiographs of pediatric scoliosis patients.

METHODS: Anterior-posterior (AP) and lateral spine radiographs were extracted from the institutional picture archive for patients with scoliosis. Overall, there were 7777 AP images and 5621 lateral images. Radiographs were manually classified into ten categories: two preoperative and three postoperative categories each for AP and lateral images. The images were split into training, validation, and testing sets (70:15:15 proportional split). A deep learning classifier using the EfficientNet B6 architecture was trained on the spine training set. Hyperparameters and model architecture were tuned against the performance of the models in the validation set.

RESULTS: The trained classifiers had an overall accuracy on the test set of 1.00 on 1166 AP images and 1.00 on 843 lateral images. Precision ranged from 0.98 to 1.00 in the AP images, and from 0.91 to 1.00 on the lateral images. Lower performance was observed on classes with fewer than 100 images in the dataset. Final performance metrics were calculated on the assigned test set, including accuracy, precision, recall, and F1 score (the harmonic mean of precision and recall).

CONCLUSIONS: A deep learning convolutional neural network classifier was trained to a high degree of accuracy to distinguish between 10 categories pre- and postoperative spine radiographs of patients with scoliosis. Observed performance was higher in more prevalent categories. These models represent an important step in developing an automatic system for data ingestion into large, labeled imaging registries.

PMID:39039392 | DOI:10.1007/s43390-024-00933-9

Categories: Literature Watch

Contextual AI models for single-cell protein biology

Mon, 2024-07-22 06:00

Nat Methods. 2024 Jul 22. doi: 10.1038/s41592-024-02341-3. Online ahead of print.

ABSTRACT

Understanding protein function and developing molecular therapies require deciphering the cell types in which proteins act as well as the interactions between proteins. However, modeling protein interactions across biological contexts remains challenging for existing algorithms. Here we introduce PINNACLE, a geometric deep learning approach that generates context-aware protein representations. Leveraging a multiorgan single-cell atlas, PINNACLE learns on contextualized protein interaction networks to produce 394,760 protein representations from 156 cell type contexts across 24 tissues. PINNACLE's embedding space reflects cellular and tissue organization, enabling zero-shot retrieval of the tissue hierarchy. Pretrained protein representations can be adapted for downstream tasks: enhancing 3D structure-based representations for resolving immuno-oncological protein interactions, and investigating drugs' effects across cell types. PINNACLE outperforms state-of-the-art models in nominating therapeutic targets for rheumatoid arthritis and inflammatory bowel diseases and pinpoints cell type contexts with higher predictive capability than context-free models. PINNACLE's ability to adjust its outputs on the basis of the context in which it operates paves the way for large-scale context-specific predictions in biology.

PMID:39039335 | DOI:10.1038/s41592-024-02341-3

Categories: Literature Watch

YOLO-Granada: a lightweight attentioned Yolo for pomegranates fruit detection

Mon, 2024-07-22 06:00

Sci Rep. 2024 Jul 22;14(1):16848. doi: 10.1038/s41598-024-67526-4.

ABSTRACT

Pomegranate is an important fruit crop that is usually managed manually through experience. Intelligent management systems for pomegranate orchards can improve yields and address labor shortages. Fast and accurate detection of pomegranates is one of the key technologies of this management system, crucial for yield and scientific management. Currently, most solutions use deep learning to achieve pomegranate detection, but deep learning is not effective in detecting small targets and large parameters, and the computation speed is slow; therefore, there is room for improving the pomegranate detection task. Based on the improved You Only Look Once version 5 (YOLOv5) algorithm, a lightweight pomegranate growth period detection algorithm YOLO-Granada is proposed. A lightweight ShuffleNetv2 network is used as the backbone to extract pomegranate features. Using grouped convolution reduces the computational effort of ordinary convolution, and using channel shuffle increases the interaction between different channels. In addition, the attention mechanism can help the neural network suppress less significant features in the channels or space, and the Convolutional Block Attention Module attention mechanism can improve the effect of attention and optimize the object detection accuracy by using the contribution factor of weights. The average accuracy of the improved network reaches 0.922. It is only less than 1% lower than the original YOLOv5s model (0.929) but brings a speed increase and a compression of the model size. and the detection speed is 17.3% faster than the original network. The parameters, floating-point operations, and model size of this network are compressed to 54.7%, 51.3%, and 56.3% of the original network, respectively. In addition, the algorithm detects 8.66 images per second, achieving real-time results. In this study, the Nihui convolutional neural network framework was further utilized to develop an Android-based application for real-time pomegranate detection. The method provides a more accurate and lightweight solution for intelligent management devices in pomegranate orchards, which can provide a reference for the design of neural networks in agricultural applications.

PMID:39039263 | DOI:10.1038/s41598-024-67526-4

Categories: Literature Watch

Neural general circulation models for weather and climate

Mon, 2024-07-22 06:00

Nature. 2024 Jul 22. doi: 10.1038/s41586-024-07744-y. Online ahead of print.

ABSTRACT

General circulation models (GCMs) are the foundation of weather and climate prediction1,2. GCMs are physics-based simulators that combine a numerical solver for large-scale dynamics with tuned representations for small-scale processes such as cloud formation. Recently, machine-learning models trained on reanalysis data have achieved comparable or better skill than GCMs for deterministic weather forecasting3,4. However, these models have not demonstrated improved ensemble forecasts, or shown sufficient stability for long-term weather and climate simulations. Here we present a GCM that combines a differentiable solver for atmospheric dynamics with machine-learning components and show that it can generate forecasts of deterministic weather, ensemble weather and climate on par with the best machine-learning and physics-based methods. NeuralGCM is competitive with machine-learning models for one- to ten-day forecasts, and with the European Centre for Medium-Range Weather Forecasts ensemble prediction for one- to fifteen-day forecasts. With prescribed sea surface temperature, NeuralGCM can accurately track climate metrics for multiple decades, and climate forecasts with 140-kilometre resolution show emergent phenomena such as realistic frequency and trajectories of tropical cyclones. For both weather and climate, our approach offers orders of magnitude computational savings over conventional GCMs, although our model does not extrapolate to substantially different future climates. Our results show that end-to-end deep learning is compatible with tasks performed by conventional GCMs and can enhance the large-scale physical simulations that are essential for understanding and predicting the Earth system.

PMID:39039241 | DOI:10.1038/s41586-024-07744-y

Categories: Literature Watch

Mediodorsal thalamus and ventral pallidum contribute to subcortical regulation of the default mode network

Mon, 2024-07-22 06:00

Commun Biol. 2024 Jul 23;7(1):891. doi: 10.1038/s42003-024-06531-9.

ABSTRACT

Humans and other animals readily transition from externally to internally focused attention, and these transitions are accompanied by activation of the default mode network (DMN). The DMN was considered a cortical network, yet recent evidence suggests subcortical structures are also involved. We investigated the role of ventral pallidum (VP) and mediodorsal thalamus (MD) in DMN regulation in tree shrew, a close relative of primates. Electrophysiology and deep learning-based classification of behavioral states revealed gamma oscillations in VP and MD coordinated with gamma in anterior cingulate (AC) cortex during DMN states. Cross-frequency coupling between gamma and delta oscillations was higher during DMN than other behaviors, underscoring the engagement of MD, VP and AC. Our findings highlight the importance of VP and MD in DMN regulation, extend homologies in DMN regulation among mammals, and underline the importance of thalamus and basal forebrain to the regulation of DMN.

PMID:39039239 | DOI:10.1038/s42003-024-06531-9

Categories: Literature Watch

Enhancing handwritten text recognition accuracy with gated mechanisms

Mon, 2024-07-22 06:00

Sci Rep. 2024 Jul 22;14(1):16800. doi: 10.1038/s41598-024-67738-8.

ABSTRACT

Handwritten Text Recognition (HTR) is a challenging task due to the complex structures and variations present in handwritten text. In recent years, the application of gated mechanisms, such as Long Short-Term Memory (LSTM) networks, has brought significant advancements to HTR systems. This paper presents an overview of HTR using a gated mechanism and highlights its novelty and advantages. The gated mechanism enables the model to capture long-term dependencies, retain relevant context, handle variable length sequences, mitigate error propagation, and adapt to contextual variations. The pipeline involves preprocessing the handwritten text images, extracting features, modeling the sequential dependencies using the gated mechanism, and decoding the output into readable text. The training process utilizes annotated datasets and optimization techniques to minimize transcription discrepancies. HTR using a gated mechanism has found applications in digitizing historical documents, automatic form processing, and real-time transcription. The results show improved accuracy and robustness compared to traditional HTR approaches. The advancements in HTR using a gated mechanism open up new possibilities for effectively recognizing and transcribing handwritten text in various domains. This research does a better job than the most recent iteration of the HTR system when compared to five different handwritten datasets (Washington, Saint Gall, RIMES, Bentham and IAM). Smartphones and robots are examples of low-cost computing devices that can benefit from this research.

PMID:39039237 | DOI:10.1038/s41598-024-67738-8

Categories: Literature Watch

The artistic image processing for visual healing in smart city

Mon, 2024-07-22 06:00

Sci Rep. 2024 Jul 22;14(1):16846. doi: 10.1038/s41598-024-68082-7.

ABSTRACT

This study investigates the processing methods of artistic images within the context of Smart city (SC) initiatives, focusing on the visual healing effects of artistic image processing to enhance urban residents' mental health and quality of life. Firstly, it examines the role of artistic image processing techniques in visual healing. Secondly, deep learning technology is introduced and improved, proposing the overlapping segmentation vision transformer (OSViT) for image blocks, and further integrating the bidirectional long short-term memory (BiLSTM) algorithm. An innovative artistic image processing and classification recognition model based on OSViT-BiLSTM is then constructed. Finally, the visual healing effect of the processed art images in different scenes is analyzed. The results demonstrate that the proposed model achieves a classification recognition accuracy of 92.9% for art images, which is at least 6.9% higher than that of other existing model algorithms. Additionally, over 90% of users report satisfaction with the visual healing effects of the artistic images. Therefore, it is found that the proposed model can accurately identify artistic images, enhance their beauty and artistry, and improve the visual healing effect. This study provides an experimental reference for incorporating visual healing into SC initiatives.

PMID:39039163 | DOI:10.1038/s41598-024-68082-7

Categories: Literature Watch

Pages