Deep learning

Automated assessment of cardiac dynamics in aging and dilated cardiomyopathy Drosophila models using machine learning

Fri, 2024-06-07 06:00

Commun Biol. 2024 Jun 7;7(1):702. doi: 10.1038/s42003-024-06371-7.

ABSTRACT

The Drosophila model is pivotal in deciphering the pathophysiological underpinnings of various human ailments, notably aging and cardiovascular diseases. Cutting-edge imaging techniques and physiology yield vast high-resolution videos, demanding advanced analysis methods. Our platform leverages deep learning to segment optical microscopy images of Drosophila hearts, enabling the quantification of cardiac parameters in aging and dilated cardiomyopathy (DCM). Validation using experimental datasets confirms the efficacy of our aging model. We employ two innovative approaches deep-learning video classification and machine-learning based on cardiac parameters to predict fly aging, achieving accuracies of 83.3% (AUC 0.90) and 79.1%, (AUC 0.87) respectively. Moreover, we extend our deep-learning methodology to assess cardiac dysfunction associated with the knock-down of oxoglutarate dehydrogenase (OGDH), revealing its potential in studying DCM. This versatile approach promises accelerated cardiac assays for modeling various human diseases in Drosophila and holds promise for application in animal and human cardiac physiology under diverse conditions.

PMID:38849449 | DOI:10.1038/s42003-024-06371-7

Categories: Literature Watch

Estimation of the amount of pear pollen based on flowering stage detection using deep learning

Fri, 2024-06-07 06:00

Sci Rep. 2024 Jun 7;14(1):13163. doi: 10.1038/s41598-024-63611-w.

ABSTRACT

Pear pollination is performed by artificial pollination because the pollination rate through insect pollination is not stable. Pollen must be collected to secure sufficient pollen for artificial pollination. However, recently, collecting sufficient amounts of pollen in Japan has become difficult, resulting in increased imports from overseas. To solve this problem, improving the efficiency of pollen collection and strengthening the domestic supply and demand system is necessary. In this study, we proposed an Artificial Intelligence (AI)-based method to estimate the amount of pear pollen. The proposed method used a deep learning-based object detection algorithm, You Only Look Once (YOLO), to classify and detect flower shapes in five stages, from bud to flowering, and to estimate the pollen amount. In this study, the performance of the proposed method was discussed by analyzing the accuracy and error of classification for multiple flower varieties. Although this study only discussed the performance of estimating the amount of pollen collected, in the future, we aim to establish a technique for estimating the time of maximum pollen collection using the method proposed in this study.

PMID:38849427 | DOI:10.1038/s41598-024-63611-w

Categories: Literature Watch

An enhanced speech emotion recognition using vision transformer

Fri, 2024-06-07 06:00

Sci Rep. 2024 Jun 7;14(1):13126. doi: 10.1038/s41598-024-63776-4.

ABSTRACT

In human-computer interaction systems, speech emotion recognition (SER) plays a crucial role because it enables computers to understand and react to users' emotions. In the past, SER has significantly emphasised acoustic properties extracted from speech signals. The use of visual signals for enhancing SER performance, however, has been made possible by recent developments in deep learning and computer vision. This work utilizes a lightweight Vision Transformer (ViT) model to propose a novel method for improving speech emotion recognition. We leverage the ViT model's capabilities to capture spatial dependencies and high-level features in images which are adequate indicators of emotional states from mel spectrogram input fed into the model. To determine the efficiency of our proposed approach, we conduct a comprehensive experiment on two benchmark speech emotion datasets, the Toronto English Speech Set (TESS) and the Berlin Emotional Database (EMODB). The results of our extensive experiment demonstrate a considerable improvement in speech emotion recognition accuracy attesting to its generalizability as it achieved 98%, 91%, and 93% (TESS-EMODB) accuracy respectively on the datasets. The outcomes of the comparative experiment show that the non-overlapping patch-based feature extraction method substantially improves the discipline of speech emotion recognition. Our research indicates the potential for integrating vision transformer models into SER systems, opening up fresh opportunities for real-world applications requiring accurate emotion recognition from speech compared with other state-of-the-art techniques.

PMID:38849422 | DOI:10.1038/s41598-024-63776-4

Categories: Literature Watch

xECGArch: a trustworthy deep learning architecture for interpretable ECG analysis considering short-term and long-term features

Fri, 2024-06-07 06:00

Sci Rep. 2024 Jun 7;14(1):13122. doi: 10.1038/s41598-024-63656-x.

ABSTRACT

Deep learning-based methods have demonstrated high classification performance in the detection of cardiovascular diseases from electrocardiograms (ECGs). However, their blackbox character and the associated lack of interpretability limit their clinical applicability. To overcome existing limitations, we present a novel deep learning architecture for interpretable ECG analysis (xECGArch). For the first time, short- and long-term features are analyzed by two independent convolutional neural networks (CNNs) and combined into an ensemble, which is extended by methods of explainable artificial intelligence (xAI) to whiten the blackbox. To demonstrate the trustworthiness of xECGArch, perturbation analysis was used to compare 13 different xAI methods. We parameterized xECGArch for atrial fibrillation (AF) detection using four public ECG databases ( n = 9854 ECGs) and achieved an F1 score of 95.43% in AF versus non-AF classification on an unseen ECG test dataset. A systematic comparison of xAI methods showed that deep Taylor decomposition provided the most trustworthy explanations ( + 24 % compared to the second-best approach). xECGArch can account for short- and long-term features corresponding to clinical features of morphology and rhythm, respectively. Further research will focus on the relationship between xECGArch features and clinical features, which may help in medical applications for diagnosis and therapy.

PMID:38849417 | DOI:10.1038/s41598-024-63656-x

Categories: Literature Watch

Deep learning for colorectal cancer detection in contrast-enhanced CT without bowel preparation: a retrospective, multicentre study

Fri, 2024-06-07 06:00

EBioMedicine. 2024 Jun 6;104:105183. doi: 10.1016/j.ebiom.2024.105183. Online ahead of print.

ABSTRACT

BACKGROUND: Contrast-enhanced CT scans provide a means to detect unsuspected colorectal cancer. However, colorectal cancers in contrast-enhanced CT without bowel preparation may elude detection by radiologists. We aimed to develop a deep learning (DL) model for accurate detection of colorectal cancer, and evaluate whether it could improve the detection performance of radiologists.

METHODS: We developed a DL model using a manually annotated dataset (1196 cancer vs 1034 normal). The DL model was tested using an internal test set (98 vs 115), two external test sets (202 vs 265 in 1, and 252 vs 481 in 2), and a real-world test set (53 vs 1524). We compared the detection performance of the DL model with radiologists, and evaluated its capacity to enhance radiologists' detection performance.

FINDINGS: In the four test sets, the DL model had the area under the receiver operating characteristic curves (AUCs) ranging between 0.957 and 0.994. In both the internal test set and external test set 1, the DL model yielded higher accuracy than that of radiologists (97.2% vs 86.0%, p < 0.0001; 94.9% vs 85.3%, p < 0.0001), and significantly improved the accuracy of radiologists (93.4% vs 86.0%, p < 0.0001; 93.6% vs 85.3%, p < 0.0001). In the real-world test set, the DL model delivered sensitivity comparable to that of radiologists who had been informed about clinical indications for most cancer cases (94.3% vs 96.2%, p > 0.99), and it detected 2 cases that had been missed by radiologists.

INTERPRETATION: The developed DL model can accurately detect colorectal cancer and improve radiologists' detection performance, showing its potential as an effective computer-aided detection tool.

FUNDING: This study was supported by National Science Fund for Distinguished Young Scholars of China (No. 81925023); Regional Innovation and Development Joint Fund of National Natural Science Foundation of China (No. U22A20345); National Natural Science Foundation of China (No. 82072090 and No. 82371954); Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application (No. 2022B1212010011); High-level Hospital Construction Project (No. DFJHBF202105).

PMID:38848616 | DOI:10.1016/j.ebiom.2024.105183

Categories: Literature Watch

Transcription Factor Binding Site Prediction Using CnNet Approach

Fri, 2024-06-07 06:00

IEEE/ACM Trans Comput Biol Bioinform. 2024 Jun 7;PP. doi: 10.1109/TCBB.2024.3411024. Online ahead of print.

ABSTRACT

Controlling the gene expression is the most important development in a living organism, which makes it easier to find different kinds of diseases and their causes. It's very difficult to know what factors control the gene expression. Transcription Factor (TF) is a protein that plays an important role in gene expression. Discovering the transcription factor has immense biological significance, however, it is challenging to develop novel techniques and evaluation for regulatory developments in biological structures. In this research, we mainly focus on 'sequence specificities' that can be ascertained from experimental data with 'deep learning' techniques, which offer a scalable, flexible and unified computational approach for predicting transcription factor binding. Specifically, Multiple Expression motifs for Motif Elicitation (MEME) technique with Convolution Neural Network (CNN) named as CnNet, has been used for discovering the 'sequence specificities' of DNA gene sequences dataset. This process involves two steps: a) discovering the motifs that are capable of identifying useful TF binding site by using MEME technique, and b) computing a score indicating the likelihood of a given sequence being a useful binding site by using CNN technique. The proposed CnNet approach predicts the TF binding score with much better accuracy compared to existing approaches. The source code and datasets used in this work are available at https://github.com/masoodbai/CnNet-Approach-for-TFBS.git.

PMID:38848239 | DOI:10.1109/TCBB.2024.3411024

Categories: Literature Watch

Neural Disparity Refinement

Fri, 2024-06-07 06:00

IEEE Trans Pattern Anal Mach Intell. 2024 Jun 7;PP. doi: 10.1109/TPAMI.2024.3411292. Online ahead of print.

ABSTRACT

We propose a framework that combines traditional, hand-crafted algorithms and recent advances in deep learning to obtain high-quality, high-resolution disparity maps from stereo images. By casting the refinement process as a continuous feature sampling strategy, our neural disparity refinement network can estimate an enhanced disparity map at any output resolution. Our solution can process any disparity map produced by classical stereo algorithms, as well as those predicted by modern stereo networks or even different depth-from-images approaches, such as the COLMAP structure-from-motion pipeline. Nonetheless, when deployed in the former configuration, our framework performs at its best in terms of zero-shot generalization from synthetic to real images. Moreover, its continuous formulation allows for easily handling the unbalanced stereo setup very diffused in mobile phones.

PMID:38848234 | DOI:10.1109/TPAMI.2024.3411292

Categories: Literature Watch

SeqAFNet: A Beat-Wise Sequential Neural Network for Atrial Fibrillation Classification in Adhesive Patch-Type Electrocardiographs

Fri, 2024-06-07 06:00

IEEE J Biomed Health Inform. 2024 Jun 7;PP. doi: 10.1109/JBHI.2024.3411056. Online ahead of print.

ABSTRACT

Due to their convenience, adhesive patch-type electrocardiographs are commonly used for arrhythmia screening. This study aimed to develop a reliable method that can improve the classification performance of atrial fibrillation (AF) using these devices based on the 2020 European Society of Cardiology (ESC) guidelines for AF diagnosis in clinical practice. We developed a deep learning model that utilizes RR interval frames for precise, beat-wise classification of electrocardiogram (ECG) signals. This model is specifically designed to sequentially classify each R peak on the ECG, considering the rhythms surrounding each beat. It features a two-stage bidirectional Recurrent Neural Network (RNN) with a many-to-many architecture, which is particularly optimized for processing sequential and time-series data. The structure aims to extract local features and capture long-term dependencies associated with AF. After inference, outputs which indicating either AF or non-AF, derived from various temporal sequences are combined through an ensembling technique to enhance prediction accuracy. We collected AF data from a clinical trial that utilized the MEMO Patch, an adhesive patch-type electrocardiograph. When trained on public databases, the model demonstrated high accuracy on the patch dataset (accuracy: 0.986, precision: 0.981, sensitivity: 0.979, specificity: 0.992, and F1 score: 0.98), maintaining consistent performance across public datasets. SeqAFNet was robust for AF classification, making it a potential tool in real-world applications.

PMID:38848232 | DOI:10.1109/JBHI.2024.3411056

Categories: Literature Watch

A Novel Hierarchical Cross-Stream Aggregation Neural Network for Semantic Segmentation of 3-D Dental Surface Models

Fri, 2024-06-07 06:00

IEEE Trans Neural Netw Learn Syst. 2024 Jun 7;PP. doi: 10.1109/TNNLS.2024.3404276. Online ahead of print.

ABSTRACT

Accurate teeth delineation on 3-D dental models is essential for individualized orthodontic treatment planning. Pioneering works like PointNet suggest a promising direction to conduct efficient and accurate 3-D dental model analyses in end-to-end learnable fashions. Recent studies further imply that multistream architectures to concurrently learn geometric representations from different inputs/views (e.g., coordinates and normals) are beneficial for segmenting teeth with varying conditions. However, such multistream networks typically adopt simple late-fusion strategies to combine features captured from raw inputs that encode complementary but fundamentally different geometric information, potentially hampering their accuracy in end-to-end semantic segmentation. This article presents a hierarchical cross-stream aggregation (HiCA) network to learn more discriminative point/cell-wise representations from multiview inputs for fine-grained 3-D semantic segmentation. Specifically, based upon our multistream backbone with input-tailored feature extractors, we first design a contextual cross-steam aggregation (CA) module conditioned on interstream consistency to boost each view's contextual representation learning jointly. Then, before the late fusion of different streams' outputs for segmentation, we further deploy a discriminative cross-stream aggregation (DA) module to concurrently update all views' discriminative representation learning by leveraging a specific graph attention strategy induced by multiview prototype learning. On both public and in-house datasets of real-patient dental models, our method significantly outperformed state-of-the-art (SOTA) deep learning methods for teeth semantic segmentation. In addition, extended experimental results suggest the applicability of HiCA to other general 3-D shape segmentation tasks. The code is available at https://github.com/ladderlab-xjtu/HiCA.

PMID:38848227 | DOI:10.1109/TNNLS.2024.3404276

Categories: Literature Watch

Multi-Modal Sleep Stage Classification With Two-Stream Encoder-Decoder

Fri, 2024-06-07 06:00

IEEE Trans Neural Syst Rehabil Eng. 2024;32:2096-2105. doi: 10.1109/TNSRE.2024.3394738.

ABSTRACT

Sleep staging serves as a fundamental assessment for sleep quality measurement and sleep disorder diagnosis. Although current deep learning approaches have successfully integrated multimodal sleep signals, enhancing the accuracy of automatic sleep staging, certain challenges remain, as follows: 1) optimizing the utilization of multi-modal information complementarity, 2) effectively extracting both long- and short-range temporal features of sleep information, and 3) addressing the class imbalance problem in sleep data. To address these challenges, this paper proposes a two-stream encode-decoder network, named TSEDSleepNet, which is inspired by the depth sensitive attention and automatic multi-modal fusion (DSA2F) framework. In TSEDSleepNet, a two-stream encoder is used to extract the multiscale features of electrooculogram (EOG) and electroencephalogram (EEG) signals. And a self-attention mechanism is utilized to fuse the multiscale features, generating multi-modal saliency features. Subsequently, the coarser-scale construction module (CSCM) is adopted to extract and construct multi-resolution features from the multiscale features and the salient features. Thereafter, a Transformer module is applied to capture both long- and short-range temporal features from the multi-resolution features. Finally, the long- and short-range temporal features are restored with low-layer details and mapped to the predicted classification results. Additionally, the Lovász loss function is applied to alleviate the class imbalance problem in sleep datasets. Our proposed method was tested on the Sleep-EDF-39 and Sleep-EDF-153 datasets, and it achieved classification accuracies of 88.9% and 85.2% and Macro-F1 scores of 84.8% and 79.7%, respectively, thus outperforming conventional traditional baseline models. These results highlight the efficacy of the proposed method in fusing multi-modal information. This method has potential for application as an adjunct tool for diagnosing sleep disorders.

PMID:38848223 | DOI:10.1109/TNSRE.2024.3394738

Categories: Literature Watch

Research on prediction and evaluation algorithm of sports athletes performance based on neural network

Fri, 2024-06-07 06:00

Technol Health Care. 2024 Apr 30. doi: 10.3233/THC-232000. Online ahead of print.

ABSTRACT

BACKGROUND: The Ultimate Fighting Championship (UFC) stands as a prominent global platform for professional mixed martial arts, captivating audiences worldwide. With its continuous growth and globalization efforts, UFC events have garnered significant attention and achieved commendable results. However, as the scale of development expands, the operational demands on UFC events intensify. At its core, UFC thrives on the exceptional performances of its athletes, which serve as the primary allure for audiences.

OBJECTIVE: This study aims to enhance the allure of UFC matches and cultivate exceptional athletes by predicting athlete performance on the field. To achieve this, a recurrent neural network prediction model based on Bidirectional Long Short-Term Memory (BiLSTM) is proposed. The model seeks to leverage athlete portraits and characteristics for performance prediction.

METHODS: The proposed methodology involves constructing athlete portraits and analyzing athlete characteristics to develop the prediction model. The BiLSTM-based recurrent neural network is utilized for its ability to capture temporal dependencies in sequential data. The model's performance is assessed through experimental analysis.

RESULTS: Experimental results demonstrate that the athlete performance prediction model achieved an overall accuracy of 0.7524. Comparative analysis reveals that the proposed BiLSTM model outperforms traditional methods such as Linear Regression and Multilayer Perceptron (MLP), showcasing superior prediction accuracy.

CONCLUSION: This study introduces a novel approach to predicting athlete performance in UFC matches using a BiLSTM-based recurrent neural network. By leveraging athlete portraits and characteristics, the proposed model offers improved accuracy compared to classical methods. Enhancing the predictive capabilities in UFC not only enriches the viewing experience but also contributes to the development of exceptional athletes in the sport.

PMID:38848203 | DOI:10.3233/THC-232000

Categories: Literature Watch

Machine Learning Approaches for Dementia Detection Through Speech and Gait Analysis: A Systematic Literature Review

Fri, 2024-06-07 06:00

J Alzheimers Dis. 2024 Jun 3. doi: 10.3233/JAD-231459. Online ahead of print.

ABSTRACT

BACKGROUND: Dementia is a general term for several progressive neurodegenerative disorders including Alzheimer's disease. Timely and accurate detection is crucial for early intervention. Advancements in artificial intelligence present significant potential for using machine learning to aid in early detection.

OBJECTIVE: Summarize the state-of-the-art machine learning-based approaches for dementia prediction, focusing on non-invasive methods, as the burden on the patients is lower. Specifically, the analysis of gait and speech performance can offer insights into cognitive health through clinically cost-effective screening methods.

METHODS: A systematic literature review was conducted following the PRISMA protocol (Preferred Reporting Items for Systematic Reviews and Meta-Analyses). The search was performed on three electronic databases (Scopus, Web of Science, and PubMed) to identify the relevant studies published between 2017 to 2022. A total of 40 papers were selected for review.

RESULTS: The most common machine learning methods employed were support vector machine followed by deep learning. Studies suggested the use of multimodal approaches as they can provide comprehensive and better prediction performance. Deep learning application in gait studies is still in the early stages as few studies have applied it. Moreover, including features of whole body movement contribute to better classification accuracy. Regarding speech studies, the combination of different parameters (acoustic, linguistic, cognitive testing) produced better results.

CONCLUSIONS: The review highlights the potential of machine learning, particularly non-invasive approaches, in the early prediction of dementia. The comparable prediction accuracies of manual and automatic speech analysis indicate an imminent fully automated approach for dementia detection.

PMID:38848181 | DOI:10.3233/JAD-231459

Categories: Literature Watch

LDCT image biomarkers that matter most for the deep learning classification of indeterminate pulmonary nodules

Fri, 2024-06-07 06:00

Cancer Biomark. 2024 May 22. doi: 10.3233/CBM-230444. Online ahead of print.

ABSTRACT

BACKGROUND: Continued improvement in deep learning methodologies has increased the rate at which deep neural networks are being evaluated for medical applications, including diagnosis of lung cancer. However, there has been limited exploration of the underlying radiological characteristics that the network relies on to identify lung cancer in computed tomography (CT) images.

OBJECTIVE: In this study, we used a combination of image masking and saliency activation maps to systematically explore the contributions of both parenchymal and tumor regions in a CT image to the classification of indeterminate lung nodules.

METHODS: We selected individuals from the National Lung Screening Trial (NLST) with solid pulmonary nodules 4-20 mm in diameter. Segmentation masks were used to generate three distinct datasets; 1) an Original Dataset containing the complete low-dose CT scans from the NLST, 2) a Parenchyma-Only Dataset in which the tumor regions were covered by a mask, and 3) a Tumor-Only Dataset in which only the tumor regions were included.

RESULTS: The Original Dataset significantly outperformed the Parenchyma-Only Dataset and the Tumor-Only Dataset with an AUC of 80.80 ± 3.77% compared to 76.39 ± 3.16% and 78.11 ± 4.32%, respectively. Gradient-weighted class activation mapping (Grad-CAM) of the Original Dataset showed increased attention was being given to the nodule and the tumor-parenchyma boundary when nodules were classified as malignant. This pattern of attention remained unchanged in the case of the Parenchyma-Only Dataset. Nodule size and first-order statistical features of the nodules were significantly different with the average malignant and benign nodule maximum 3d diameter being 23 mm and 12 mm, respectively.

CONCLUSION: We conclude that network performance is linked to textural features of nodules such as kurtosis, entropy and intensity, as well as morphological features such as sphericity and diameter. Furthermore, textural features are more positively associated with malignancy than morphological features.

PMID:38848168 | DOI:10.3233/CBM-230444

Categories: Literature Watch

Eichner classification based on panoramic X-ray images using deep learning: A pilot study

Fri, 2024-06-07 06:00

Biomed Mater Eng. 2024 Jun 7. doi: 10.3233/BME-230217. Online ahead of print.

ABSTRACT

BACKGROUND: Research using panoramic X-ray images using deep learning has been progressing in recent years. There is a need to propose methods that can classify and predict from image information.

OBJECTIVE: In this study, Eichner classification was performed on image processing based on panoramic X-ray images. The Eichner classification was based on the remaining teeth, with the aim of making partial dentures. This classification was based on the condition that the occlusal position was supported by the remaining teeth in the upper and lower jaws.

METHODS: Classification models were constructed using two convolutional neural network methods: the sequential and VGG19 models. The accuracy was compared with the accuracy of Eichner classification using the sequential and VGG19 models.

RESULTS: Both accuracies were greater than 81%, and they had sufficient functions for the Eichner classification.

CONCLUSION: We were able to build a highly accurate prediction model using deep learning scratch sequential model and VGG19. This predictive model will become part of the basic considerations for future AI research in dentistry.

PMID:38848165 | DOI:10.3233/BME-230217

Categories: Literature Watch

Unsupervised stain augmentation enhanced glomerular instance segmentation on pathology images

Fri, 2024-06-07 06:00

Int J Comput Assist Radiol Surg. 2024 Jun 7. doi: 10.1007/s11548-024-03154-7. Online ahead of print.

ABSTRACT

PURPOSE: In pathology images, different stains highlight different glomerular structures, so a supervised deep learning-based glomerular instance segmentation model trained on individual stains performs poorly on other stains. However, it is difficult to obtain a training set with multiple stains because the labeling of pathology images is very time-consuming and tedious. Therefore, in this paper, we proposed an unsupervised stain augmentation-based method for segmentation of glomerular instances.

METHODS: In this study, we successfully realized the conversion between different staining methods such as PAS, MT and PASM by contrastive unpaired translation (CUT), thus improving the staining diversity of the training set. Moreover, we replaced the backbone of mask R-CNN with swin transformer to further improve the efficiency of feature extraction and thus achieve better performance in instance segmentation task.

RESULTS: To validate the method presented in this paper, we constructed a dataset from 216 WSIs of the three stains in this study. After conducting in-depth experiments, we verified that the instance segmentation method based on stain augmentation outperforms existing methods across all metrics for PAS, PASM, and MT stains. Furthermore, ablation experiments are performed in this paper to further demonstrate the effectiveness of the proposed module.

CONCLUSION: This study successfully demonstrated the potential of unsupervised stain augmentation to improve glomerular segmentation in pathology analysis. Future research could extend this approach to other complex segmentation tasks in the pathology image domain to further explore the potential of applying stain augmentation techniques in different domains of pathology image analysis.

PMID:38848032 | DOI:10.1007/s11548-024-03154-7

Categories: Literature Watch

Layer-selective deep representation to improve esophageal cancer classification

Fri, 2024-06-07 06:00

Med Biol Eng Comput. 2024 Jun 7. doi: 10.1007/s11517-024-03142-8. Online ahead of print.

ABSTRACT

Even though artificial intelligence and machine learning have demonstrated remarkable performances in medical image computing, their accountability and transparency level must be improved to transfer this success into clinical practice. The reliability of machine learning decisions must be explained and interpreted, especially for supporting the medical diagnosis. For this task, the deep learning techniques' black-box nature must somehow be lightened up to clarify its promising results. Hence, we aim to investigate the impact of the ResNet-50 deep convolutional design for Barrett's esophagus and adenocarcinoma classification. For such a task, and aiming at proposing a two-step learning technique, the output of each convolutional layer that composes the ResNet-50 architecture was trained and classified for further definition of layers that would provide more impact in the architecture. We showed that local information and high-dimensional features are essential to improve the classification for our task. Besides, we observed a significant improvement when the most discriminative layers expressed more impact in the training and classification of ResNet-50 for Barrett's esophagus and adenocarcinoma classification, demonstrating that both human knowledge and computational processing may influence the correct learning of such a problem.

PMID:38848031 | DOI:10.1007/s11517-024-03142-8

Categories: Literature Watch

Large language models from OpenAI, Google, Meta, X and Co. : The role of "closed" and "open" models in radiology

Fri, 2024-06-07 06:00

Radiologie (Heidelb). 2024 Jun 7. doi: 10.1007/s00117-024-01327-8. Online ahead of print.

ABSTRACT

BACKGROUND: In 2023, the release of ChatGPT triggered an artificial intelligence (AI) boom. The underlying large language models (LLM) of the nonprofit organization "OpenAI" are not freely available under open-source licenses, which does not allow on-site implementation inside secure clinic networks. However, efforts are being made by open-source communities, start-ups and large tech companies to democratize the use of LLMs. This opens up the possibility of using LLMs in a data protection-compliant manner and even adapting them to our own data.

OBJECTIVES: This paper aims to explain the potential of privacy-compliant local LLMs for radiology and to provide insights into the "open" versus "closed" dynamics of the currently rapidly developing field of AI.

MATERIALS AND METHODS: PubMed search for radiology articles with LLMs and subjective selection of references in the sense of a narrative key topic article.

RESULTS: Various stakeholders, including large tech companies such as Meta, Google and X, but also European start-ups such as Mistral AI, contribute to the democratization of LLMs by publishing the models (open weights) or by publishing the model and source code (open source). Their performance is lower than current "closed" LLMs, such as GPT‑4 from OpenAI.

CONCLUSION: Despite differences in performance, open and thus locally implementable LLMs show great promise for improving the efficiency and quality of diagnostic reporting as well as interaction with patients and enable retrospective extraction of diagnostic information for secondary use of clinical free-text databases for research, teaching or clinical application.

PMID:38847898 | DOI:10.1007/s00117-024-01327-8

Categories: Literature Watch

Discovery of a Novel and Potent LCK Inhibitor for Leukemia Treatment via Deep Learning and Molecular Docking

Fri, 2024-06-07 06:00

J Chem Inf Model. 2024 Jun 7. doi: 10.1021/acs.jcim.4c00151. Online ahead of print.

ABSTRACT

The lymphocyte-specific protein tyrosine kinase (LCK) plays a crucial role in both T-cell development and activation. Dysregulation of LCK signaling has been demonstrated to drive the oncogenesis of T-cell acute lymphoblastic leukemia (T-ALL), thus providing a therapeutic target for leukemia treatment. In this study, we introduced a sophisticated virtual screening strategy combined with biological evaluations to discover potent LCK inhibitors. Our initial approach involved utilizing the PLANET algorithm to assess and contrast various scoring methodologies suitable for LCK inhibitor screening. After effectively evaluating PLANET, we progressed to devise a virtual screening workflow that synergistically combines the strengths of PLANET with the capabilities of Schrödinger's suite. This integrative strategy led to the efficient identification of four potential LCK inhibitors. Among them, compound 1232030-35-1 stood out as the most promising candidate with an IC50 of 0.43 nM. Further in vitro bioassays revealed that 1232030-35-1 exhibited robust antiproliferative effects on T-ALL cells, which was attributed to its ability to suppress the phosphorylations of key molecules in the LCK signaling pathway. More importantly, 1232030-35-1 treatment demonstrated profound in vivo antileukemia efficacy in a human T-ALL xenograft model. In addition, complementary molecular dynamics simulations provided deeper insight into the binding kinetics between 1232030-35-1 and LCK, highlighting the formation of a hydrogen bond with Met319. Collectively, our study established a robust and effective screening strategy that integrates AI-driven and conventional methodologies for the identification of LCK inhibitors, positioning 1232030-35-1 as a highly promising and novel drug-like candidate for potential applications in treating T-ALL.

PMID:38847742 | DOI:10.1021/acs.jcim.4c00151

Categories: Literature Watch

Spatial acoustic properties recovery with deep learning

Fri, 2024-06-07 06:00

J Acoust Soc Am. 2024 Jun 1;155(6):3690-3701. doi: 10.1121/10.0026231.

ABSTRACT

The physics-informed neural network (PINN) can recover partial differential equation (PDE) coefficients that remain constant throughout the spatial domain directly from measurements. We propose a spatially dependent physics-informed neural network (SD-PINN), which enables recovering coefficients in spatially dependent PDEs using one neural network, eliminating the requirement for domain-specific physical expertise. The network is trained by minimizing a combination of loss functions involving data-fitting and physical constraints, in which the requirement for satisfying the assumed governing PDE is encoded. For the recovery of spatially two-dimensional (2D) PDEs, we store the PDE coefficients at all locations in the 2D region of interest into a matrix and incorporate a low-rank assumption for this matrix to recover the coefficients at locations without measurements. We apply the SD-PINN to recovering spatially dependent coefficients of the wave equation to reveal the spatial distribution of acoustic properties in the inhomogeneous medium.

PMID:38847594 | DOI:10.1121/10.0026231

Categories: Literature Watch

Graph Neural Networks with Multi-Features for Predicting Cocrystals Using APIs and coformers Interactions

Fri, 2024-06-07 06:00

Curr Med Chem. 2024 May 22. doi: 10.2174/0109298673290511240404053224. Online ahead of print.

ABSTRACT

Active pharmaceutical ingredients (APIs) have gained direct pharmaceutical interest, along with their in vitro properties, and thus utilized as auxiliary solid dosage forms upon FDA guidance and approval on pharmaceutical cocrystals when reacting with coformers, as a potential and attractive route for drug substance development. However, screening and selecting suitable and appropriate coformers that may potentially react with APIs to successfully form cocrystals is a time-consuming, inefficient, costly, and labour intensive task. In this study, we implemented graph neural networks to predict the formation of cocrystals using our first created API coformers interactions graph dataset. We further compared our work with previous studies that implemented descriptor-based models (e.g., random forest, support vector machine, extreme gradient boosting, and artificial neural networks). All built graph-based models show compelling performance accuracies (i.e., 91.36, 94.60 and 95. 95% for GCN, GraphSAGE, and R-GCN respectively). Furthermore, R-GCN prevailed among the built graph-based models because of its capability to learn the topological structure of the graph from the additionally provided information (i.e., non-ionic and non-covalent interactions or link information) between APIs and coformers.

PMID:38847382 | DOI:10.2174/0109298673290511240404053224

Categories: Literature Watch

Pages