Deep learning

The evaluation of a novel single-lead biopotential device for home sleep testing

Wed, 2024-10-23 06:00

Sleep. 2024 Oct 23:zsae248. doi: 10.1093/sleep/zsae248. Online ahead of print.

ABSTRACT

STUDY OBJECTIVES: This paper reports on the clinical evaluation of the sleep staging performance of a novel single-lead biopotential device.

METHODS: 133 patients suspected of obstructive sleep apnea were included in a multi-site cohort. All patients underwent polysomnography and received the study device, a single-lead biopotential measurement device attached to the forehead. Clinical endpoint parameters were selected to evaluate the device's ability to determine sleep stages. Finally, the device's performance was compared to the clinical study results of comparable devices.

RESULTS: Concurrent PSG and study device data were successfully acquired for 106 of the 133 included patients. The results of this study demonstrated significant similarity in overall sleep staging performance (5-stage Cohen's Kappa of 0.70) to the best-performing reduced-lead biopotential device to which it was compared (5-stage Cohen's Kappa of 0.73). Contrary to the comparator devices, the study device reported a higher Cohen's Kappa for REM (0.78) compared to N3 (0.61), which can be explained by its particular measuring electrode placement (diagonally across the lateral cross-section of the eye). This placement was optimized to ensure the polarity of rapid eye movements could be adequately captured, enhancing the capacity to discriminate between N3 and REM sleep when using only a single-lead setup.

CONCLUSIONS: The results of this study demonstrate the feasibility of incorporating a single-lead biopotential extension in a reduced-channel home sleep apnea testing setup. Such incorporation could narrow the gap in the functionality of reduced-channel home sleep testing and in-lab polysomnography without compromising the patient's ease of use and comfort.

PMID:39441980 | DOI:10.1093/sleep/zsae248

Categories: Literature Watch

Uncertainty Qualification for Deep Learning-Based Elementary Reaction Property Prediction

Wed, 2024-10-23 06:00

J Chem Inf Model. 2024 Oct 23. doi: 10.1021/acs.jcim.4c01358. Online ahead of print.

ABSTRACT

The prediction of the thermodynamic and kinetic properties of elementary reactions has shown rapid improvement due to the implementation of deep learning (DL) methods. While various studies have reported the success in predicting reaction properties, the quantification of prediction uncertainty has seldom been investigated, thus compromising the confidence in using these predicted properties in practical applications. Here, we integrated graph convolutional neural networks (GCNN) with three uncertainty prediction techniques, including deep ensemble, Monte Carlo (MC)-dropout, and evidential learning, to provide insights into the uncertainty quantification and utility. The deep ensemble model outperforms others in accuracy and shows the highest reliability in estimating prediction uncertainty across all elementary reaction property data sets. We also verified that the deep ensemble model showed a satisfactory capability in recognizing epistemic and aleatoric uncertainties. Additionally, we adopted a Monte Carlo Tree Search method for extracting the explainable reaction substructures, providing a chemical explanation for DL predicted properties and corresponding uncertainties. Finally, to demonstrate the utility of uncertainty qualification in practical applications, we performed an uncertainty-guided calibration of the DL-constructed kinetic model, which achieved a 25% higher hit ratio in identifying dominant reaction pathways compared to that of the calibration without uncertainty guidance.

PMID:39441973 | DOI:10.1021/acs.jcim.4c01358

Categories: Literature Watch

EuDockScore: Euclidean graph neural networks for scoring protein-protein interfaces

Wed, 2024-10-23 06:00

Bioinformatics. 2024 Oct 23:btae636. doi: 10.1093/bioinformatics/btae636. Online ahead of print.

ABSTRACT

MOTIVATION: Protein-protein interactions are essential for a variety of biological phenomena including mediating bio-chemical reactions, cell signaling, and the immune response. Proteins seek to form interfaces which reduce overall system energy. Although determination of single polypeptide chain protein structures has been revolutionized by deep learning techniques, complex prediction has still not been perfected. Additionally, experimentally determining structures is incredibly resource and time expensive. An alternative is the technique of computational docking, which takes the solved individual structures of proteins to produce candidate interfaces (decoys). Decoys are then scored using a mathematical function that assess the quality of the system, know as a scoring functions. Beyond docking, scoring functions are a critical component of assessing structures produced by many protein generative models. Scoring models are also used as a final filtering in many generative deep learning models including those that generate antibody binders, and those which perform docking.

RESULTS: In this work we present improved scoring functions for protein-protein interactions which utilizes cutting-edge euclidean graph neural network architectures, to assess protein-protein interfaces. These euclidean docking score models are known as EuDockScore, and EuDockScore-Ab with the latter being antibody-antigen dock specific. Finally, we provided EuDockScore-AFM a model trained on antibody-antigen outputs from AlphaFold-Multimer which proves useful in re-ranking large numbers of AlphaFold-Multimer outputs.

AVAILABILITY: The code for these models is available at https://gitlab.com/mcfeemat/eudockscore.

PMID:39441796 | DOI:10.1093/bioinformatics/btae636

Categories: Literature Watch

An ensemble deep learning model for medical image fusion with Siamese neural networks and VGG-19

Wed, 2024-10-23 06:00

PLoS One. 2024 Oct 23;19(10):e0309651. doi: 10.1371/journal.pone.0309651. eCollection 2024.

ABSTRACT

Multimodal medical image fusion methods, which combine complementary information from many multi-modality medical images, are among the most important and practical approaches in numerous clinical applications. Various conventional image fusion techniques have been developed for multimodality image fusion. Complex procedures for weight map computing, fixed fusion strategy and lack of contextual understanding remain difficult in conventional and machine learning approaches, usually resulting in artefacts that degrade the image quality. This work proposes an efficient hybrid learning model for medical image fusion using pre-trained and non-pre-trained networks i.e. VGG-19 and SNN with stacking ensemble method. The model leveraging the unique capabilities of each architecture, can effectively preserve the detailed information with high visual quality, for numerous combinations of image modalities in image fusion challenges, notably improved contrast, increased resolution, and lower artefacts. Additionally, this ensemble model can be more robust in the fusion of various combinations of source images that are publicly available from Havard-Medical-Image-Fusion Datasets, GitHub. and Kaggle. Our proposed model performance is superior in terms of visual quality and performance metrics to that of the existing fusion methods in literature like PCA+DTCWT, NSCT, DWT, DTCWT+NSCT, GADCT, CNN and VGG-19.

PMID:39441782 | DOI:10.1371/journal.pone.0309651

Categories: Literature Watch

Grade-Skewed Domain Adaptation via Asymmetric Bi-Classifier Discrepancy Minimization for Diabetic Retinopathy Grading

Wed, 2024-10-23 06:00

IEEE Trans Med Imaging. 2024 Oct 23;PP. doi: 10.1109/TMI.2024.3485064. Online ahead of print.

ABSTRACT

Diabetic retinopathy (DR) is a leading cause of preventable blindness worldwide. Deep learning has exhibited promising performance in the grading of DR. Certain deep learning strategies have facilitated convenient regular eye check-ups, which are crucial for managing DR and preventing severe visual impairment. However, the generalization performance on cross-center, cross-vendor, and cross-user test datasets is compromised due to domain shift. Furthermore, the presence of small lesions and the imbalanced grade distribution, resulting from the characteristics of DR grading (e.g., the progressive nature of DR disease and the design of grading standards), complicates image-level domain adaptation for DR grading. The general predictions of the models trained on grade-skewed source domains will be significantly biased toward the majority grades, which further increases the adaptation difficulty. We formulate this problem as a grade-skewed domain adaptation challenge. Under the grade-skewed domain adaptation problem, we propose a novel method for image-level supervised DR grading via Asymmetric Bi-Classifier Discrepancy Minimization (ABiD). First, we propose optimizing the feature extractor by minimizing the discrepancy between the predictions of the asymmetric bi-classifier based on two classification criteria to encourage the exploration of crucial features in adjacent grades and stretch the distribution of adjacent grades in the latent space. Moreover, the classifier difference is maximized by using the forward and inverse distribution compensation mechanism to locate easily confused instances, which avoids pseudolabel bias on the target domain. The experimental results on two public DR datasets and one private DR dataset demonstrate that our method outperforms state-of-the-art methods significantly.

PMID:39441682 | DOI:10.1109/TMI.2024.3485064

Categories: Literature Watch

Robust Myocardial Perfusion MRI Quantification with DeepFermi

Wed, 2024-10-23 06:00

IEEE Trans Biomed Eng. 2024 Oct 23;PP. doi: 10.1109/TBME.2024.3485233. Online ahead of print.

ABSTRACT

Stress perfusion cardiac magnetic resonance is an important technique for examining and assessing the blood supply of the myocardium. Currently, the majority of clinical perfusion scans are evaluated based on visual assessment by experienced clinicians. This makes the process subjective, and to this end, quantitative methods have been proposed to offer a more user-independent assessment of perfusion. These methods, however, rely on time-consuming deconvolution analysis and are susceptible to data outliers caused by artifacts due to cardiac or respiratory motion. In our work, we introduce a novel deep-learning method that integrates the commonly used Fermi function with a neural network architecture for fast, accurate, and robust myocardial perfusion quantification. This approach employs the Fermi model to ensure that the perfusion maps are consistent with measured data, while also utilizing a prior based on a 3D convolutional neural network to generalize spatio-temporal information across different patient data. Our network is trained within a self-supervised learning framework, which circumvents the need for ground-truth perfusion labels that are challenging to obtain. Furthermore, we extended this training methodology by adopting a technique that ensures estimations are resistant to data outliers, thereby improving robustness against motion artifacts. Our simulation experiments demonstrated an overall improvement in the accuracy and robustness of perfusion parameter estimation, consistently outperforming traditional deconvolution analysis algorithms across varying Signal-to-Noise Ratio scenarios in the presence of data outliers. For the in vivo studies, our method generated robust perfusion estimates that aligned with clinical diagnoses, while being approximately five times faster than conventional algorithms.

PMID:39441677 | DOI:10.1109/TBME.2024.3485233

Categories: Literature Watch

Clinical Utility of Deep Learning Assistance for Detecting Various Abnormal Findings in Color Retinal Fundus Images: A Reader Study

Wed, 2024-10-23 06:00

Transl Vis Sci Technol. 2024 Oct 1;13(10):34. doi: 10.1167/tvst.13.10.34.

ABSTRACT

PURPOSE: To evaluate the clinical usefulness of a deep learning-based detection device for multiple abnormal findings on retinal fundus photographs for readers with varying expertise.

METHODS: Fourteen ophthalmologists (six residents, eight specialists) assessed 399 fundus images with respect to 12 major ophthalmologic findings, with or without the assistance of a deep learning algorithm, in two separate reading sessions. Sensitivity, specificity, and reading time per image were compared.

RESULTS: With algorithmic assistance, readers significantly improved in sensitivity for all 12 findings (P < 0.05) but tended to be less specific (P < 0.05) for hemorrhage, drusen, membrane, and vascular abnormality, more profoundly so in residents. Sensitivity without algorithmic assistance was significantly lower in residents (23.1%∼75.8%) compared to specialists (55.1%∼97.1%) in nine findings, but it improved to similar levels with algorithmic assistance (67.8%∼99.4% in residents, 83.2%∼99.5% in specialists) with only hemorrhage remaining statistically significantly lower. Variances in sensitivity were significantly reduced for all findings. Reading time per image decreased in images with fewer than three findings per image, more profoundly in residents. When simulated based on images acquired from a health screening center, average reading time was estimated to be reduced by 25.9% (from 16.4 seconds to 12.1 seconds per image) for residents, and by 2.0% (from 9.6 seconds to 9.4 seconds) for specialists.

CONCLUSIONS: Deep learning-based computer-assisted detection devices increase sensitivity, reduce inter-reader variance in sensitivity, and reduce reading time in less complicated images.

TRANSLATIONAL RELEVANCE: This study evaluated the influence that algorithmic assistance in detecting abnormal findings on retinal fundus photographs has on clinicians, possibly predicting its influence on clinical application.

PMID:39441571 | DOI:10.1167/tvst.13.10.34

Categories: Literature Watch

Deep learning-based approach for acquisition time reduction in ventilation SPECT in patients after lung transplantation

Wed, 2024-10-23 06:00

Radiol Phys Technol. 2024 Oct 23. doi: 10.1007/s12194-024-00853-3. Online ahead of print.

ABSTRACT

We aimed to evaluate the image quality and diagnostic performance of chronic lung allograft dysfunction (CLAD) with lung ventilation single-photon emission computed tomography (SPECT) images acquired briefly using a convolutional neural network (CNN) in patients after lung transplantation and to explore the feasibility of short acquisition times. We retrospectively identified 93 consecutive lung-transplant recipients who underwent ventilation SPECT/computed tomography (CT). We employed a CNN to distinguish the images acquired in full time from those acquired in a short time. The image quality was evaluated using the structural similarity index (SSIM) loss and normalized mean square error (NMSE). The correlation between functional volume/morphological volume (F/M) ratios of full-time SPECT images and predicted SPECT images was evaluated. Differences in the F/M ratio were evaluated using Bland-Altman plots, and the diagnostic performance was compared using the area under the curve (AUC). The learning curve, obtained using MSE, converged within 100 epochs. The NMSE was significantly lower (P < 0.001) and the SSIM was significantly higher (P < 0.001) for the CNN-predicted SPECT images compared to the short-time SPECT images. The F/M ratio of full-time SPECT images and predicted SPECT images showed a significant correlation (r = 0.955, P < 0.0001). The Bland-Altman plot revealed a bias of -7.90% in the F/M ratio. The AUC values were 0.942 for full-time SPECT images, 0.934 for predicted SPECT images and 0.872 for short-time SPECT images. Our findings suggest that a deep-learning-based approach can significantly curtail the acquisition time of ventilation SPECT, while preserving the image quality and diagnostic accuracy for CLAD.

PMID:39441494 | DOI:10.1007/s12194-024-00853-3

Categories: Literature Watch

Deep learning super-resolution reconstruction for fast and high-quality cine cardiovascular magnetic resonance

Wed, 2024-10-23 06:00

Eur Radiol. 2024 Oct 23. doi: 10.1007/s00330-024-11145-0. Online ahead of print.

ABSTRACT

OBJECTIVES: To compare standard-resolution balanced steady-state free precession (bSSFP) cine images with cine images acquired at low resolution but reconstructed with a deep learning (DL) super-resolution algorithm.

MATERIALS AND METHODS: Cine cardiovascular magnetic resonance (CMR) datasets (short-axis and 4-chamber views) were prospectively acquired in healthy volunteers and patients at normal (cineNR: 1.89 × 1.96 mm2, reconstructed at 1.04 × 1.04 mm2) and at a low-resolution (2.98 × 3.00 mm2, reconstructed at 1.04 × 1.04 mm2). Low-resolution images were reconstructed using compressed sensing DL denoising and resolution upscaling (cineDL). Left ventricular ejection fraction (LVEF), end-diastolic volume index (LVEDVi), and strain were assessed. Apparent signal-to-noise (aSNR) and contrast-to-noise ratios (aCNR) were calculated. Subjective image quality was assessed on a 5-point Likert scale. Student's paired t-test, Wilcoxon matched-pairs signed-rank-test, and intraclass correlation coefficient (ICC) were used for statistical analysis.

RESULTS: Thirty participants were analyzed (37 ± 16 years; 20 healthy volunteers and 10 patients). Short-axis views whole-stack acquisition duration of cineDL was shorter than cineNR (57.5 ± 8.7 vs 98.7 ± 12.4 s; p < 0.0001). No differences were noted for: LVEF (59 ± 7 vs 59 ± 7%; ICC: 0.95 [95% confidence interval: 0.94, 0.99]; p = 0.17), LVEDVi (85.0 ± 13.5 vs 84.4 ± 13.7 mL/m2; ICC: 0.99 [0.98, 0.99]; p = 0.12), longitudinal strain (-19.5 ± 4.3 vs -19.8 ± 3.9%; ICC: 0.94 [0.88, 0.97]; p = 0.52), short-axis aSNR (81 ± 49 vs 69 ± 38; p = 0.32), aCNR (53 ± 31 vs 45 ± 27; p = 0.33), or subjective image quality (5.0 [IQR 4.9, 5.0] vs 5.0 [IQR 4.7, 5.0]; p = 0.99).

CONCLUSION: Deep-learning reconstruction of cine images acquired at a lower spatial resolution led to a decrease in acquisition times of 42% with shorter breath-holds without affecting volumetric results or image quality.

KEY POINTS: Question Cine CMR acquisitions are time-intensive and vulnerable to artifacts. Findings Low-resolution upscaled reconstructions using DL super-resolution decreased acquisition times by 35-42% without a significant difference in volumetric results or subjective image quality. Clinical relevance DL super-resolution reconstructions of bSSFP cine images acquired at a lower spatial resolution reduce acquisition times while preserving diagnostic accuracy, improving the clinical feasibility of cine imaging by decreasing breath hold duration.

PMID:39441391 | DOI:10.1007/s00330-024-11145-0

Categories: Literature Watch

Accelerating Evidence Synthesis in Observational Studies: Development of a Living Natural Language Processing-Assisted Intelligent Systematic Literature Review System

Wed, 2024-10-23 06:00

JMIR Med Inform. 2024 Oct 23;12:e54653. doi: 10.2196/54653.

ABSTRACT

BACKGROUND: Systematic literature review (SLR), a robust method to identify and summarize evidence from published sources, is considered to be a complex, time-consuming, labor-intensive, and expensive task.

OBJECTIVE: This study aimed to present a solution based on natural language processing (NLP) that accelerates and streamlines the SLR process for observational studies using real-world data.

METHODS: We followed an agile software development and iterative software engineering methodology to build a customized intelligent end-to-end living NLP-assisted solution for observational SLR tasks. Multiple machine learning-based NLP algorithms were adopted to automate article screening and data element extraction processes. The NLP prediction results can be further reviewed and verified by domain experts, following the human-in-the-loop design. The system integrates explainable articificial intelligence to provide evidence for NLP algorithms and add transparency to extracted literature data elements. The system was developed based on 3 existing SLR projects of observational studies, including the epidemiology studies of human papillomavirus-associated diseases, the disease burden of pneumococcal diseases, and cost-effectiveness studies on pneumococcal vaccines.

RESULTS: Our Intelligent SLR Platform covers major SLR steps, including study protocol setting, literature retrieval, abstract screening, full-text screening, data element extraction from full-text articles, results summary, and data visualization. The NLP algorithms achieved accuracy scores of 0.86-0.90 on article screening tasks (framed as text classification tasks) and macroaverage F1 scores of 0.57-0.89 on data element extraction tasks (framed as named entity recognition tasks).

CONCLUSIONS: Cutting-edge NLP algorithms expedite SLR for observational studies, thus allowing scientists to have more time to focus on the quality of data and the synthesis of evidence in observational studies. Aligning the living SLR concept, the system has the potential to update literature data and enable scientists to easily stay current with the literature related to observational studies prospectively and continuously.

PMID:39441204 | DOI:10.2196/54653

Categories: Literature Watch

Self-Supervised Learning for Generic Raman Spectrum Denoising

Wed, 2024-10-23 06:00

Anal Chem. 2024 Oct 23. doi: 10.1021/acs.analchem.4c01550. Online ahead of print.

ABSTRACT

Raman and surface-enhanced Raman scattering (SERS) spectroscopy are highly specific and sensitive optical modalities that have been extensively investigated in diverse applications. Noise reduction is demanding in the preprocessing procedure, especially for weak Raman/SERS spectra. Existing denoising methods require manual optimization of parameters, which is time-consuming and laborious and cannot always achieve satisfactory performance. Deep learning has been increasingly applied in Raman/SERS spectral denoising but usually requires massive training data, where the true labels may not exist. Aiming at these challenges, this work presents a generic Raman spectrum denoising algorithm with self-supervised learning for accurate, rapid, and robust noise reduction. A specialized network based on U-Net is established, which first extracts high-level features and then restores key peak profiles of the spectra. A subsampling strategy is proposed to refine the raw Raman spectrum and avoid the underlying biased interference. The effectiveness of the proposed approach has been validated by a broad range of spectral data, exhibiting its strong generalization ability. In the context of photosafe detection of deep-seated tumors, our method achieved signal-to-noise ratio enhancement by over 400%, which resulted in a significant increase in the limit of detection thickness from 10 to 18 cm. Our approach demonstrates superior denoising performance compared to the state-of-the-art denoising methods. The occlusion method further showed that the proposed algorithm automatically focuses on characterized peaks, enhancing the interpretability of our approach explicitly in Raman and SERS spectroscopy.

PMID:39441128 | DOI:10.1021/acs.analchem.4c01550

Categories: Literature Watch

Extraction of agricultural plastic greenhouses based on a U-Net Convolutional Neural Network Coupled with edge expansion and loss function improvement

Wed, 2024-10-23 06:00

J Air Waste Manag Assoc. 2024 Oct 23. doi: 10.1080/10962247.2024.2412708. Online ahead of print.

ABSTRACT

Compared to traditional interpretation methods, which suffer problems such as heavy workloads, small adaptation ranges and poor repeatability, deep learning network models can better extract target features from remote sensing images. In this study, we used GF-7 image data to improve the traditional U-Net convolutional neural network (CNN) model. The Canny operator and Gaussian kernel (GK) function were used for sample edge expansion, and the binary cross-entropy and GK functions were used to jointly constrain the loss. Finally, APGs were accurately extracted and compared to those obtained with the original model. The results indicated that the APG extraction accuracy of the U-Net network was improved through the expansion of sample edge information and adoption of joint loss function constraints.

PMID:39440842 | DOI:10.1080/10962247.2024.2412708

Categories: Literature Watch

A deep-learning-based histopathology classifier for focal cortical dysplasia (FCD) unravels a complex scenario of comorbid FCD subtypes

Wed, 2024-10-23 06:00

Epilepsia. 2024 Oct 23. doi: 10.1111/epi.18161. Online ahead of print.

ABSTRACT

OBJECTIVE: Recently, we developed a first artificial intelligence (AI)-based digital pathology classifier for focal cortical dysplasia (FCD) as defined by the ILAE classification. Herein, we tested the usefulness of the classifier in a retrospective histopathology workup scenario.

METHODS: Eighty-six new cases with histopathologically confirmed FCD ILAE type Ia (FCDIa), FCDIIa, FCDIIb, mild malformation of cortical development with oligodendroglial hyperplasia in epilepsy (MOGHE), or mild malformations of cortical development were selected, 20 of which had confirmed gene mosaicism.

RESULTS: The classifier always recognized the correct histopathology diagnosis in four or more 1000 × 1000-μm digital tiles in all cases. Furthermore, the final diagnosis overlapped with the largest batch of tiles assigned by the algorithm to one diagnostic entity in 80.2% of all cases. However, 86.2% of all cases revealed more than one diagnostic category. As an example, FCDIIb was identified in all of the 23 patients with histopathologically assigned FCDIIb, whereas the classifier correctly recognized FCDIIa tiles in 19 of these cases (83%), that is, dysmorphic neurons but no balloon cells. In contrast, the classifier misdiagnosed FCDIIb tiles in seven of 23 cases histopathologically assigned to FCDIIa (33%). This mandates a second look by the signing histopathologist to either confirm balloon cells or differentiate from reactive astrocytes. The algorithm also recognized coexisting architectural dysplasia, for example, vertically oriented microcolumns as in FCDIa, in 22% of cases classified as FCDII and in 62% of cases with MOGHE. Microscopic review confirmed microcolumns in the majority of tiles, suggesting that vertically oriented architectural abnormalities are more common than previously anticipated.

SIGNIFICANCE: An AI-based diagnostic classifier will become a helpful tool in our future histopathology laboratory, in particular when large anatomical resections from epilepsy surgery require extensive resources. We also provide an open access web application allowing the histopathologist to virtually review digital tiles obtained from epilepsy surgery to corroborate their final diagnosis.

PMID:39440630 | DOI:10.1111/epi.18161

Categories: Literature Watch

CellSAM: Advancing Pathologic Image Cell Segmentation via Asymmetric Large-Scale Vision Model Feature Distillation Aggregation Network

Wed, 2024-10-23 06:00

Microsc Res Tech. 2024 Oct 23. doi: 10.1002/jemt.24716. Online ahead of print.

ABSTRACT

Segment anything model (SAM) has attracted extensive interest as a potent large-scale image segmentation model, with prior efforts adapting it for use in medical imaging. However, the precise segmentation of cell nucleus instances remains a formidable challenge in computational pathology, given substantial morphological variations and the dense clustering of nuclei with unclear boundaries. This study presents an innovative cell segmentation algorithm named CellSAM. CellSAM has the potential to improve the effectiveness and precision of disease identification and therapy planning. As a variant of SAM, CellSAM integrates dual-image encoders and employs techniques such as knowledge distillation and mask fusion. This innovative model exhibits promising capabilities in capturing intricate cell structures and ensuring adaptability in resource-constrained scenarios. The experimental results indicate that this structure effectively enhances the quality and precision of cell segmentation. Remarkably, CellSAM demonstrates outstanding results even with minimal training data. In the evaluation of particular cell segmentation tasks, extensive comparative analyzes show that CellSAM outperforms both general fundamental models and state-of-the-art (SOTA) task-specific models. Comprehensive evaluation metrics yield scores of 0.884, 0.876, and 0.768 for mean accuracy, recall, and precision respectively. Extensive experiments show that CellSAM excels in capturing subtle details and complex structures and is capable of segmenting cells in images accurately. Additionally, CellSAM demonstrates excellent performance on clinical data, indicating its potential for robust applications in treatment planning and disease diagnosis, thereby further improving the efficiency of computer-aided medicine.

PMID:39440549 | DOI:10.1002/jemt.24716

Categories: Literature Watch

Reconstructing Molecular Networks by Causal Diffusion Do-Calculus Analysis with Deep Learning

Wed, 2024-10-23 06:00

Adv Sci (Weinh). 2024 Oct 23:e2409170. doi: 10.1002/advs.202409170. Online ahead of print.

ABSTRACT

Quantifying molecular regulations between genes/molecules causally from observed data is crucial for elucidating the molecular mechanisms underlying biological processes at the network level. Presently, most methods for inferring gene regulatory and biological networks rely on association studies or observational causal-analysis approaches. This study introduces a novel approach that combines intervention operations and diffusion models within a do-calculus framework by deep learning, i.e., Causal Diffusion Do-calculus (CDD) analysis, to infer causal networks between molecules. CDD can extract causal relations from observed data owing to its intervention operations, thereby significantly enhancing the accuracy and generalizability of causal network inference. Computationally, CDD has been applied to both simulated data and real omics data, which demonstrates that CDD outperforms existing methods in accurately inferring gene regulatory networks and identifying causal links from genes to disease phenotypes. Especially, compared with the Mendelian randomization algorithm and other existing methods, the CDD can reliably identify the disease genes or molecules for complex diseases with better performances. In addition, the causal analysis between various diseases and the potential factors in different populations from the UK Biobank database is also conducted, which further validated the effectiveness of CDD.

PMID:39440482 | DOI:10.1002/advs.202409170

Categories: Literature Watch

Text-image multimodal fusion model for enhanced fake news detection

Wed, 2024-10-23 06:00

Sci Prog. 2024 Oct-Dec;107(4):368504241292685. doi: 10.1177/00368504241292685.

ABSTRACT

In the era of rapid internet expansion and technological progress, discerning real from fake news poses a growing challenge, exposing users to potential misinformation. The existing literature primarily focuses on analyzing individual features in fake news, overlooking multimodal feature fusion recognition. Compared to single-modal approaches, multimodal fusion allows for a more comprehensive and enriched capture of information from different data modalities (such as text and images), thereby improving the performance and effectiveness of the model. This study proposes a model using multimodal fusion to identify fake news, aiming to curb misinformation. The framework integrates textual and visual information, using early fusion, joint fusion and late fusion strategies to combine them. The proposed framework processes textual and visual information through data cleaning and feature extraction before classification. Fake news classification is accomplished through a model, achieving accuracy of 85% and 90% in the Gossipcop and Fakeddit datasets, with F1-scores of 90% and 88%, showcasing its performance. The study presents outcomes across different training periods, demonstrating the effectiveness of multimodal fusion in combining text and image recognition for combating fake news. This research contributes significantly to addressing the critical issue of misinformation, emphasizing a comprehensive approach for detection accuracy enhancement.

PMID:39440371 | DOI:10.1177/00368504241292685

Categories: Literature Watch

Development of AI-assisted microscopy frameworks through realistic simulation with pySTED

Wed, 2024-10-23 06:00

Nat Mach Intell. 2024;6(10):1197-1215. doi: 10.1038/s42256-024-00903-w. Epub 2024 Sep 26.

ABSTRACT

The integration of artificial intelligence into microscopy systems significantly enhances performance, optimizing both image acquisition and analysis phases. Development of artificial intelligence-assisted super-resolution microscopy is often limited by access to large biological datasets, as well as by difficulties to benchmark and compare approaches on heterogeneous samples. We demonstrate the benefits of a realistic stimulated emission depletion microscopy simulation platform, pySTED, for the development and deployment of artificial intelligence strategies for super-resolution microscopy. pySTED integrates theoretically and empirically validated models for photobleaching and point spread function generation in stimulated emission depletion microscopy, as well as simulating realistic point-scanning dynamics and using a deep learning model to replicate the underlying structures of real images. This simulation environment can be used for data augmentation to train deep neural networks, for the development of online optimization strategies and to train reinforcement learning models. Using pySTED as a training environment allows the reinforcement learning models to bridge the gap between simulation and reality, as showcased by its successful deployment on a real microscope system without fine tuning.

PMID:39440349 | PMC:PMC11491398 | DOI:10.1038/s42256-024-00903-w

Categories: Literature Watch

Deep neural network-based robotic visual servoing for satellite target tracking

Wed, 2024-10-23 06:00

Front Robot AI. 2024 Oct 8;11:1469315. doi: 10.3389/frobt.2024.1469315. eCollection 2024.

ABSTRACT

In response to the costly and error-prone manual satellite tracking on the International Space Station (ISS), this paper presents a deep neural network (DNN)-based robotic visual servoing solution to the automated tracking operation. This innovative approach directly addresses the critical issue of motion decoupling, which poses a significant challenge in current image moment-based visual servoing. The proposed method uses DNNs to estimate the manipulator's pose, resulting in a significant reduction of coupling effects, which enhances control performance and increases tracking precision. Real-time experimental tests are carried out using a 6-DOF Denso manipulator equipped with an RGB camera and an object, mimicking the targeting pin. The test results demonstrate a 32.04% reduction in pose error and a 21.67% improvement in velocity precision compared to conventional methods. These findings demonstrate that the method has the potential to improve efficiency and accuracy significantly in satellite target tracking and capturing.

PMID:39440297 | PMC:PMC11494149 | DOI:10.3389/frobt.2024.1469315

Categories: Literature Watch

Developing predictive precision medicine models by exploiting real-world data using machine learning methods

Wed, 2024-10-23 06:00

J Appl Stat. 2024 Feb 13;51(14):2980-3003. doi: 10.1080/02664763.2024.2315451. eCollection 2024.

ABSTRACT

Computational Medicine encompasses the application of Statistical Machine Learning and Artificial Intelligence methods on several traditional medical approaches, including biochemical testing which is extremely valuable both for early disease prognosis and long-term individual monitoring, as it can provide important information about a person's health status. However, using Statistical Machine Learning and Artificial Intelligence algorithms to analyze biochemical test data from Electronic Health Records requires several preparatory steps, such as data manipulation and standardization. This study presents a novel approach for utilizing Electronic Health Records from large, real-world databases to develop predictive precision medicine models by exploiting Artificial Intelligence. Furthermore, to demonstrate the effectiveness of this approach, we compare the performance of various traditional Statistical Machine Learning and Deep Learning algorithms in predicting individuals' future biochemical test outcomes. Specifically, using data from a large real-world database, we exploit a longitudinal format of the data in order to predict the future values of 15 biochemical tests and identify individuals at high risk. The proposed approach and the extensive model comparison contribute to the personalized approach that modern medicine aims to achieve.

PMID:39440239 | PMC:PMC11492405 | DOI:10.1080/02664763.2024.2315451

Categories: Literature Watch

Deep learning-driven ultrasound-assisted diagnosis: optimizing GallScopeNet for precise identification of biliary atresia

Wed, 2024-10-23 06:00

Front Med (Lausanne). 2024 Oct 8;11:1445069. doi: 10.3389/fmed.2024.1445069. eCollection 2024.

ABSTRACT

BACKGROUND: Biliary atresia (BA) is a severe congenital biliary developmental abnormality threatening neonatal health. Traditional diagnostic methods rely heavily on experienced radiologists, making the process time-consuming and prone to variability. The application of deep learning for the automated diagnosis of BA remains underexplored.

METHODS: This study introduces GallScopeNet, a deep learning model designed to improve diagnostic efficiency and accuracy through innovative architecture and advanced feature extraction techniques. The model utilizes data from a carefully constructed dataset of gallbladder ultrasound images. A dataset comprising thousands of ultrasound images was employed, with the majority used for training and validation and a subset reserved for external testing. The model's performance was evaluated using five-fold cross-validation and external assessment, employing metrics such as accuracy and the area under the receiver operating characteristic curve (AUC), compared against clinical diagnostic standards.

RESULTS: GallScopeNet demonstrated exceptional performance in distinguishing BA from non-BA cases. In the external test dataset, GallScopeNet achieved an accuracy of 81.21% and an AUC of 0.85, indicating strong diagnostic capabilities. The results highlighted the model's ability to maintain high classification performance, reducing misdiagnosis and missed diagnosis.

CONCLUSION: GallScopeNet effectively differentiates between BA and non-BA images, demonstrating significant potential and reliability for early diagnosis. The system's high efficiency and accuracy suggest it could serve as a valuable diagnostic tool in clinical settings, providing substantial technical support for improving diagnostic workflows.

PMID:39440041 | PMC:PMC11493747 | DOI:10.3389/fmed.2024.1445069

Categories: Literature Watch

Pages