Deep learning

Who is WithMe? EEG features for attention in a visual task, with auditory and rhythmic support

Mon, 2025-01-27 06:00

Front Neurosci. 2025 Jan 10;18:1434444. doi: 10.3389/fnins.2024.1434444. eCollection 2024.

ABSTRACT

INTRODUCTION: The study of attention has been pivotal in advancing our comprehension of cognition. The goal of this study is to investigate which EEG data representations or features are most closely linked to attention, and to what extent they can handle the cross-subject variability.

METHODS: We explore the features obtained from the univariate time series from a single EEG channel, such as time domain features and recurrence plots, as well as representations obtained directly from the multivariate time series, such as global field power or functional brain networks. To address the cross-subject variability in EEG data, we also investigate persistent homology features that are robust to different types of noise. The performance of the different EEG representations is evaluated with the Support Vector Machine (SVM) accuracy on the WithMe data derived from a modified digit span experiment, and is benchmarked against baseline EEG-specific models, including a deep learning architecture known for effectively learning task-specific features.

RESULTS: The raw EEG time series outperform each of the considered data representations, but can fall short in comparison with the black-box deep learning approach that learns the best features.

DISCUSSION: The findings are limited to the WithMe experimental paradigm, highlighting the need for further studies on diverse tasks to provide a more comprehensive understanding of their utility in the analysis of EEG data.

PMID:39867449 | PMC:PMC11758281 | DOI:10.3389/fnins.2024.1434444

Categories: Literature Watch

Reusable specimen-level inference in computational pathology

Mon, 2025-01-27 06:00

ArXiv [Preprint]. 2025 Jan 10:arXiv:2501.05945v1.

ABSTRACT

Foundation models for computational pathology have shown great promise for specimen-level tasks and are increasingly accessible to researchers. However, specimen-level models built on these foundation models remain largely unavailable, hindering their broader utility and impact. To address this gap, we developed SpinPath, a toolkit designed to democratize specimen-level deep learning by providing a zoo of pretrained specimen-level models, a Python-based inference engine, and a JavaScript-based inference platform. We demonstrate the utility of SpinPath in metastasis detection tasks across nine foundation models. SpinPath may foster reproducibility, simplify experimentation, and accelerate the adoption of specimen-level deep learning in computational pathology research.

PMID:39867428 | PMC:PMC11759856

Categories: Literature Watch

An easy-to-use three-dimensional protein-structure-prediction online platform "DPL3D" based on deep learning algorithms

Mon, 2025-01-27 06:00

Curr Res Struct Biol. 2025 Jan 3;9:100163. doi: 10.1016/j.crstbi.2024.100163. eCollection 2025 Jun.

ABSTRACT

The change in the three-dimensional (3D) structure of a protein can affect its own function or interaction with other protein(s), which may lead to disease(s). Gene mutations, especially missense mutations, are the main cause of changes in protein structure. Due to the lack of protein crystal structure data, about three-quarters of human mutant proteins cannot be predicted or accurately predicted, and the pathogenicity of missense mutations can only be indirectly evaluated by evolutionary conservation. Recently, many computational methods have been developed to predict protein 3D structures with accuracy comparable to experiments. This progress enables the information of structural biology to be further utilized by clinicians. Thus, we developed a user-friendly platform named DPL3D (http://nsbio.tech:3000) which can predict and visualize the 3D structure of mutant proteins. The crystal structure and other information of proteins were downloaded together with the software including AlphaFold 2, RoseTTAFold, RoseTTAFold All-Atom, and trRosettaX-Single. We implemented a query module for 210,180 molecular structures, including 52,248 human proteins. Visualization of protein two-dimensional (2D) and 3D structure prediction can be generated via LiteMol automatically or manually and interactively. This platform will allow users to easily and quickly retrieve large-scale structural information for biological discovery.

PMID:39867105 | PMC:PMC11761317 | DOI:10.1016/j.crstbi.2024.100163

Categories: Literature Watch

Technological Advancements in Augmented, Mixed, and Virtual Reality Technologies for Surgery: A Systematic Review

Mon, 2025-01-27 06:00

Cureus. 2024 Dec 26;16(12):e76428. doi: 10.7759/cureus.76428. eCollection 2024 Dec.

ABSTRACT

Recent advancements in artificial intelligence (AI) have shown significant potential in the medical field, although many applications are still in the research phase. This paper provides a comprehensive review of advancements in augmented reality (AR), mixed reality (MR), and virtual reality (VR) for surgical applications from 2019 to 2024 to accelerate the transition of AI from the research to the clinical phase. This paper also provides an overview of proposed databases for further use in extended reality (XR), which includes AR, MR, and VR, as well as a summary of typical research applications involving XR in surgical practices. Additionally, this paper concludes by discussing challenges and proposed solutions for the application of XR in the medical field. Although the areas of focus and specific implementations vary among AR, MR, and VR, current trends in XR focus mainly on reducing workload and minimizing surgical errors through navigation, training, and machine learning-based visualization. Through analyzing these trends, AR and MR have greater advantages for intraoperative surgical functions, whereas VR is limited to preoperative training and surgical preparation. VR faces additional limitations, and its use has been reduced in research since the first applications of XR, which likely suggests the same will happen with further development. Nonetheless, with increased access to technology and the ability to overcome the black box problem, XR's applications in medical fields and surgery will increase to guarantee further accuracy and precision while reducing risk and workload.

PMID:39867005 | PMC:PMC11763273 | DOI:10.7759/cureus.76428

Categories: Literature Watch

Enhancing prostate cancer segmentation in bpMRI: Integrating zonal awareness into attention-guided U-Net

Mon, 2025-01-27 06:00

Digit Health. 2025 Jan 24;11:20552076251314546. doi: 10.1177/20552076251314546. eCollection 2025 Jan-Dec.

ABSTRACT

PURPOSE: Prostate cancer (PCa) is the second most common cancer in males worldwide, requiring improvements in diagnostic imaging to identify and treat it at an early stage. Bi-parametric magnetic resonance imaging (bpMRI) is recognized as an essential diagnostic technique for PCa, providing shorter acquisition times and cost-effectiveness. Nevertheless, accurate diagnosis using bpMRI images is difficult due to the inconspicuous and diverse characteristics of malignant tumors and the intricate structure of the prostate gland. An automated system is required to assist the medical professionals in accurate and early diagnosis with less effort.

METHOD: This study recognizes the impact of zonal features on the advancement of the disease. The aim is to improve the diagnostic performance through a novel automated approach of a two-step mechanism using bpMRI images. First, pretraining a convolutional neural network (CNN)-based attention-guided U-Net model for segmenting the region of interest which is carried out in the prostate zone. Secondly, pretraining the same type of Attention U-Net is performed for lesion segmentation.

RESULTS: The performance of the pretrained models and training an attention-guided U-Net from the scratch for segmenting tumors on the prostate region is analyzed. The proposed attention-guided U-Net model achieved an area under the curve (AUC) of 0.85 and a dice similarity coefficient value of 0.82, outperforming some other pretrained deep learning models.

CONCLUSION: Our approach greatly enhances the identification and categorization of clinically significant PCa by including zonal data. Our approach exhibits exceptional performance in the accurate segmentation of bpMRI images compared to current techniques, as evidenced by thorough validation of a diverse dataset. This research not only enhances the field of medical imaging for oncology but also underscores the potential of deep learning models to progress PCa diagnosis and personalized patient care.

PMID:39866889 | PMC:PMC11758924 | DOI:10.1177/20552076251314546

Categories: Literature Watch

G2PDeep-v2: a web-based deep-learning framework for phenotype prediction and biomarker discovery for all organisms using multi-omics data

Mon, 2025-01-27 06:00

Res Sq [Preprint]. 2025 Jan 9:rs.3.rs-5776937. doi: 10.21203/rs.3.rs-5776937/v1.

ABSTRACT

The G2PDeep-v2 server is a web-based platform powered by deep learning, for phenotype prediction and markers discovery from multi-omics data in any organisms including humans, plants, animals, and viruses. The server provides multiple services for researchers to create deep-learning models through an interactive interface and train these models using an automated hyperparameter tuning algorithm on high-performance computing resources. Users can visualize the results of phenotype and markers predictions and perform Gene Set Enrichment Analysis for the significant markers to provide insights into the molecular mechanisms underlying complex diseases, conditions and other biological phenotypes being studied. The G2PDeep-v2 server is publicly available at https://g2pdeep.org/ and can be utilized for all organisms.

PMID:39866874 | PMC:PMC11760241 | DOI:10.21203/rs.3.rs-5776937/v1

Categories: Literature Watch

OCDet: A comprehensive ovarian cell detection model with channel attention on immunohistochemical and morphological pathology images

Sun, 2025-01-26 06:00

Comput Biol Med. 2025 Jan 25;186:109713. doi: 10.1016/j.compbiomed.2025.109713. Online ahead of print.

ABSTRACT

BACKGROUND: Ovarian cancer is among the most lethal gynecologic malignancy that threatens women's lives. Pathological diagnosis is a key tool for early detection and diagnosis of ovarian cancer, guiding treatment strategies. The evaluation of various ovarian cancer-related cells, based on morphological and immunohistochemical pathology images, is deemed an important step. Currently, the lack of a comprehensive deep learning framework for detecting various ovarian cells poses a performance bottleneck in ovarian cancer pathological diagnosis.

METHOD: This paper presents OCDet, an object detection model with channel attention, which achieves comprehensive detection of CD3, CD8, and CD20 positive lymphocytes in immunohistochemical pathology slides, and neutrophils and polyploid giant cancer cells in H&E slides of ovarian cancer. OCDet, utilizing CSPDarkNet as its backbone, incorporates an Efficient Channel Attention module for Resolution-Specified Embedding Refinement and Multi-Resolution Embedding Fusion, enabling the efficient extraction of pathological features.

RESULT: The experiment demonstrated that OCDet performed well in target detection of three types of positive lymphocytes in immunohistochemical images, as well as neutrophils and polyploid giant cancer cells in H&E images. The mAP@0.5 reached 98.82 %, 92.91 %, and 90.49 % respectively, all surpassing other compared models. The ablation experiment further highlighted the superiority of the introduced Efficient Channel Attention (ECA) mechanism.

CONCLUSION: The proposed OCDet enables accurate detection of multiple types of cells in immunohistochemical and morphological pathology images of ovarian cancer, serving as an efficient application tool for pathological diagnosis thereof. The proposed framework has the potential to be further applied to other cancer types.

PMID:39864335 | DOI:10.1016/j.compbiomed.2025.109713

Categories: Literature Watch

A robust and generalized framework in diabetes classification across heterogeneous environments

Sun, 2025-01-26 06:00

Comput Biol Med. 2025 Jan 25;186:109720. doi: 10.1016/j.compbiomed.2025.109720. Online ahead of print.

ABSTRACT

Diabetes mellitus (DM) represents a major global health challenge, affecting a diverse range of demographic populations across all age groups. It has particular implications for women during pregnancy and the postpartum period. The contemporary prevalence of sedentary lifestyle patterns and suboptimal dietary practices has substantially contributed to the escalating incidence of this metabolic disorder. The timely identification of diabetes mellitus (DM) in the female population is crucial for preventing related complications and facilitating the implementation of effective therapeutic interventions. However, conventional predictive models frequently demonstrate limited external validity when applied across heterogeneous datasets, potentially compromising clinical utility. This study proposes a robust machine learning (ML) framework for diabetes prediction across diverse populations using two distinct datasets: the PIMA and BD datasets. The framework employs intra-dataset, inter-dataset, and partial fusion dataset validation techniques to comprehensively assess the generalizability and performance of various models. In intra-dataset validation, the Extreme Gradient Boosting (XGBoost) model achieved the highest accuracy on the PIMA dataset with 79%. In contrast, the Random Forest (RF) and Gradient Boosting (GB) models demonstrated accuracy close to 99% on the BD dataset. For inter-dataset validation, where models were trained on one dataset and tested on the other, the ensemble model outperformed others with 88% accuracy when trained on PIMA and tested on BD. However, model performance declined when trained on BD and tested on PIMA (74%), reflecting the challenges of inter-dataset generalization ability. Finally, during partial fusion data validation, the deep learning (DL) model achieved 74% accuracy when trained on the BD dataset augmented with 30% of the PIMA dataset. This accuracy increased to 98% when training on the PIMA dataset combined with 30% of the BD data. These findings emphasize the importance of dataset diversity and the partial fusion dataset that can significantly enhance the model's robustness and generalizability. This framework offers valuable insights into the complexities of diabetes prediction across heterogeneous environments.

PMID:39864329 | DOI:10.1016/j.compbiomed.2025.109720

Categories: Literature Watch

Sensitivity-enhanced hydrogel digital RT-LAMP with in situ enrichment and interfacial reaction for norovirus quantification in food and water

Sun, 2025-01-26 06:00

J Hazard Mater. 2025 Jan 21;488:137325. doi: 10.1016/j.jhazmat.2025.137325. Online ahead of print.

ABSTRACT

Low levels of human norovirus (HuNoV) in food and environment present challenges for nucleic acid detection. This study reported an evaporation-enhanced hydrogel digital reverse transcription loop-mediated isothermal amplification (HD RT-LAMP) with interfacial enzymatic reaction for sensitive HuNoV quantification in food and water. By drying samples on a chamber array chip, HuNoV particles were enriched in situ. The interfacial amplification of nucleic acid at the hydrogel-chip interface was triggered after coating HD RT-LAMP system. Nanoconfined spaces in hydrogels provided a simple and rapid "digital format" to quantify single virus within 15 min. Through in situ evaporation for enrichment, the sensitivity level was increased by 20 times. The universality of the sensitivity-enhanced assay was also verified using other bacteria and virus. Furthermore, a deep learning model and smartphone app were developed for automatic amplicon analysis. Multiple actual samples, including 3 lake waters, strawberry, tap water and drinking water, were in situ enriched and detected for norovirus quantification using the chamber arrays. Therefore, the sensitivity-enhanced HD RT-LAMP is an efficient assay for testing biological hazards in food safety monitoring and environmental surveillance.

PMID:39864200 | DOI:10.1016/j.jhazmat.2025.137325

Categories: Literature Watch

Integrating a novel algorithm in assessing the impact of floods on the genetic diversity of a high commercial value fish (Cyprinidae: Spinibarbus sp.) in Lang Son province of Vietnam

Sun, 2025-01-26 06:00

Zoology (Jena). 2025 Jan 20;168:126240. doi: 10.1016/j.zool.2025.126240. Online ahead of print.

ABSTRACT

Floods, which occur when the amount of precipitation surpasses the capacity of an area to drain it adequately, have detrimental consequences on the survival and future generations of fishes. However, few works have reported the prediction of this natural phenomenon in a relation to certain fish species, especially in fast-flowing rivers. In the specific context of the northern mountainous provinces of Vietnam, where the Spinibarbus sp. fish species resides, it has been observed through the current study that the fish population in Lang Son exhibits the lowest genetic diversity and genetic distance. Consequently, the population of Spinibarbus sp. in Lang Son shows a heightened susceptibility to floods, resulting in reduction in population size and compromised population resilience. In order to provide decision support information for managers, conservationists, and researchers, we have employed a genetic algorithm-support vector machine regression (GA-SVR) predictive model to map flood vulnerability using thirteen dependent variables. The study findings have unveiled a significant negative correlation between flood-sensitive regions and genetic diversity. These discoveries emphasize the significance of considering the impact of floods on the genetic diversity of Spinibarbus sp. in Lang Son through flood vulnerability mapping. This underscores the value of establishing a comprehensive framework based on the GA-SVR algorithm for early flood detection, thereby facilitating the implementation of effective measures to minimize damages and conserve this commercial fish species.

PMID:39864169 | DOI:10.1016/j.zool.2025.126240

Categories: Literature Watch

Artificial Intelligence in Pancreatic Imaging: A Systematic Review

Sun, 2025-01-26 06:00

United European Gastroenterol J. 2025 Jan 26. doi: 10.1002/ueg2.12723. Online ahead of print.

ABSTRACT

The rising incidence of pancreatic diseases, including acute and chronic pancreatitis and various pancreatic neoplasms, poses a significant global health challenge. Pancreatic ductal adenocarcinoma (PDAC) for example, has a high mortality rate due to late-stage diagnosis and its inaccessible location. Advances in imaging technologies, though improving diagnostic capabilities, still necessitate biopsy confirmation. Artificial intelligence, particularly machine learning and deep learning, has emerged as a revolutionary force in healthcare, enhancing diagnostic precision and personalizing treatment. This narrative review explores Artificial intelligence's role in pancreatic imaging, its technological advancements, clinical applications, and associated challenges. Following the PRISMA-DTA guidelines, a comprehensive search of databases including PubMed, Scopus, and Cochrane Library was conducted, focusing on Artificial intelligence, machine learning, deep learning, and radiomics in pancreatic imaging. Articles involving human subjects, written in English, and published up to March 31, 2024, were included. The review process involved title and abstract screening, followed by full-text review and refinement based on relevance and novelty. Recent Artificial intelligence advancements have shown promise in detecting and diagnosing pancreatic diseases. Deep learning techniques, particularly convolutional neural networks (CNNs), have been effective in detecting and segmenting pancreatic tissues as well as differentiating between benign and malignant lesions. Deep learning algorithms have also been used to predict survival time, recurrence risk, and therapy response in pancreatic cancer patients. Radiomics approaches, extracting quantitative features from imaging modalities such as CT, MRI, and endoscopic ultrasound, have enhanced the accuracy of these deep learning models. Despite the potential of Artificial intelligence in pancreatic imaging, challenges such as legal and ethical considerations, algorithm transparency, and data security remain. This review underscores the transformative potential of Artificial intelligence in enhancing the diagnosis and treatment of pancreatic diseases, ultimately aiming to improve patient outcomes and survival rates.

PMID:39865461 | DOI:10.1002/ueg2.12723

Categories: Literature Watch

PSMA PET/CT based multimodal deep learning model for accurate prediction of pelvic lymph-node metastases in prostate cancer patients identified as candidates for extended pelvic lymph node dissection by preoperative nomograms

Sun, 2025-01-26 06:00

Eur J Nucl Med Mol Imaging. 2025 Jan 27. doi: 10.1007/s00259-024-07065-2. Online ahead of print.

ABSTRACT

PURPOSE: To develop and validate a prostate-specific membrane antigen (PSMA) PET/CT based multimodal deep learning model for predicting pathological lymph node invasion (LNI) in prostate cancer (PCa) patients identified as candidates for extended pelvic lymph node dissection (ePLND) by preoperative nomograms.

METHODS: [68Ga]Ga-PSMA-617 PET/CT scan of 116 eligible PCa patients (82 in the training cohort and 34 in the test cohort) who underwent radical prostatectomy with ePLND were analyzed in our study. The Med3D deep learning network was utilized to extract discriminative features from the entire prostate volume of interest on the PET/CT images. Subsequently, a multimodal model i.e., Multi kernel Support Vector Machine was constructed to combine the PET/CT deep learning features, quantitative PET and clinical parameters. The performance of the multimodal models was assessed using final histopathology as the reference standard, with evaluation metrics including area under the receiver operating characteristic curve (AUC), calibration curve, decision curve analysis, and compared with available nomograms and PET/CT visual evaluation result.

RESULTS: Our multimodal model incorporated clinical information, maximum standardized uptake value (SUVmax), and PET/CT deep learning features. The AUC for predicting LNI was 0.89 (95% confidence interval [CI] 0.81-0.97) for the final model. The proposed model demonstrated superior predictive accuracy in the test cohort compared to PET/CT visual evaluation result, the Memorial Sloan Kettering Cancer Center (MSKCC) and the Briganti-2017 nomograms (AUC 0.85 [95% CI 0.69-1.00] vs. 0.80 [95% CI 0.64-0.95] vs. 0.79 [95% CI 0.61-0.97] and 0.69 [95% CI 0.50-0.88], respectively). The proposed model showed similar calibration and higher net benefit as compared to the traditional nomograms.

CONCLUSION: Our multimodal deep learning model, which incorporates preoperative PSMA PET/CT imaging, shows enhanced predictive capabilities for LNI in clinically localized PCa compared to PSMA PET/CT visual evaluation result and existing nomograms like the MSKCC and Briganti-2017 nomograms. This model has the potential to reduce unnecessary ePLND procedures while minimizing the risk of missing cases of LNI.

PMID:39865180 | DOI:10.1007/s00259-024-07065-2

Categories: Literature Watch

Capsule network approach for monkeypox (CAPSMON) detection and subclassification in medical imaging system

Sun, 2025-01-26 06:00

Sci Rep. 2025 Jan 26;15(1):3296. doi: 10.1038/s41598-025-87993-7.

ABSTRACT

In response to the pressing need for the detection of Monkeypox caused by the Monkeypox virus (MPXV), this study introduces the Enhanced Spatial-Awareness Capsule Network (ESACN), a Capsule Network architecture designed for the precise multi-class classification of dermatological images. Addressing the shortcomings of traditional Machine Learning and Deep Learning models, our ESACN model utilizes the dynamic routing and spatial hierarchy capabilities of CapsNets to differentiate complex patterns such as those seen in monkeypox, chickenpox, measles, and normal skin presentations. CapsNets' inherent ability to recognize and process crucial spatial relationships within images outperforms conventional CNNs, particularly in tasks that require the distinction of visually similar classes. Our model's superior performance, demonstrated through rigorous evaluation, exhibits significant improvements in accuracy, precision, recall, and F1 score, even with limited data. The results highlight the potential of ESACN as a reliable tool for enhancing diagnostic accuracy in medical settings. In our case study, the ESACN model was applied to a dataset comprising 659 images across four classes: 178 images of Monkeypox, 171 of Chickenpox, 80 of Measles, and 230 of Normal skin conditions. This case study underscores the model's effectiveness in real-world applications, providing robust and accurate classification that could greatly aid in early diagnosis and treatment planning in clinical environments.

PMID:39865160 | DOI:10.1038/s41598-025-87993-7

Categories: Literature Watch

Disorder-induced enhancement of lithium-ion transport in solid-state electrolytes

Sun, 2025-01-26 06:00

Nat Commun. 2025 Jan 26;16(1):1057. doi: 10.1038/s41467-025-56322-x.

ABSTRACT

Enhancing the ion conduction in solid electrolytes is critically important for the development of high-performance all-solid-state lithium-ion batteries (LIBs). Lithium thiophosphates are among the most promising solid electrolytes, as they exhibit superionic conductivity at room temperature. However, the lack of comprehensive understanding of their ion conduction mechanism, especially the effect of structural disorder on ionic conductivity, is a long-standing problem that limits further innovations in all-solid-state LIBs. Here, we address this challenge by establishing and employing a deep learning potential to simulate Li3PS4 electrolyte systems with varying levels of disorder. The results show that disorder-driven diffusion dynamics significantly enhances the room-temperature conductivity. We further establish bridges between dynamical characteristics, local structural features, and atomic rearrangements by applying a machine learning-based structure fingerprint termed "softness". This metric allows the classification of the disorder-induced "soft" hopping lithium ions. Our findings offer insights into ion conduction mechanisms in complex disordered structures, thereby contributing to the development of superior solid-state electrolytes for LIBs.

PMID:39865086 | DOI:10.1038/s41467-025-56322-x

Categories: Literature Watch

Potential Use and Limitation of Artificial Intelligence to Screen Diabetes Mellitus in Clinical Practice: A Literature Review

Sun, 2025-01-26 06:00

Acta Med Indones. 2024 Oct;56(4):563-570.

ABSTRACT

The burden of undiagnosed diabetes mellitus (DM) is substantial, with approximately 240 million individuals globally unaware of their condition, disproportionately affecting low- and middle-income countries (LMICs), including Indonesia. Without screening, DM and its complications will impose significant pressure on healthcare systems. Current clinical practices for screening and diagnosing DM primarily involve blood or laboratory-based testing which possess limitations on access and cost. To address these challenges, researchers have developed risk-scoring tools to identify high-risk populations. However, considering generalizability, artificial intelligence (AI) technologies offer a promising approach, leveraging diverse data sources for improved accuracy. AI models (i.e., machine learning and deep learning) have yielded prediction performances of up to 98% in various diseases. This article underscores the potential of AI-driven approaches in reducing the burden of DM through accurate prediction of undiagnosed diabetes while highlighting the need for continued innovation and collaboration in healthcare delivery.

PMID:39865054

Categories: Literature Watch

Regional Image Quality Scoring for 2-D Echocardiography Using Deep Learning

Sun, 2025-01-26 06:00

Ultrasound Med Biol. 2025 Jan 25:S0301-5629(24)00469-1. doi: 10.1016/j.ultrasmedbio.2024.12.008. Online ahead of print.

ABSTRACT

OBJECTIVE: To develop and compare methods to automatically estimate regional ultrasound image quality for echocardiography separate from view correctness.

METHODS: Three methods for estimating image quality were developed: (i) classic pixel-based metric: the generalized contrast-to-noise ratio (gCNR), computed on myocardial segments (region of interest) and left ventricle lumen (background), extracted by a U-Net segmentation model; (ii) local image coherence: the average local coherence as predicted by a U-Net model that predicts image coherence from B-mode ultrasound images at the pixel level; (iii) deep convolutional network: an end-to-end deep-learning model that predicts the quality of each region in the image directly. These methods were evaluated against manual regional quality annotations provided by three experienced cardiologists.

RESULTS: The results indicated poor performance of the gCNR metric, with Spearman correlation to annotations of ρ = 0.24. The end-to-end learning model obtained the best result, ρ = 0.69, comparable to the inter-observer correlation, ρ = 0.63. Finally, the coherence-based method, with ρ = 0.58, out-performed the classical metrics and was more generic than the end-to-end approach.

CONCLUSION: The deep convolutional network provided the most accurate regional quality prediction, while the coherence-based method offered a more generalizable solution. gCNR showed limited effectiveness in this study. The image quality prediction tool is available as an open-source Python library at https://github.com/GillesVanDeVyver/arqee.

PMID:39864961 | DOI:10.1016/j.ultrasmedbio.2024.12.008

Categories: Literature Watch

Development of a Clinically Applicable Deep Learning System Based on Sparse Training Data to Accurately Detect Acute Intracranial Hemorrhage from Non-enhanced Head Computed Tomography

Sun, 2025-01-26 06:00

Neurol Med Chir (Tokyo). 2025 Jan 24. doi: 10.2176/jns-nmc.2024-0163. Online ahead of print.

ABSTRACT

Non-enhanced head computed tomography is widely used for patients presenting with head trauma or stroke, given acute intracranial hemorrhage significantly influences clinical decision-making. This study aimed to develop a deep learning algorithm, referred to as DeepCT, to detect acute intracranial hemorrhage on non-enhanced head computed tomography images and evaluate its clinical applicability. We retrospectively collected 1,815 computed tomography image sets from a single center for model training. Additional computed tomography sets from 3 centers were used to construct an independent validation dataset (VAL) and 2 test datasets (GPS-C and DICH). A third test dataset (US-TW) comprised 150 cases, each from 1 hospital in Taiwan and 1 hospital in the United States of America. Our deep learning model, based on U-Net and ResNet architectures, was implemented using PyTorch. The deep learning algorithm exhibited high accuracy across the validation and test datasets, with overall accuracy ranging from 0.9343 to 0.9820. Our findings show that the deep learning algorithm effectively identifies acute intracranial hemorrhage in non-enhanced head computed tomography studies. Clinically, this algorithm can be used for hyperacute triage, reducing reporting times, and enhancing the accuracy of radiologist interpretations. The evaluation of the algorithm on both United States and Taiwan datasets further supports its universal reliability for detecting acute intracranial hemorrhage.

PMID:39864839 | DOI:10.2176/jns-nmc.2024-0163

Categories: Literature Watch

Multimodal optimal matching and augmentation method for small sample gesture recognition

Sun, 2025-01-26 06:00

Biosci Trends. 2025 Jan 25. doi: 10.5582/bst.2024.01370. Online ahead of print.

ABSTRACT

In human-computer interaction, gesture recognition based on physiological signals offers advantages such as a more natural and fast interaction mode and less constrained by the environment than visual-based. Surface electromyography-based gesture recognition has significantly progressed. However, since individuals have physical differences, researchers must collect data multiple times from each user to train the deep learning model. This data acquisition process can be particularly burdensome for non-healthy users. Researchers are currently exploring transfer learning and data augmentation techniques to enhance the accuracy of small-sample gesture recognition models. However, challenges persist, such as negative transfer and limited diversity in training samples, leading to suboptimal recognition performance. Therefore, We introduce motion information into sEMG-based recognition and propose a multimodal optimal matching and augmentation method for small sample gesture recognition, achieving efficient gesture recognition with only one acquisition per gesture. Firstly, this method utilizes the optimal matching signal selection module to select the most similar signals from the existing data to the new user as the training set, reducing inter-domain differences. Secondly, the similarity calculation augmentation module enhances the diversity of the training set. Finally, the Modal-type embedding enhances the information interaction between each mode signal. We evaluated the effectiveness on Self-collected Stroke Patient, the Ninapro DB1 dataset and the Ninapro DB5 dataset and achieved accuracies of 93.69%, 91.65% and 98.56%, respectively. These results demonstrate that the method achieved performance comparable to traditional recognition models while significantly reducing the collected data.

PMID:39864830 | DOI:10.5582/bst.2024.01370

Categories: Literature Watch

Joint image reconstruction and segmentation of real-time cardiac MRI in free-breathing using a model based on disentangled representation learning

Sun, 2025-01-26 06:00

J Cardiovasc Magn Reson. 2025 Jan 24:101844. doi: 10.1016/j.jocmr.2025.101844. Online ahead of print.

ABSTRACT

PURPOSE: To investigate image quality and agreement of derived cardiac function parameters in a novel joint image reconstruction and segmentation approach based on disentangled representation learning, enabling real-time cardiac cine imaging during free-breathing.

METHODS: A multi-tasking neural network architecture, incorporating disentangled representation learning, was trained using simulated examinations based on data from a public repository along with MR scans specifically acquired for model development. An exploratory feasibility study evaluated the method on undersampled real-time acquisitions using an in-house developed spiral bSSFP pulse sequence in eight healthy participants and five patients with intermittent atrial fibrillation. Images and predicted LV segmentations were compared to the reference standard of ECG-gated segmented Cartesian cine with repeated breath-holds and corresponding manual segmentation.

RESULTS: On a 5-point Likert scale, image quality of the real-time breath-hold approach and Cartesian cine was comparable in healthy participants (RT-BH: 1.99 ±.98, Cartesian: 1.94 ±.86, p=.052), but slightly inferior in free-breathing (RT-FB: 2.40 ±.98, p<.001). In patients with arrhythmia, both real-time approaches demonstrated favourable image quality (RT-BH: 2.10 ± 1.28, p<.001, RT-FB: 2.40 ± 1.13, p<.01, Cartesian: 2.68 ± 1.13). Intra-observer reliability was good (ICC=.77,95%-confidence interval [.75,.79], p<.001). In functional analysis, a positive bias was observed for ejection fractions derived from the proposed model compared to the clinical reference standard (RT-BH mean: 58.5 ± 5.6%, bias: +3.47%, 95%-confidence interval [-.86, 7.79%], RT-FB mean: 57.9 ± 10.6%, bias: +1.45%, [-3.02, 5.91%], Cartesian mean: 54.9 ± 6.7%).

CONCLUSION: The introduced real-time MR imaging technique enables high-quality cardiac cine data acquisitions in 1-2minutes, eliminating the need for ECG gating and breath-holds. This approach offers a promising alternative to the current clinical practice of segmented acquisition, with shorter scan times, improved patient comfort, and increased robustness to arrhythmia and patient non-compliance.

PMID:39864743 | DOI:10.1016/j.jocmr.2025.101844

Categories: Literature Watch

Quantifying Knee-Adjacent Subcutaneous Fat in the Entire OAI Baseline Dataset - Associations with Cartilage MRI T<sub>2</sub>, Thickness and Pain, Independent of BMI

Sun, 2025-01-26 06:00

Osteoarthritis Cartilage. 2025 Jan 24:S1063-4584(25)00018-4. doi: 10.1016/j.joca.2025.01.001. Online ahead of print.

ABSTRACT

OBJECTIVE: Knee-adjacent subcutaneous fat (kaSCF) has emerged as a potential biomarker and risk factor for OA progression. This study aims to develop an AI-based tool for the automatic segmentation of kaSCF thickness and evaluate the cross-sectional associations between kaSCF, cartilage thickness, MRI-based cartilage T2 relaxation time, knee pain, and muscle strength independent of BMI.

DESIGN: Baseline 3.0T MR images of the right knee from the entire Osteoarthritis Initiative (OAI) cohort (n=4796) were used to quantify average values of kaSCF, cartilage thickness, and T2 using deep learning algorithms. Regression models (adjusted for age, gender, BMI, and race) were used to evaluate the associations between standardized kaSCF and outcomes of cartilage thickness, T2, pain, and knee extension strength.

RESULTS: Model prediction CVs for kaSCF thickness ranged from 3.57% to 9.87% across femoral and tibial regions. Greater average kaSCF was associated with thinner cartilage in men (std. β= -0.029, 95% CI: -0.050 to -0.007, p=0.010) and higher T2 in women (std. β=0.169, 95% CI: 0.072 to 0.265, p=0.001). Greater kaSCF was also associated with lower knee extension force (std. β= -15.36, 95% CI: -20.39 to -10.33, p<0.001) and higher odds of frequent knee pain (std. OR=1.156, 95% CI: 1.046 to 1.278, p=0.005) across all participants.

CONCLUSIONS: Greater kaSCF was associated with thinner cartilage in men, higher T2 in women, reduced knee strength, and greater knee pain, independent of BMI. These findings suggest a potential role of kaSCF as a predictor for KOA-related structural, functional, and clinical outcomes independent of the effects of BMI.

PMID:39864732 | DOI:10.1016/j.joca.2025.01.001

Categories: Literature Watch

Pages