Deep learning

Improving TMJ Diagnosis: A Deep Learning Approach for Detecting Mandibular Condyle Bone Changes

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 17;15(8):1022. doi: 10.3390/diagnostics15081022.

ABSTRACT

Objectives: This paper evaluates the potential of using deep learning approaches for the detection of degenerative bone changes in the mandibular condyle. The aim of this study is to enable the detection and diagnosis of mandibular condyle degenerations, which are difficult to observe and diagnose on panoramic radiographs, using deep learning methods. Methods: A total of 3875 condylar images were obtained from panoramic radiographs. Condylar bone changes were represented by flattening, osteophyte, and erosion, and images in which two or more of these changes were observed were labeled as "other". Due to the limited number of images containing osteophytes and erosion, two approaches were used. In the first approach, images containing osteophytes and erosion were combined into the "other" group, resulting in three groups: normal, flattening, and deformation ("deformation" encompasses the "other" group, together with osteophyte and erosion). In the second approach, images containing osteophytes and erosion were completely excluded, resulting in three groups: normal, flattening, and other. The study utilizes a range of advanced deep learning algorithms, including Dense Networks, Residual Networks, VGG Networks, and Google Networks, which are pre-trained with transfer learning techniques. Model performance was evaluated using datasets with different distributions, specifically 70:30 and 80:20 training-test splits. Results: The GoogleNet architecture achieved the highest accuracy. Specifically, with the 80:20 split of the normal-flattening-deformation dataset and the Adamax optimizer, an accuracy of 95.23% was achieved. The results demonstrate that CNN-based methods are highly successful in determining mandibular condyle bone changes. Conclusions: This study demonstrates the potential of deep learning, particularly CNNs, for the accurate and efficient detection of TMJ-related condylar bone changes from panoramic radiographs. This approach could assist clinicians in identifying patients requiring further intervention. Future research may involve using cross-sectional imaging methods and training the right and left condyles together to potentially increase the success rate. This approach has the potential to improve the early detection of TMJ-related condylar bone changes, enabling timely referrals and potentially preventing disease progression.

PMID:40310446 | DOI:10.3390/diagnostics15081022

Categories: Literature Watch

Artificial Intelligence in Oral Diagnosis: Detecting Coated Tongue with Convolutional Neural Networks

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 17;15(8):1024. doi: 10.3390/diagnostics15081024.

ABSTRACT

Background/Objectives: Coated tongue is a common oral condition with notable clinical relevance, often overlooked due to its asymptomatic nature. Its presence may reflect poor oral hygiene and can serve as an early indicator of underlying systemic diseases. This study aimed to develop a robust diagnostic model utilizing convolutional neural networks and machine learning classifiers to improve the detection of coated tongue lesions. Methods: A total of 200 tongue images (100 coated and 100 healthy) were analyzed. Images were acquired using a DSLR camera (Nikon D5500 with Sigma Macro 105 mm lens, Nikon, Tokyo, Japan) under standardized daylight conditions. Following preprocessing, feature vectors were extracted using CNN architectures (VGG16, VGG19, ResNet, MobileNet, and NasNet) and classified using Support Vector Machine (SVM), K-Nearest Neighbors (KNN), and Multi-Layer Perceptron (MLP) classifiers. Performance metrics included sensitivity, specificity, accuracy, and F1 score. Results: The SVM + VGG19 hybrid model achieved the best performance among all tested configurations, with a sensitivity of 82.6%, specificity of 88.23%, accuracy of 85%, and an F1 score of 86.36%. Conclusions: The SVM + VGG19 model demonstrated high accuracy and reliability in diagnosing coated tongue lesions, highlighting its potential as an effective clinical decision support tool. Future research with larger datasets may further enhance model robustness and applicability in diverse populations.

PMID:40310445 | DOI:10.3390/diagnostics15081024

Categories: Literature Watch

Integrating Machine Learning and Deep Learning for Predicting Non-Surgical Root Canal Treatment Outcomes Using Two-Dimensional Periapical Radiographs

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 16;15(8):1009. doi: 10.3390/diagnostics15081009.

ABSTRACT

Background/Objectives: In a previous study, we utilized categorical variables and machine learning (ML) algorithms to predict the success of non-surgical root canal treatments (NSRCTs) in apical periodontitis (AP), classifying the outcome as either success (healed) or failure (not healed). Given the importance of radiographic imaging in diagnosis, the present study evaluates the efficacy of deep learning (DL) in predicting NSRCT outcomes using two-dimensional (2D) periapical radiographs, comparing its performance with ML models. Methods: The DL model was trained and validated using leave-one-out cross-validation (LOOCV). Its output was incorporated into the set of categorical variables, and the ML study was reproduced using backward stepwise selection (BSS). The chi-square test was applied to assess the association between this new variable and NSRCT outcomes. Finally, after identifying the best-performing method from the ML study reproduction, statistical comparisons were conducted between this method, clinical professionals, and the image-based model using Fisher's exact test. Results: The association study yielded a p-value of 0.000000127, highlighting the predictive capability of 2D radiographs. After incorporating the DL-based predictive variable, the ML algorithm that demonstrated the best performance was logistic regression (LR), differing from the previous study, where random forest (RF) was the top performer. When comparing the deep learning-logistic regression (DL-LR) model with the clinician's prognosis (DP), DL-LR showed superior performance with a statistically significant difference (p-value < 0.05) in sensitivity, NPV, and accuracy. The same trend was observed in the DL vs. DP comparison. However, no statistically significant differences were found in the comparisons of RF vs. DL-LR, RF vs. DL, or DL vs. DL-LR. Conclusions: The findings of this study suggest that image-based artificial intelligence models exhibit superior predictive capability compared with those relying exclusively on categorical data. Moreover, they outperform clinician prognosis.

PMID:40310439 | DOI:10.3390/diagnostics15081009

Categories: Literature Watch

Stain Normalization of Histopathological Images Based on Deep Learning: A Review

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 18;15(8):1032. doi: 10.3390/diagnostics15081032.

ABSTRACT

Histopathological images stained with hematoxylin and eosin (H&E) are crucial for cancer diagnosis and prognosis. However, color variations caused by differences in tissue preparation and scanning devices can lead to data distribution discrepancies, adversely affecting the performance of downstream algorithms in tasks like classification, segmentation, and detection. To address these issues, stain normalization methods have been developed to standardize color distributions across images from various sources. Recent advancements in deep learning-based stain normalization methods have shown significant promise due to their minimal preprocessing requirements, independence from reference templates, and robustness. This review examines 115 publications to explore the latest developments in this field. We first outline the evaluation metrics and publicly available datasets used for assessing stain normalization methods. Next, we systematically review deep learning-based approaches, including supervised, unsupervised, and self-supervised methods, categorizing them by core technologies and analyzing their contributions and limitations. Finally, we discuss current challenges and future directions, aiming to provide researchers with a comprehensive understanding of the field, promote further development, and accelerate the progress of intelligent cancer diagnosis.

PMID:40310413 | DOI:10.3390/diagnostics15081032

Categories: Literature Watch

Machine Learning and Deep Learning for Healthcare Data Processing and Analyzing: Towards Data-Driven Decision-Making and Precise Medicine

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 21;15(8):1051. doi: 10.3390/diagnostics15081051.

ABSTRACT

Artificial intelligence (AI) is reshaping the landscape of healthcare data [...].

PMID:40310409 | DOI:10.3390/diagnostics15081051

Categories: Literature Watch

Fractal-Based Architectures with Skip Connections and Attention Mechanism for Improved Segmentation of MS Lesions in Cervical Spinal Cord

Thu, 2025-05-01 06:00

Diagnostics (Basel). 2025 Apr 19;15(8):1041. doi: 10.3390/diagnostics15081041.

ABSTRACT

Background/Objectives: Multiple sclerosis (MS) is an autoimmune disease that damages the myelin sheath of the central nervous system, which includes the brain and spinal cord. Although MS lesions in the brain are more frequently investigated, MS lesions in the cervical spinal cord (CSC) can be much more specific for the diagnosis of the disease. Furthermore, as lesion burden in the CSC is directly related to disease progression, the presence of lesions in the CSC may help to differentiate MS from other neurological diseases. Methods: In this study, two novel deep learning models based on fractal architectures are proposed for the automatic detection and segmentation of MS lesions in the CSC by improving the convolutional and connection structures used in the layers of the U-Net architecture. In our previous study, we introduced the FractalSpiNet architecture by incorporating fractal convolutional block structures into the U-Net framework to develop a deeper network for segmenting MS lesions in the CPC. In this study, to improve the detection of smaller structures and finer details in the images, an attention mechanism is integrated into the FractalSpiNet architecture, resulting in the Att-FractalSpiNet model. In addition, in the second hybrid model, a fractal convolutional block is incorporated into the skip connection structure of the U-Net architecture, resulting in the development of the Con-FractalU-Net model. Results: Experimental studies were conducted using U-Net, FractalSpiNet, Con-FractalU-Net, and Att-FractalSpiNet architectures to detect the CSC region and the MS lesions within its boundaries. In segmenting the CSC region, the proposed Con-FractalU-Net architecture achieved the highest Dice Similarity Coefficient (DSC) score of 98.89%. Similarly, in detecting MS lesions within the CSC region, the Con-FractalU-Net model again achieved the best performance with a DSC score of 91.48%. Conclusions: For segmentation of the CSC region and detection of MS lesions, the proposed fractal-based Con-FractalU-Net and Att-FractalSpiNet architectures achieved higher scores than the baseline U-Net architecture, particularly in segmenting small and complex structures.

PMID:40310404 | DOI:10.3390/diagnostics15081041

Categories: Literature Watch

TabNet and TabTransformer: Novel Deep Learning Models for Chemical Toxicity Prediction in Comparison With Machine Learning

Thu, 2025-05-01 06:00

J Appl Toxicol. 2025 May 1. doi: 10.1002/jat.4803. Online ahead of print.

ABSTRACT

The prediction of chemical toxicity is crucial for applications in drug discovery, environmental safety, and regulatory assessments. This study aims to evaluate the performance of advanced deep learning architectures, TabNet and TabTransformer, in comparison to traditional machine learning methods, for predicting the toxicity of chemical compounds across 12 toxicological endpoints. The dataset consisted of 12,228 training and 3057 test samples, each characterized by 801 molecular descriptors representing chemical and structural features. Traditional machine learning models, including XGBoost, CatBoost, SVM, and a voting classifier, were paired with feature selection techniques such as principal component analysis (PCA), recursive feature elimination (RFE), and mutual information (MI). Advanced architectures, TabNet and TabTransformer, were trained directly on the full feature set without dimensionality reduction. Model performance was assessed using accuracy, F1-score, AUC-ROC, AUPR, and Matthews correlation coefficient (MCC), alongside SHAP analysis to interpret feature importance and enhance model transparency under class imbalance conditions. Cross-validation and test set evaluations ensured robust comparisons across all models and toxicological endpoints. TabNet and TabTransformer consistently outperformed traditional classifiers, achieving AUC-ROC values up to 96% for endpoints such as SR.ARE and SR.p53. TabTransformer showed the highest performance on complex labels, benefiting from self-attention mechanisms that captured intricate feature relationships, while TabNet achieved competitive outcomes with an efficient, dynamic feature selection. In addition to standard metrics, we reported AUPR and MCC to better evaluate model performance under class imbalance, with both models maintaining high scores across endpoints. Although traditional classifiers, particularly the voting classifier, performed well when combined with feature selection-achieving up to 94% AUC-ROC on SR.p53-they lagged behind the deep learning models in generalizability and feature interaction modeling. SHAP analysis further highlighted the interpretability of the proposed architectures by identifying influential descriptors such as VSAEstate6 and MoRSEE8. This study highlights the superiority of TabNet and TabTransformer in predicting chemical toxicity while ensuring interpretability through SHAP analysis. These models offer a promising alternative to traditional in vitro and in vivo approaches, paving the way for cost-effective and ethical toxicity assessments.

PMID:40309751 | DOI:10.1002/jat.4803

Categories: Literature Watch

EfficientNetB0-Based End-to-End Diagnostic System for Diabetic Retinopathy Grading and Macular Edema Detection

Thu, 2025-05-01 06:00

Diabetes Metab Syndr Obes. 2025 Apr 26;18:1311-1321. doi: 10.2147/DMSO.S506494. eCollection 2025.

ABSTRACT

PURPOSE: This study aims to develop and validate a deep learning-based automated diagnostic system that utilizes fluorescein angiography (FFA) images for the rapid and accurate diagnosis of diabetic retinopathy (DR) and its complications.

METHODS: We collected 19,031 FFA images from 2753 patients between June 2017 and March 2024 to construct and evaluate our analytical framework. The images were preprocessed and annotated for training and validating the deep learning model. The study employed a two-stage deep learning system: the first stage used EfficientNetB0 for a five-class classification task to differentiate between normal retinal conditions, various stages of DR, and post-laser treatment status; the second stage focused on images classified as abnormal in the first stage, further detecting the presence of diabetic macular edema (DME). Model performance was evaluated using multiple classification metrics, including accuracy, AUC, precision, recall, F1-score, and Cohen's kappa coefficient.

RESULTS: In the first stage, the model achieved an accuracy of 0.7036 and an AUC of 0.9062 on the test set, demonstrating high accuracy and discriminative ability. In the second stage, the model achieved an accuracy of 0.7258 and an AUC of 0.7530, performing well. Additionally, through Grad-CAM (gradient-weighted class activation mapping), we visualized the most influential image regions in the model's decision-making process, enhancing the model's interpretability.

CONCLUSION: This study successfully developed an end-to-end DR diagnostic system based on the EfficientNetB0 model. The system not only automates the grading of FFA images but also detects DME, significantly reducing the time required for image interpretation by clinicians and providing an effective tool to improve the efficiency and accuracy of DR diagnosis.

PMID:40309724 | PMC:PMC12042962 | DOI:10.2147/DMSO.S506494

Categories: Literature Watch

Approach for enhancing the accuracy of semantic segmentation of chest X-ray images by edge detection and deep learning integration

Thu, 2025-05-01 06:00

Front Artif Intell. 2025 Apr 16;8:1522730. doi: 10.3389/frai.2025.1522730. eCollection 2025.

ABSTRACT

INTRODUCTION: Accurate segmentation of anatomical structures in chest X-ray images remains challenging, especially for regions with low contrast and overlapping structures. This limitation significantly affects the diagnosis of cardiothoracic diseases. Existing deep learning methods often struggle with preserving structural boundaries, leading to segmentation artifacts.

METHODS: To address these challenges, I propose a novel segmentation approach that integrates contour detection techniques with the U-net deep learning architecture. Specifically, the method employs Sobel and Scharr edge detection filters to enhance structural boundaries in chest X-ray images before segmentation. The pipeline involves pre-processing using contour detection, followed by segmentation with a U-net model trained to identify lungs, heart, and clavicles.

RESULTS: Experimental evaluation demonstrated that using edge-enhancing filters, particularly the Sobel operator, leads to a marked improvement in segmentation accuracy. For lung segmentation, the model achieved an accuracy of 99.26%, a Dice coefficient of 98.88%, and a Jaccard index of 97.54%. Heart segmentation results included 99.47% accuracy and 94.14% Jaccard index, while clavicle segmentation reached 99.79% accuracy and 89.57% Jaccard index. These results consistently outperform the baseline U-net model without edge enhancement.

DISCUSSION: The integration of contour detection methods with the U-net model significantly improves the segmentation quality of complex anatomical regions in chest X-rays. Among the tested filters, the Sobel operator proved to be the most effective in enhancing boundary information and reducing segmentation artifacts. This approach offers a promising direction for more accurate and robust computer-aided diagnosis systems in radiology.

PMID:40309721 | PMC:PMC12040918 | DOI:10.3389/frai.2025.1522730

Categories: Literature Watch

Deep linear matrix approximate reconstruction with integrated BOLD signal denoising reveals reproducible hierarchical brain connectivity networks from multiband multi-echo fMRI

Thu, 2025-05-01 06:00

Front Neurosci. 2025 Apr 16;19:1577029. doi: 10.3389/fnins.2025.1577029. eCollection 2025.

ABSTRACT

The hierarchical modular functional structure in the human brain has not been adequately depicted by conventional functional magnetic resonance imaging (fMRI) acquisition techniques and traditional functional connectivity reconstruction methods. Fortunately, rapid advancements in fMRI scanning techniques and deep learning methods open a novel frontier to map the spatial hierarchy within Brain Connectivity Networks (BCNs). The novel multiband multi-echo (MBME) fMRI technique has increased spatiotemporal resolution and peak functional sensitivity, while the advanced deep linear model (multilayer-stacked) named DEep Linear Matrix Approximate Reconstruction (DELMAR) enables the identification of hierarchical features without extensive hyperparameter tuning. We incorporate a multi-echo blood oxygenation level-dependent (BOLD) signal and DELMAR for denoising in its first layer, thereby eliminating the need for a separate multi-echo independent component analysis (ME-ICA) denoising step. Our results demonstrate that the DELMAR/Denoising/Mapping strategy produces more accurate and reproducible hierarchical BCNs than traditional ME-ICA denoising followed by DELMAR. Additionally, we showcase that MBME fMRI outperforms multiband (MB) fMRI in terms of hierarchical BCN mapping accuracy and precision. These reproducible spatial hierarchies in BCNs have significant potential for developing improved fMRI diagnostic and prognostic biomarkers of functional connectivity across a wide range of neurological and psychiatric disorders.

PMID:40309655 | PMC:PMC12040835 | DOI:10.3389/fnins.2025.1577029

Categories: Literature Watch

Dynamic Prediction and Intervention of Serum Sodium in Patients with Stroke Based on Attention Mechanism Model

Thu, 2025-05-01 06:00

J Healthc Inform Res. 2025 Mar 6;9(2):174-190. doi: 10.1007/s41666-025-00192-x. eCollection 2025 Jun.

ABSTRACT

Abnormal serum sodium levels are a common and severe complication in stroke patients, significantly increasing mortality risk and prolonging ICU stays. Accurate real-time prediction of serum sodium fluctuations is crucial for optimizing clinical interventions. However, existing predictive models face limitations in handling complex dynamic features and long time series data, making them less effective in guiding individualized treatment. To address this challenge, this study developed a deep learning model based on a multi-head attention mechanism to enable real-time prediction of serum sodium concentrations and provide personalized intervention recommendations for ICU stroke patients. This study utilized publicly available MIMIC-III (n = 2346) and MIMIC-IV (n = 896) datasets, extracting time series data from 10 key clinical indicators closely associated with serum sodium levels. To address the complexity of long time series data, a moving sliding window sub-sampling segmentation method was employed, effectively transforming extensive sequences into more manageable inputs while preserving critical temporal dependencies. By leveraging advanced mathematical modeling, meaningful insights were extracted from sparse and irregular time series data. The resulting time-feature fusion multi-head attention (TFF-MHA) model underwent rigorous validation using public datasets and demonstrated superior performance in predicting both serum sodium values and corresponding intervention measures compared to existing models. This study contributes to the field of healthcare informatics by introducing an innovative, data-driven approach for dynamic serum sodium prediction and intervention recommendation, providing a valuable clinical decision-support tool for optimizing sodium management strategies in critically ill stroke patients.

PMID:40309130 | PMC:PMC12037442 | DOI:10.1007/s41666-025-00192-x

Categories: Literature Watch

Enhancing plant morphological trait identification in herbarium collections through deep learning-based segmentation

Thu, 2025-05-01 06:00

Appl Plant Sci. 2025 Feb 13;13(2):e70000. doi: 10.1002/aps3.70000. eCollection 2025 Mar-Apr.

ABSTRACT

PREMISE: Deep learning has become increasingly important in the analysis of digitized herbarium collections, which comprise millions of scans that provide valuable resources for studying plant evolution and biodiversity. However, leveraging deep learning algorithms to analyze these scans presents significant challenges, partly due to the heterogeneous nature of the non-plant material that forms the background of the scans. We hypothesize that removing such backgrounds can improve the performance of these algorithms.

METHODS: We propose a novel method based on deep learning to segment and generate plant masks from herbarium scans and subsequently remove the non-plant backgrounds. The semi-automatic preprocessing stages involve the identification and removal of non-plant elements, substantially reducing the manual effort required to prepare the training dataset.

RESULTS: The results highlight the importance of effective image segmentation, which achieved an F1 score of up to 96.6%. Moreover, when used in classification models for plant morphological trait identification, the images resulting from segmentation improved classification accuracy by up to 3% and F1 score by up to 7% compared to non-segmented images.

DISCUSSION: Our approach isolates plant elements in herbarium scans by removing background elements to improve classification tasks. We demonstrate that image segmentation significantly enhances the performance of plant morphological trait identification models.

PMID:40308899 | PMC:PMC12038731 | DOI:10.1002/aps3.70000

Categories: Literature Watch

Navigating the Multiverse: a Hitchhiker's guide to selecting harmonization methods for multimodal biomedical data

Thu, 2025-05-01 06:00

Biol Methods Protoc. 2025 Apr 17;10(1):bpaf028. doi: 10.1093/biomethods/bpaf028. eCollection 2025.

ABSTRACT

The application of machine learning (ML) techniques in predictive modelling has greatly advanced our comprehension of biological systems. There is a notable shift in the trend towards integration methods that specifically target the simultaneous analysis of multiple modes or types of data, showcasing superior results compared to individual analyses. Despite the availability of diverse ML architectures for researchers interested in embracing a multimodal approach, the current literature lacks a comprehensive taxonomy that includes the pros and cons of these methods to guide the entire process. Closing this gap is imperative, necessitating the creation of a robust framework. This framework should not only categorize the diverse ML architectures suitable for multimodal analysis but also offer insights into their respective advantages and limitations. Additionally, such a framework can serve as a valuable guide for selecting an appropriate workflow for multimodal analysis. This comprehensive taxonomy would provide a clear guidance and support informed decision-making within the progressively intricate landscape of biomedical and clinical data analysis. This is an essential step towards advancing personalized medicine. The aims of the work are to comprehensively study and describe the harmonization processes that are performed and reported in the literature and present a working guide that would enable planning and selecting an appropriate integrative model. We present harmonization as a dual process of representation and integration, each with multiple methods and categories. The taxonomy of the various representation and integration methods are classified into six broad categories and detailed with the advantages, disadvantages and examples. A guide flowchart describing the step-by-step processes that are needed to adopt a multimodal approach is also presented along with examples and references. This review provides a thorough taxonomy of methods for harmonizing multimodal data and introduces a foundational 10-step guide for newcomers to implement a multimodal workflow.

PMID:40308831 | PMC:PMC12043205 | DOI:10.1093/biomethods/bpaf028

Categories: Literature Watch

HR-NeRF: advancing realism and accuracy in highlight scene representation

Thu, 2025-05-01 06:00

Front Neurorobot. 2025 Apr 16;19:1558948. doi: 10.3389/fnbot.2025.1558948. eCollection 2025.

ABSTRACT

NeRF and its variants excel in novel view synthesis but struggle with scenes featuring specular highlights. To address this limitation, we introduce the Highlight Recovery Network (HRNet), a new architecture that enhances NeRF's ability to capture specular scenes. HRNet incorporates Swish activation functions, affine transformations, multilayer perceptrons (MLPs), and residual blocks, which collectively enable smooth non-linear transformations, adaptive feature scaling, and hierarchical feature extraction. The residual connections help mitigate the vanishing gradient problem, ensuring stable training. Despite the simplicity of HRNet's components, it achieves impressive results in recovering specular highlights. Additionally, a density voxel grid enhances model efficiency. Evaluations on four inward-facing benchmarks demonstrate that our approach outperforms NeRF and its variants, achieving a 3-5 dB PSNR improvement on each dataset while accurately capturing scene details. Furthermore, our method effectively preserves image details without requiring positional encoding, rendering a single scene in ~18 min on an NVIDIA RTX 3090 Ti GPU.

PMID:40308477 | PMC:PMC12041011 | DOI:10.3389/fnbot.2025.1558948

Categories: Literature Watch

An android-smartphone application for rice panicle detection and rice growth stage recognition using a lightweight YOLO network

Thu, 2025-05-01 06:00

Front Plant Sci. 2025 Apr 16;16:1561632. doi: 10.3389/fpls.2025.1561632. eCollection 2025.

ABSTRACT

INTRODUCTION: Detection of rice panicles and recognition of rice growth stages can significantly improve precision field management, which is crucial for maximizing grain yield. This study explores the use of deep learning on mobile phones as a platform for rice phenotype applications.

METHODS: An improved YOLOv8 model, named YOLO_Efficient Computation Optimization (YOLO_ECO), was proposed to detect rice panicles at the booting, heading, and filling stages, and to recognize growth stages. YOLO_ECO introduced key improvements, including the C2f-FasterBlock-Effective Multi-scale Attention (C2f-Faster-EMA) replacing the original C2f module in the backbone, adoption of Slim Neck to reduce neck complexity, and the use of a Lightweight Shared Convolutional Detection (LSCD) head to enhance efficiency. An Android application, YOLO-RPD, was developed to facilitate rice phenotype detection in complex field environments.

RESULTS AND DISCUSSION: The performance impact of YOLO-RPD using models with different backbone networks, quantitative models, and input image sizes was analyzed. Experimental results demonstrated that YOLO_ECO outperformed traditional deep learning models, achieving average precision values of 96.4%, 93.2%, and 81.5% at the booting, heading, and filling stages, respectively. Furthermore, YOLO_ECO exhibited advantages in detecting occlusion and small panicles, while significantly optimizing parameter count, computational demand, and model size. The YOLO_ECO FP32-1280 achieved a mean average precision (mAP) of 90.4%, with 1.8 million parameters and 4.1 billion floating-point operations (FLOPs). The YOLO-RPD application demonstrates the feasibility of deploying deep learning models on mobile devices for precision agriculture, providing rice growers with a practical, lightweight tool for real-time monitoring.

PMID:40308302 | PMC:PMC12040913 | DOI:10.3389/fpls.2025.1561632

Categories: Literature Watch

Deep learning in neurosurgery: a systematic literature review with a structured analysis of applications across subspecialties

Thu, 2025-05-01 06:00

Front Neurol. 2025 Apr 16;16:1532398. doi: 10.3389/fneur.2025.1532398. eCollection 2025.

ABSTRACT

OBJECTIVE: This study systematically reviewed deep learning (DL) applications in neurosurgical practice to provide a comprehensive understanding of DL in neurosurgery. The review process included a systematic overview of recent developments in DL technologies, an examination of the existing literature on their applications in neurosurgery, and insights into the future of neurosurgery. The study also summarized the most widely used DL algorithms, their specific applications in neurosurgical practice, their limitations, and future directions.

MATERIALS AND METHODS: An advanced search using medical subject heading terms was conducted in Medline (via PubMed), Scopus, and Embase databases restricted to articles published in English. Two independent neurosurgically experienced reviewers screened selected articles.

RESULTS: A total of 456 articles were initially retrieved. After screening, 162 were found eligible and included in the study. Reference lists of all 162 articles were checked, and 19 additional articles were found eligible and included in the study. The 181 included articles were divided into 6 categories according to the subspecialties: general neurosurgery (n = 64), neuro-oncology (n = 49), functional neurosurgery (n = 32), vascular neurosurgery (n = 17), neurotrauma (n = 9), and spine and peripheral nerve (n = 10). The leading procedures in which DL algorithms were most commonly used were deep brain stimulation and subthalamic and thalamic nuclei localization (n = 24) in the functional neurosurgery group; segmentation, identification, classification, and diagnosis of brain tumors (n = 29) in the neuro-oncology group; and neuronavigation and image-guided neurosurgery (n = 13) in the general neurosurgery group. Apart from various video and image datasets, computed tomography, magnetic resonance imaging, and ultrasonography were the most frequently used datasets to train DL algorithms in all groups overall (n = 79). Although there were few studies involving DL applications in neurosurgery in 2016, research interest began to increase in 2019 and has continued to grow in the 2020s.

CONCLUSION: DL algorithms can enhance neurosurgical practice by improving surgical workflows, real-time monitoring, diagnostic accuracy, outcome prediction, volumetric assessment, and neurosurgical education. However, their integration into neurosurgical practice involves challenges and limitations. Future studies should focus on refining DL models with a wide variety of datasets, developing effective implementation techniques, and assessing their affect on time and cost efficiency.

PMID:40308224 | PMC:PMC12040697 | DOI:10.3389/fneur.2025.1532398

Categories: Literature Watch

Selective laser cleaning of microbeads using deep learning

Wed, 2025-04-30 06:00

Sci Rep. 2025 Apr 30;15(1):15160. doi: 10.1038/s41598-025-99646-w.

ABSTRACT

Laser cleaning is widely used industrially to remove surface contaminants with high precision. Conventional methods, however, lack real-time monitoring and feedback loops, often necessitating over-machining to ensure complete contaminant removal, which leads to inefficient energy use and potential substrate damage. In this work, we demonstrate a concept of selective laser cleaning via the application of femtosecond laser pulses and polystyrene microbeads with a diameter of 15 μm. These microbeads model challenging scenarios in high-precision optical work and delicate surface treatments across laboratory and production settings. To enable adaptive, real-time cleaning, we integrated a neural network that predicts the sample's appearance after each laser pulse into a feedback loop, tailoring the cleaning process to a bespoke target pattern. This method ensures precise contaminant removal with minimal energy use, making it highly promising for applications demanding strict material control, such as wafer cleaning, sensitive surface treatments, and heritage restoration. By combining machine learning with ultrafast laser technology, our approach significantly enhances the efficiency and precision of cleaning processes.

PMID:40307358 | DOI:10.1038/s41598-025-99646-w

Categories: Literature Watch

A hybrid deep learning framework for early detection of diabetic retinopathy using retinal fundus images

Wed, 2025-04-30 06:00

Sci Rep. 2025 Apr 30;15(1):15166. doi: 10.1038/s41598-025-99309-w.

ABSTRACT

Recent advancements in deep learning have significantly impacted medical image processing domain, enabling sophisticated and accurate diagnostic tools. This paper presents a novel hybrid deep learning framework that combines convolutional neural networks (CNNs) and recurrent neural networks (RNNs) for diabetic retinopathy (DR) early detection and progression monitoring using retinal fundus images. Utilizing the sequential nature of disease progression, the proposed method integrates temporal information across multiple retinal scans to enhance detection accuracy. The proposed model utilizes publicly available DRIVE and Kaggle diabetic retinopathy datasets to evaluate the performance. The benchmark datasets provide a diverse set of annotated retinal images and the proposed hybrid model employs a CNN to extract spatial features from retinal images. The spatial feature extraction is enhanced by multi-scale feature extraction to capture fine details and broader patterns. These enriched spatial features are then fed into an RNN with attention mechanism to capture temporal dependencies so that most relevant data aspects can be considered for analysis. This combined approach enables the model to consider both current and previous states of the retina, improving its ability to detect subtle changes indicative of early-stage DR. Proposed model experimental evaluation demonstrate the superior performance over traditional deep learning models like CNN, RNN, InceptionV3, VGG19 and LSTM in terms of both sensitivity and specificity, achieving 97.5% accuracy on the DRIVE dataset, 94.04% on the Kaggle dataset, 96.9% on the Eyepacs Dataset. This research work not only advances the field of automated DR detection but also provides a framework for utilizing temporal information in medical image analysis.

PMID:40307328 | DOI:10.1038/s41598-025-99309-w

Categories: Literature Watch

A digital photography dataset for Vaccinia Virus plaque quantification using Deep Learning

Wed, 2025-04-30 06:00

Sci Data. 2025 Apr 30;12(1):719. doi: 10.1038/s41597-025-05030-8.

ABSTRACT

Virological plaque assay is the major method of detecting and quantifying infectious viruses in research and diagnostic samples. Furthermore, viral plaque phenotypes contain information about the life cycle and spreading mechanism of the virus forming them. While some modernisations have been proposed, the conventional assay typically involves manual quantification of plaque phenotypes, which is both laborious and time-consuming. Here, we present an annotated dataset of digital photographs of plaque assay plates of Vaccinia virus - a prototypic propoxvirus. We demonstrate how analysis of these plates can be performed using deep learning by training models based on the leading architecture for biomedical instance segmentation - StarDist. Finally, we show that the entire analysis can be achieved in a single step by HydraStarDist - the modified architecture we propose.

PMID:40307255 | DOI:10.1038/s41597-025-05030-8

Categories: Literature Watch

Effects of Deep Learning-Based Reconstruction on the Quality of Accelerated Contrast-Enhanced Neck MRI

Wed, 2025-04-30 06:00

Korean J Radiol. 2025 May;26(5):446-445. doi: 10.3348/kjr.2024.1059.

ABSTRACT

OBJECTIVE: To compare the quality of deep learning-reconstructed turbo spin-echo (DL-TSE) and conventionally interpolated turbo spin-echo (Conv-TSE) techniques in contrast-enhanced MRI of the neck.

MATERIALS AND METHODS: Contrast-enhanced T1-weighted DL-TSE and Conv-TSE images were acquired using 3T scanners from 106 patients. DL-TSE employed a closed-source, 'work-in-progress' (WIP No. 1062, iTSE, version 10; Siemens Healthineers) algorithm for interpolation and denoising to achieve the same in-plane resolution (axial: 0.26 × 0.26 mm²; coronal: 0.29 × 0.29 mm²) while reducing scan times by 15.9% and 52.6% for axial and coronal scans, respectively. The full width at half maximum (FWHM) and percent signal ghosting were measured using stationary and flow phantom scans, respectively. In patient images, non-uniformity (NU), contrast-to-noise ratio (CNR), and regional mucosal FWHM were evaluated. Two neuroradiologists visually rated the patient images for overall quality, sharpness, regional mucosal conspicuity, artifacts, and lesions using a 5-point Likert scale.

RESULTS: FWHM in the stationary phantom scan was consistently sharper in DL-TSE. The percent signal ghosting outside the flow phantom was lower in DL-TSE (0.06% vs. 0.14%) but higher within the phantom (8.92% vs. 1.75%) compared to Conv-TSE. In patient scans, DL-TSE showed non-inferior NU and higher CNR. Regional mucosal FWHM was significantly better in DL-TSE, particularly in the oropharynx (coronal: 1.08 ± 0.31 vs. 1.52 ± 0.46 mm) and hypopharynx (coronal: 1.26 ± 0.35 vs. 1.91 ± 0.56 mm) (both P < 0.001). DL-TSE demonstrated higher overall image quality (axial: 4.61 ± 0.49 vs. 3.32 ± 0.54) and sharpness (axial: 4.40 ± 0.56 vs. 3.11 ± 0.53) (both P < 0.001). In addition, mucosal conspicuity was improved, especially in the oropharynx (axial: 4.41 ± 0.67 vs. 3.40 ± 0.69) and hypopharynx (axial: 4.45 ± 0.58 vs. 3.58 ± 0.63) (both P < 0.001). Extracorporeal ghost artifacts were reduced in DL-TSE (axial: 4.32 ± 0.60 vs. 3.90 ± 0.71, P < 0.001) but artifacts overlapping anatomical structures were slightly more pronounced (axial: 3.78 ± 0.74 vs. 3.95 ± 0.72, P < 0.001). Lesions were detected with higher confidence in DL-TSE.

CONCLUSION: DL-based reconstruction applied to accelerated neck MRI improves overall image quality, sharpness, mucosal conspicuity in motion-prone regions, and lesion detection confidence. Despite more pronounced ghost artifacts overlapping anatomical structures, DL-TSE enables substantial scan time reduction while enhancing diagnostic performance.

PMID:40307199 | DOI:10.3348/kjr.2024.1059

Categories: Literature Watch

Pages