Deep learning
AVP-GPT2: A Transformer-Powered Platform for De Novo Generation, Screening, and Explanation of Antiviral Peptides
Viruses. 2024 Dec 25;17(1):14. doi: 10.3390/v17010014.
ABSTRACT
Human respiratory syncytial virus (RSV) remains a significant global health threat, particularly for vulnerable populations. Despite extensive research, effective antiviral therapies are still limited. To address this urgent need, we present AVP-GPT2, a deep-learning model that significantly outperforms its predecessor, AVP-GPT, in designing and screening antiviral peptides. Trained on a significantly expanded dataset, AVP-GPT2 employs a transformer-based architecture to generate diverse peptide sequences. A multi-modal screening approach, incorporating Star-Transformer and Vision Transformer, enables accurate prediction of antiviral activity and toxicity, leading to the identification of potent and safe candidates. SHAP analysis further enhances interpretability by explaining the underlying mechanisms of peptide activity. Our in vitro experiments confirmed the antiviral efficacy of peptides generated by AVP-GPT2, with some exhibiting EC50 values as low as 0.01 μM and CC50 values > 30 μM. This represents a substantial improvement over AVP-GPT and traditional methods. AVP-GPT2 has the potential to significantly impact antiviral drug discovery by accelerating the identification of novel therapeutic agents. Future research will explore its application to other viral targets and its integration into existing drug development pipelines.
PMID:39861804 | DOI:10.3390/v17010014
A Feature-Enhanced Small Object Detection Algorithm Based on Attention Mechanism
Sensors (Basel). 2025 Jan 20;25(2):589. doi: 10.3390/s25020589.
ABSTRACT
With the rapid development of AI algorithms and computational power, object recognition based on deep learning frameworks has become a major research direction in computer vision. UAVs equipped with object detection systems are increasingly used in fields like smart transportation, disaster warning, and emergency rescue. However, due to factors such as the environment, lighting, altitude, and angle, UAV images face challenges like small object sizes, high object density, and significant background interference, making object detection tasks difficult. To address these issues, we use YOLOv8s as the basic framework and introduce a multi-level feature fusion algorithm. Additionally, we design an attention mechanism that links distant pixels to improve small object feature extraction. To address missed detections and inaccurate localization, we replace the detection head with a dynamic head, allowing the model to route objects to the appropriate head for final output. We also introduce Slideloss to improve the model's learning of difficult samples and ShapeIoU to better account for the shape and scale of bounding boxes. Experiments on datasets like VisDrone2019 show that our method improves accuracy by nearly 10% and recall by about 11% compared to the baseline. Additionally, on the AI-TODv1.5 dataset, our method improves the mAP50 from 38.8 to 45.2.
PMID:39860960 | DOI:10.3390/s25020589
Intelligent Intrusion Detection System Against Various Attacks Based on a Hybrid Deep Learning Algorithm
Sensors (Basel). 2025 Jan 20;25(2):580. doi: 10.3390/s25020580.
ABSTRACT
The Internet of Things (IoT) has emerged as a crucial element in everyday life. The IoT environment is currently facing significant security concerns due to the numerous problems related to its architecture and supporting technology. In order to guarantee the complete security of the IoT, it is important to deal with these challenges. This study centers on employing deep learning methodologies to detect attacks. In general, this research aims to improve the performance of existing deep learning models. To mitigate data imbalances and enhance learning outcomes, the synthetic minority over-sampling technique (SMOTE) is employed. Our approach contributes to a multistage feature extraction process where autoencoders (AEs) are used initially to extract robust features from unstructured data on the model architecture's left side. Following this, long short-term memory (LSTM) networks on the right analyze these features to recognize temporal patterns indicative of abnormal behavior. The extracted and temporally refined features are inputted into convolutional neural networks (CNNs) for final classification. This structured arrangement harnesses the distinct capabilities of each model to process and classify IoT security data effectively. Our framework is specifically designed to address various attacks, including denial of service (DoS) and Mirai attacks, which are particularly harmful to IoT systems. Unlike conventional intrusion detection systems (IDSs) that may employ a singular model or simple feature extraction methods, our multistage approach provides more comprehensive analysis and utilization of data, enhancing detection capabilities and accuracy in identifying complex cyber threats in IoT environments. This research highlights the potential benefits that can be gained by applying deep learning methods to improve the effectiveness of IDSs in IoT security. The results obtained indicate a potential improvement for enhancing security measures and mitigating emerging threats.
PMID:39860948 | DOI:10.3390/s25020580
D-MGDCN-CLSTM: A Traffic Prediction Model Based on Multi-Graph Gated Convolution and Convolutional Long-Short-Term Memory
Sensors (Basel). 2025 Jan 19;25(2):561. doi: 10.3390/s25020561.
ABSTRACT
Real-time and accurate traffic forecasting aids in traffic planning and design and helps to alleviate congestion. Addressing the negative impacts of partial data loss in traffic forecasting, and the challenge of simultaneously capturing short-term fluctuations and long-term trends, this paper presents a traffic forecasting model, D-MGDCN-CLSTM, based on Multi-Graph Gated Dilated Convolution and Conv-LSTM. The model uses the DTWN algorithm to fill in missing data. To better capture the dual characteristics of short-term fluctuations and long-term trends in traffic, the model employs the DWT for multi-scale decomposition to obtain approximation and detail coefficients. The Conv-LSTM processes the approximation coefficients to capture the long-term characteristics of the time series, while the multiple layers of the MGDCN process the detail coefficients to capture short-term fluctuations. The outputs of the two branches are then merged to produce the forecast results. The model is tested against 10 algorithms using the PeMSD7(M) and PeMSD7(L) datasets, improving MAE, RMSE, and ACC by an average of 1.38% and 13.89%, 1% and 1.24%, and 5.92% and 1%, respectively. Ablation experiments, parameter impact analysis, and visual analysis all demonstrate the superiority of our decompositional approach in handling the dual characteristics of traffic data.
PMID:39860932 | DOI:10.3390/s25020561
A Deep Learning Approach for Mental Fatigue State Assessment
Sensors (Basel). 2025 Jan 19;25(2):555. doi: 10.3390/s25020555.
ABSTRACT
This study investigates mental fatigue in sports activities by leveraging deep learning techniques, deviating from the conventional use of heart rate variability (HRV) feature analysis found in previous research. The study utilizes a hybrid deep neural network model, which integrates Residual Networks (ResNet) and Bidirectional Long Short-Term Memory (Bi-LSTM) for feature extraction, and a transformer for feature fusion. The model achieves an impressive accuracy of 95.29% in identifying fatigue from original ECG data, 2D spectral characteristics and physiological information of subjects. In comparison to traditional methods, such as Support Vector Machines (SVMs) and Random Forests (RFs), as well as other deep learning methods, including Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM), the proposed approach demonstrates significantly improved experimental outcomes. Overall, this study offers a promising solution for accurately recognizing fatigue through the analysis of physiological signals, with potential applications in sports and physical fitness training contexts.
PMID:39860925 | DOI:10.3390/s25020555
Remaining Useful Life Prediction of Rolling Bearings Based on CBAM-CNN-LSTM
Sensors (Basel). 2025 Jan 19;25(2):554. doi: 10.3390/s25020554.
ABSTRACT
Predicting the Remaining Useful Life (RUL) is vital for ensuring the reliability and safety of equipment and components. This study introduces a novel method for predicting RUL that utilizes the Convolutional Block Attention Module (CBAM) to address the problem that Convolutional Neural Networks (CNNs) do not effectively leverage data channel features and spatial features in residual life prediction. Firstly, Fast Fourier Transform (FFT) is applied to convert the data into the frequency domain. The resulting frequency domain data is then used as input to the convolutional neural network for feature extraction; Then, the weights of channel features and spatial features are assigned to the extracted features by CBAM, and the weighted features are then input into the Long Short-Term Memory (LSTM) network to learn temporal features. Finally, the effectiveness of the proposed model is verified using the PHM2012 bearing dataset. Compared to several existing RUL prediction methods, the mean squared error, mean absolute error, and root mean squared error of the proposed method in this paper are reduced by 53%, 16.87%, and 31.68%, respectively, which verifies the superiority of the method. Meanwhile, the experimental results demonstrate that the proposed method achieves good RUL prediction accuracy across various failure modes.
PMID:39860924 | DOI:10.3390/s25020554
TBF-YOLOv8n: A Lightweight Tea Bud Detection Model Based on YOLOv8n Improvements
Sensors (Basel). 2025 Jan 18;25(2):547. doi: 10.3390/s25020547.
ABSTRACT
Tea bud localization detection not only ensures tea quality, improves picking efficiency, and advances intelligent harvesting, but also fosters tea industry upgrades and enhances economic benefits. To solve the problem of the high computational complexity of deep learning detection models, we developed the Tea Bud DSCF-YOLOv8n (TBF-YOLOv8n)lightweight detection model. Improvement of the Cross Stage Partial Bottleneck Module with Two Convolutions(C2f) module via efficient Distributed Shift Convolution (DSConv) yields the C2f module with DSConv(DSCf)module, which reduces the model's size. Additionally, the coordinate attention (CA) mechanism is incorporated to mitigate interference from irrelevant factors, thereby improving mean accuracy. Furthermore, the SIOU_Loss (SCYLLA-IOU_Loss) function and the Dynamic Sample(DySample)up-sampling operator are implemented to accelerate convergence and enhance both average precision and detection accuracy. The experimental results show that compared to the YOLOv8n model, the TBF-YOLOv8n model has a 3.7% increase in accuracy, a 1.1% increase in average accuracy, a 44.4% reduction in gigabit floating point operations (GFLOPs), and a 13.4% reduction in the total number of parameters included in the model. In comparison experiments with a variety of lightweight detection models, the TBF-YOLOv8n still performs well in terms of detection accuracy while remaining more lightweight. In conclusion, the TBF-YOLOv8n model achieves a commendable balance between efficiency and precision, offering valuable insights for advancing intelligent tea bud harvesting technologies.
PMID:39860916 | DOI:10.3390/s25020547
Zero-Shot Traffic Identification with Attribute and Graph-Based Representations for Edge Computing
Sensors (Basel). 2025 Jan 18;25(2):545. doi: 10.3390/s25020545.
ABSTRACT
With the proliferation of mobile terminals and the rapid growth of network applications, fine-grained traffic identification has become increasingly challenging. Methods based on machine learning and deep learning have achieved remarkable results, but they heavily rely on the distribution of training data, which makes them ineffective in handling unseen samples. In this paper, we propose AG-ZSL, a zero-shot learning framework based on traffic behavior and attribute representations for general encrypted traffic classification. AG-ZSL primarily learns two mapping functions: one that captures traffic behavior embeddings from burst-based traffic interaction graphs, and the other that learns attribute embeddings from traffic attribute descriptions. Then, the framework minimizes the distance between these embeddings within the shared feature space. The gradient rejection algorithm and K-Nearest Neighbors are introduced to implement a two-stage method for general traffic classification. Experimental results on IoT datasets demonstrate that AG-ZSL achieves exceptional performance in classifying both known and unknown traffic, highlighting its potential for enhancing secure and efficient traffic management at the network edge.
PMID:39860913 | DOI:10.3390/s25020545
PC-CS-YOLO: High-Precision Obstacle Detection for Visually Impaired Safety
Sensors (Basel). 2025 Jan 17;25(2):534. doi: 10.3390/s25020534.
ABSTRACT
The issue of obstacle avoidance and safety for visually impaired individuals has been a major topic of research. However, complex street environments still pose significant challenges for blind obstacle detection systems. Existing solutions often fail to provide real-time, accurate obstacle avoidance decisions. In this study, we propose a blind obstacle detection system based on the PC-CS-YOLO model. The system improves the backbone network by adopting the partial convolutional feed-forward network (PCFN) to reduce computational redundancy. Additionally, to enhance the network's robustness in multi-scale feature fusion, we introduce the Cross-Scale Attention Fusion (CSAF) mechanism, which integrates features from different sensory domains to achieve superior performance. Compared to state-of-the-art networks, our system shows improvements of 2.0%, 3.9%, and 1.5% in precision, recall, and mAP50, respectively. When evaluated on a GPU, the inference speed is 20.6 ms, which is 15.3 ms faster than YOLO11, meeting the real-time requirements for blind obstacle avoidance systems.
PMID:39860905 | DOI:10.3390/s25020534
Automated detection of traumatic bleeding in CT images using 3D U-Net# and multi-organ segmentation
Biomed Phys Eng Express. 2025 Jan 24. doi: 10.1088/2057-1976/adae14. Online ahead of print.
ABSTRACT
Traumatic injury remains a leading cause of death worldwide, with traumatic bleeding being one of its most critical and fatal consequences. The use of whole-body computed tomography (WBCT) in trauma management has rapidly expanded. However, interpreting WBCT images within the limited time available before treatment is particularly challenging for acute care physicians. Our group has previously developed an automated bleeding detection method in WBCT images. However, further reduction of false positives (FPs) is necessary for clinical application. To address this issue, we propose a novel automated detection for traumatic bleeding in CT images using deep learning and multi-organ segmentation; Methods: The proposed method integrates a three-dimensional U-Net# model for bleeding detection with an FP reduction approach based on multi-organ segmentation. The multi-organ segmentation method targets the bone, kidney, and vascular regions, where FPs are primarily found during the bleeding detection process. We evaluated the proposed method using a dataset of delayed-phase contrast-enhanced trauma CT images collected from four institutions; Results: Our method detected 70.0% of bleedings with 76.2 FPs/case. The processing time for our method was 6.3 ± 1.4 min. Compared with our previous ap-proach, the proposed method significantly reduced the number of FPs while maintaining detection sensitivity.
PMID:39854772 | DOI:10.1088/2057-1976/adae14
Pedestrian POSE estimation using multi-branched deep learning pose net
PLoS One. 2025 Jan 24;20(1):e0312177. doi: 10.1371/journal.pone.0312177. eCollection 2025.
ABSTRACT
In human activity-recognition scenarios, including head and entire body pose and orientations, recognizing the pose and direction of a pedestrian is considered a complex problem. A person may be traveling in one sideway while focusing his attention on another side. It is occasionally desirable to analyze such orientation estimates using computer-vision tools for automated analysis of pedestrian behavior and intention. This article uses a deep-learning method to demonstrate the pedestrian full-body pose estimation approach. A deep-learning-based pre-trained supervised model multi-branched deep learning pose net (MBDLP-Net) is proposed for estimation and classification. For full-body pose and orientation estimation, three independent datasets, an extensive dataset for body orientation (BDBO), PKU-Reid, and TUD Multiview Pedestrians, are used. Independently, the proposed technique is trained on dataset CIFAR-100 with 100 classes. The proposed approach is meticulously tested using publicly accessible BDBO, PKU-Reid, and TUD datasets. The results show that the mean accuracy for full-body pose estimation with BDBO and PKU-Reid is 0.95%, and with TUD multiview pedestrians is 0.97%. The performance results show that the proposed technique efficiently distinguishes full-body poses and orientations in various configurations. The efficacy of the provided approach is compared with existing pretrained, robust, and state-of-the-art methodologies, providing a comprehensive understanding of its advantages.
PMID:39854382 | DOI:10.1371/journal.pone.0312177
Maize quality detection based on MConv-SwinT high-precision model
PLoS One. 2025 Jan 24;20(1):e0312363. doi: 10.1371/journal.pone.0312363. eCollection 2025.
ABSTRACT
The traditional method of corn quality detection relies heavily on the subjective judgment of inspectors and suffers from a high error rate. To address these issues, this study employs the Swin Transformer as an enhanced base model, integrating machine vision and deep learning techniques for corn quality assessment. Initially, images of high-quality, moldy, and broken corn were collected. After preprocessing, a total of 20,152 valid images were obtained for the experimental samples. The network then extracts both shallow and deep features from these maize images, which are subsequently fused. Concurrently, the extracted features undergo further processing through a specially designed convolutional block. The fused features, combined with those processed by the convolutional module, are fed into an attention layer. This attention layer assigns weights to the features, facilitating accurate final classification. Experimental results demonstrate that the MC-Swin Transformer model proposed in this paper significantly outperforms traditional convolutional neural network models in key metrics such as accuracy, precision, recall, and F1 score, achieving a recognition accuracy rate of 99.89%. Thus, the network effectively and efficiently classifies different corn qualities. This study not only offers a novel perspective and technical approach to corn quality detection but also holds significant implications for the advancement of smart agriculture.
PMID:39854315 | DOI:10.1371/journal.pone.0312363
Functional profiling of the sequence stockpile: a protein pair-based assessment of in silico prediction tools
Bioinformatics. 2025 Jan 24:btaf035. doi: 10.1093/bioinformatics/btaf035. Online ahead of print.
ABSTRACT
MOTIVATION: In silico functional annotation of proteins is crucial to narrowing the sequencing-accelerated gap in our understanding of protein activities. Numerous function annotation methods exist, and their ranks have been growing, particularly so with the recent deep learning-based developments. However, it is unclear if these tools are truly predictive. As we are not aware of any methods that can identify new terms in functional ontologies, we ask if they can, at least, identify molecular functions of proteins that are non-homologous to or far-removed from known protein families.
RESULTS: Here, we explore the potential and limitations of the existing methods in predicting molecular functions of thousands of such proteins. Lacking the "ground truth" functional annotations, we transformed the assessment of function prediction into evaluation of functional similarity of protein pairs that likely share function but are unlike any of the currently functionally annotated sequences. Notably, our approach transcends the limitations of functional annotation vocabularies, providing a means to assess different-ontology annotation methods. We find that most existing methods are limited to identifying functional similarity of homologous sequences and fail to predict function of proteins lacking reference. Curiously, despite their seemingly unlimited by-homology scope, deep learning methods also have trouble capturing the functional signal encoded in protein sequence. We believe that our work will inspire the development of a new generation of methods that push boundaries and promote exploration and discovery in the molecular function domain.
SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
PMID:39854283 | DOI:10.1093/bioinformatics/btaf035
Artificial Intelligence for Optical Coherence Tomography in Glaucoma
Transl Vis Sci Technol. 2025 Jan 2;14(1):27. doi: 10.1167/tvst.14.1.27.
ABSTRACT
PURPOSE: The integration of artificial intelligence (AI), particularly deep learning (DL), with optical coherence tomography (OCT) offers significant opportunities in the diagnosis and management of glaucoma. This article explores the application of various DL models in enhancing OCT capabilities and addresses the challenges associated with their clinical implementation.
METHODS: A review of articles utilizing DL models was conducted, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), autoencoders, and large language models (LLMs). Key developments and practical applications of these models in OCT image analysis were emphasized, particularly in the context of enhancing image quality, glaucoma diagnosis, and monitoring progression.
RESULTS: CNNs excel in segmenting retinal layers and detecting glaucomatous damage, whereas RNNs are effective in analyzing sequential OCT scans for disease progression. GANs enhance image quality and data augmentation, and autoencoders facilitate advanced feature extraction. LLMs show promise in integrating textual and visual data for comprehensive diagnostic assessments. Despite these advancements, challenges such as data availability, variability, potential biases, and the need for extensive validation persist.
CONCLUSIONS: DL models are reshaping glaucoma management by enhancing OCT's diagnostic capabilities. However, the successful translation into clinical practice requires addressing major challenges related to data variability, biases, fairness, and model validation to ensure accurate and reliable patient care.
TRANSLATIONAL RELEVANCE: This review bridges the gap between basic research and clinical care by demonstrating how AI, particularly DL models, can markedly enhance OCT's clinical utility in diagnosis, monitoring, and prediction, moving toward more individualized, personalized, and precise treatment strategies.
PMID:39854198 | DOI:10.1167/tvst.14.1.27
Risk score stratification of cutaneous melanoma patients based on whole slide images analysis by deep learning
J Eur Acad Dermatol Venereol. 2025 Jan 24. doi: 10.1111/jdv.20538. Online ahead of print.
ABSTRACT
BACKGROUND: There is a need to improve risk stratification of primary cutaneous melanomas to better guide adjuvant therapy. Taking into account that haematoxylin and eosin (HE)-stained tumour tissue contains a huge amount of clinically unexploited morphological informations, we developed a weakly-supervised deep-learning approach, SmartProg-MEL, to predict survival outcomes in stages I to III melanoma patients from HE-stained whole slide image (WSI).
METHODS: We designed a deep neural network that extracts morphological features from WSI to predict 5-y overall survival (OS), and assign a survival risk score to each patient. The model was trained and validated on a discovery cohort of primary cutaneous melanomas (IHP-MEL-1, n = 342). Performance was tested on two external and independent datasets (IHP-MEL-2, n = 161; and TCGA cohort n = 63). It was compared with well-established prognostic factors. Concordance index (c-index) was used as a metric.
RESULTS: On the discovery cohort, the SmartProg-MEL predicts the 5-y OS with a c-index of 0.78 on the cross-validation data and of 0.72 on the cross-testing series. In the external cohorts, the model achieved a c-index of 0.71 and 0.69 for the IHP-MEL-2 and TCGA dataset respectively. Furthermore, SmartProg-MEL was an independent and the most powerful prognostic factor in multivariate analysis (HR = 1.84, p-value < 0.005). Finally, the model was able to dichotomize patients in two groups-a low and a high-risk group-each associated with a significantly different 5-y OS (p-value < 0.001 for IHP-MEL-1 and p-value = 0.01 for IHP-MEL-2).
CONCLUSION: The performance of our fully automated SmartProg-MEL model outperforms the current clinicopathological factors in terms of prediction of 5-y OS and risk stratification of cutaneous melanoma patients. Incorporation of SmartProg-MEL in the clinical workflow could guide the decision-making process by improving the identification of patients that may benefit from adjuvant therapy.
PMID:39853986 | DOI:10.1111/jdv.20538
Deep-Learning Generated Synthetic Material Decomposition Images Based on Single-Energy CT to Differentiate Intracranial Hemorrhage and Contrast Staining Within 24 Hours After Endovascular Thrombectomy
CNS Neurosci Ther. 2025 Jan;31(1):e70235. doi: 10.1111/cns.70235.
ABSTRACT
AIMS: To develop a transformer-based generative adversarial network (trans-GAN) that can generate synthetic material decomposition images from single-energy CT (SECT) for real-time detection of intracranial hemorrhage (ICH) after endovascular thrombectomy.
MATERIALS: We retrospectively collected data from two hospitals, consisting of 237 dual-energy CT (DECT) scans, including matched iodine overlay maps, virtual noncontrast, and simulated SECT images. These scans were randomly divided into a training set (n = 190) and an internal validation set (n = 47) in a 4:1 ratio based on the proportion of ICH. Additionally, 26 SECT scans were included as an external validation set. We compared our trans-GAN with state-of-the-art generation methods using several physical metrics of the generated images and evaluated the diagnostic efficacy of the generated images for differentiating ICH from contrast staining.
RESULTS: In comparison with other generation methods, the images generated by trans-GAN exhibited superior quantitative performance. Meanwhile, in terms of ICH detection, the use of generated images from both the internal and external validation sets resulted in a higher area under the receiver operating characteristic curve (0.88 vs. 0.68 and 0.69 vs. 0.54, respectively) and kappa values (0.83 vs. 0.56 and 0.51 vs. 0.31, respectively) compared with input SECT images.
CONCLUSION: Our proposed trans-GAN provides a new approach based on SECT for real-time differentiation of ICH and contrast staining in hospitals without DECT conditions.
PMID:39853936 | DOI:10.1111/cns.70235
Prediction of facial nerve outcomes after surgery for vestibular schwannoma using machine learning-based models: a systematic review and meta-analysis
Neurosurg Rev. 2025 Jan 24;48(1):79. doi: 10.1007/s10143-025-03230-9.
ABSTRACT
Postoperative facial nerve (FN) dysfunction is associated with a significant impact on the quality of life of patients and can result in psychological stress and disorders such as depression and social isolation. Preoperative prediction of FN outcomes can play a critical role in vestibular schwannomas (VSs) patient care. Several studies have developed machine learning (ML)-based models in predicting FN outcomes following resection of VS. This systematic review and meta-analysis aimed to evaluate the diagnostic accuracy of ML-based models in predicting FN outcomes following resection in the setting of VS. On December 12, 2024, the four electronic databases, Pubmed, Embase, Scopus, and Web of Science, were systematically searched. Studies that evaluated the performance outcomes of the ML-based predictive models were included. The pooled sensitivity, specificity, area under the curve (AUC), and diagnostic odds ratio (DOR) were calculated through the R program. Five studies with 807 individuals with VS, encompassing 35 models, were included. The meta-analysis showed a pooled sensitivity of 82% (95%CI: 76-87%), specificity of 79% (95%CI: 74-84%), and DOR of 12.94 (95%CI: 8.65-19.34) with an AUC of 0.841. The meta-analysis of the best performance model demonstrated a pooled sensitivity of 91% (95%CI: 80-96%), specificity of 87% (95%CI: 82-91%), and DOR of 46.84 (95%CI: 19.8-110.8). Additionally, the analysis demonstrated an AUC of 0.92, a sensitivity of 0.884, and a false positive rate of 0.136 for the best performance models. ML-based models possess promising diagnostic accuracy in predicting FN outcomes following resection.
PMID:39853510 | DOI:10.1007/s10143-025-03230-9
Scanner-based real-time three-dimensional brain + body slice-to-volume reconstruction for T2-weighted 0.55-T low-field fetal magnetic resonance imaging
Pediatr Radiol. 2025 Jan 24. doi: 10.1007/s00247-025-06165-x. Online ahead of print.
ABSTRACT
BACKGROUND: Motion correction methods based on slice-to-volume registration (SVR) for fetal magnetic resonance imaging (MRI) allow reconstruction of three-dimensional (3-D) isotropic images of the fetal brain and body. However, all existing SVR methods are confined to research settings, which limits clinical integration. Furthermore, there have been no reported SVR solutions for low-field 0.55-T MRI.
OBJECTIVE: Integration of automated SVR motion correction methods directly into fetal MRI scanning process via the Gadgetron framework to enable automated T2-weighted (T2W) 3-D fetal brain and body reconstruction in the low-field 0.55-T MRI scanner within the duration of the scan.
MATERIALS AND METHODS: A deep learning fully automated pipeline was developed for T2W 3-D rigid and deformable (D/SVR) reconstruction of the fetal brain and body of 0.55-T T2W datasets. Next, it was integrated into 0.55-T low-field MRI scanner environment via a Gadgetron workflow that enables launching of the reconstruction process directly during scanning in real-time.
RESULTS: During prospective testing on 12 cases (22-40 weeks gestational age), the fetal brain and body reconstructions were available on average 6:42 ± 3:13 min after the acquisition of the final stack and could be assessed and archived on the scanner console during the ongoing fetal MRI scan. The output image data quality was rated as good to acceptable for interpretation. The retrospective testing of the pipeline on 83 0.55-T datasets demonstrated stable reconstruction quality for low-field MRI.
CONCLUSION: The proposed pipeline allows scanner-based prospective T2W 3-D motion correction for low-field 0.55-T fetal MRI via direct online integration into the scanner environment.
PMID:39853394 | DOI:10.1007/s00247-025-06165-x
Deep-Learning-Assisted Digital Fluorescence Immunoassay on Magnetic Beads for Ultrasensitive Determination of Protein Biomarkers
Anal Chem. 2025 Jan 24. doi: 10.1021/acs.analchem.4c05877. Online ahead of print.
ABSTRACT
Digital fluorescence immunoassay (DFI) based on random dispersion magnetic beads (MBs) is one of the powerful methods for ultrasensitive determination of protein biomarkers. However, in the DFI, improving the limit of detection (LOD) is challenging since the ratio of signal-to-background and the speed of manual counting beads are low. Herein, we developed a deep-learning network (ATTBeadNet) by utilizing a new hybrid attention mechanism within a UNet3+ framework for accurately and fast counting the MBs and proposed a DFI using CdS quantum dots (QDs) with narrow peak and optical stability as reported at first time. The developed ATTBeadNet was applied to counting the MBs, resulting in the F1 score (95.91%) being higher than those of other methods (ImageJ, 68.33%; computer vision-based, 92.99%; fully convolutional network, 75.00%; mask region-based convolutional neural network, 70.34%). On principle-on-proof, a sandwich MB-based DFI was proposed, in which human interleukin-6 (IL-6) was taken as a model protein biomarker, while antibody-bound streptavidin-coated MBs were used as capture MBs and antibody-HRP-tyramide-functionalized CdS QDs were used as the binding reporter. When the developed ATTBeadNet was applied to the MB-based DFI of IL-6 (20 μL), the linear range from 5 to 100 fM and an LOD of 3.1 fM were achieved, which are better than those using the ImageJ method (linear range from 30 to 100 fM and LOD of 20 fM). This work demonstrates that the integration of the deep-learning network with DFI is a promising strategy for the highly sensitive and accurate determination of protein biomarkers.
PMID:39853309 | DOI:10.1021/acs.analchem.4c05877
Deep learning-based design and experimental validation of a medicine-like human antibody library
Brief Bioinform. 2024 Nov 22;26(1):bbaf023. doi: 10.1093/bib/bbaf023.
ABSTRACT
Antibody generation requires the use of one or more time-consuming methods, namely animal immunization, and in vitro display technologies. However, the recent availability of large amounts of antibody sequence and structural data in the public domain along with the advent of generative deep learning algorithms raises the possibility of computationally generating novel antibody sequences with desirable developability attributes. Here, we describe a deep learning model for computationally generating libraries of highly human antibody variable regions whose intrinsic physicochemical properties resemble those of the variable regions of the marketed antibody-based biotherapeutics (medicine-likeness). We generated 100000 variable region sequences of antigen-agnostic human antibodies belonging to the IGHV3-IGKV1 germline pair using a training dataset of 31416 human antibodies that satisfied our computational developability criteria. The in-silico generated antibodies recapitulate intrinsic sequence, structural, and physicochemical properties of the training antibodies, and compare favorably with the experimentally measured biophysical attributes of 100 variable regions of marketed and clinical stage antibody-based biotherapeutics. A sample of 51 highly diverse in-silico generated antibodies with >90th percentile medicine-likeness and > 90% humanness was evaluated by two independent experimental laboratories. Our data show the in-silico generated sequences exhibit high expression, monomer content, and thermal stability along with low hydrophobicity, self-association, and non-specific binding when produced as full-length monoclonal antibodies. The ability to computationally generate developable human antibody libraries is a first step towards enabling in-silico discovery of antibody-based biotherapeutics. These findings are expected to accelerate in-silico discovery of antibody-based biotherapeutics and expand the druggable antigen space to include targets refractory to conventional antibody discovery methods requiring in vitro antigen production.
PMID:39851074 | DOI:10.1093/bib/bbaf023