Deep learning
Multi-dimensional perceptual recognition of tourist destination using deep learning model and geographic information system
PLoS One. 2025 Feb 7;20(2):e0318846. doi: 10.1371/journal.pone.0318846. eCollection 2025.
ABSTRACT
Perceptual recognition of tourist destinations is vital in representing the destination image, supporting destination management decision-making, and promoting tourism recommendations. However, previous studies on tourist destination perception have limitations regarding accuracy and completeness related to research methods. This study addresses these limitations by proposing an efficient strategy to achieve precise perceptual recognition of tourist destinations while ensuring the integrity of user-generated content (UGC) data and the completeness of perception dimensions. We integrated various types of UGC data, including images, texts, and spatiotemporal information, to create a comprehensive UGC dataset. Then, we adopted the improved Inception V3 model, the bidirectional long short-term memory network (BiLSTM) model with multi-head attention, and geographic information system (GIS) technology to recognize basic tourist feature information from the UGC dataset, such as the content, sentiment, and spatiotemporal perceptual dimensions of the data, achieving a recognition accuracy of over 97%. Finally, a progressive dimension combination method was proposed to visualize and analyze multiple perceptions. An experimental case study demonstrated the strategy's effectiveness, focusing on tourists' perceptions of Datong, China. Experimental results show that the approach is feasible for studying tourist destination perception. Content perception, sentiment perception, and the perception of Datong's spatial and temporal characteristics were recognized and analyzed efficiently. This study offers valuable guidance and a reference framework for selecting methods and technical routes in tourist destination perception.
PMID:39919101 | DOI:10.1371/journal.pone.0318846
scCamAge: A context-aware prediction engine for cellular age, aging-associated bioactivities, and morphometrics
Cell Rep. 2025 Feb 6;44(2):115270. doi: 10.1016/j.celrep.2025.115270. Online ahead of print.
ABSTRACT
Current deep-learning-based image-analysis solutions exhibit limitations in holistically capturing spatiotemporal cellular changes, particularly during aging. We present scCamAge, an advanced context-aware multimodal prediction engine that co-leverages image-based cellular spatiotemporal features at single-cell resolution alongside cellular morphometrics and aging-associated bioactivities such as genomic instability, mitochondrial dysfunction, vacuolar dynamics, reactive oxygen species levels, and epigenetic and proteasomal dysfunctions. scCamAge employed heterogeneous datasets comprising ∼1 million single yeast cells and was validated using pro-longevity drugs, genetic mutants, and stress-induced models. scCamAge also predicted a pro-longevity response in yeast cells under iterative thermal stress, confirmed using integrative omics analyses. Interestingly, scCamAge, trained solely on yeast images, without additional learning, surpasses generic models in predicting chemical and replication-induced senescence in human fibroblasts, indicating evolutionary conservation of aging-related morphometrics. Finally, we enhanced the generalizability of scCamAge by retraining it on human fibroblast senescence datasets, which improved its ability to predict senescent cells.
PMID:39918957 | DOI:10.1016/j.celrep.2025.115270
A novel deep learning framework for retinal disease detection leveraging contextual and local features cues from retinal images
Med Biol Eng Comput. 2025 Feb 7. doi: 10.1007/s11517-025-03314-0. Online ahead of print.
ABSTRACT
Retinal diseases are a serious global threat to human vision, and early identification is essential for effective prevention and treatment. However, current diagnostic methods rely on manual analysis of fundus images, which heavily depends on the expertise of ophthalmologists. This manual process is time-consuming and labor-intensive and can sometimes lead to missed diagnoses. With advancements in computer vision technology, several automated models have been proposed to improve diagnostic accuracy for retinal diseases and medical imaging in general. However, these methods face challenges in accurately detecting specific diseases within images due to inherent issues associated with fundus images, including inter-class similarities, intra-class variations, limited local information, insufficient contextual understanding, and class imbalances within datasets. To address these challenges, we propose a novel deep learning framework for accurate retinal disease classification. This framework is designed to achieve high accuracy in identifying various retinal diseases while overcoming inherent challenges associated with fundus images. Generally, the framework consists of three main modules. The first module is Densely Connected Multidilated Convolution Neural Network (DCM-CNN) that extracts global contextual information by effectively integrating novel Casual Dilated Dense Convolutional Blocks (CDDCBs). The second module of the framework, namely, Local-Patch-based Convolution Neural Network (LP-CNN), utilizes Class Activation Map (CAM) (obtained from DCM-CNN) to extract local and fine-grained information. To identify the correct class and minimize the error, a synergic network is utilized that takes the feature maps of both DCM-CNN and LP-CNN and connects both maps in a fully connected fashion to identify the correct class and minimize the errors. The framework is evaluated through a comprehensive set of experiments, both quantitatively and qualitatively, using two publicly available benchmark datasets: RFMiD and ODIR-5K. Our experimental results demonstrate the effectiveness of the proposed framework and achieves higher performance on RFMiD and ODIR-5K datasets compared to reference methods.
PMID:39918766 | DOI:10.1007/s11517-025-03314-0
Evaluation of deep learning-based scatter correction on a long-axial field-of-view PET scanner
Eur J Nucl Med Mol Imaging. 2025 Feb 7. doi: 10.1007/s00259-025-07120-6. Online ahead of print.
ABSTRACT
OBJECTIVE: Long-axial field-of-view (LAFOV) positron emission tomography (PET) systems allow higher sensitivity, with an increased number of detected lines of response induced by a larger angle of acceptance. However this extended angle increases the number of multiple scatters and the scatter contribution within oblique planes. As scattering affects both quality and quantification of the reconstructed image, it is crucial to correct this effect with more accurate methods than the state-of-the-art single scatter simulation (SSS) that can reach its limits with such an extended field-of-view (FOV). In this work, which is an extension of our previous assessment of deep learning-based scatter estimation (DLSE) carried out on a conventional PET system, we aim to evaluate the DLSE method performance on LAFOV total-body PET.
APPROACH: The proposed DLSE method based on an convolutional neural network (CNN) U-Net architecture uses emission and attenuation sinograms to estimate scatter sinogram. The network was trained from Monte-Carlo (MC) simulations of XCAT phantoms [ 18 F]-FDG PET acquisitions using a Siemens Biograph Vision Quadra scanner model, with multiple morphologies and dose distributions. We firstly evaluated the method performance on simulated data in both sinogram and image domain by comparing it to the MC ground truth and SSS scatter sinograms. We then tested the method on seven [ 18 F]-FDG and [ 18 F]-PSMA clinical datasets, and compare it to SSS estimations.
RESULTS: DLSE showed superior accuracy on phantom data, greater robustness to patient size and dose variations compared to SSS, and better lesion contrast recovery. It also yielded promising clinical results, improving lesion contrasts in [ 18 F]-FDG datasets and performing consistently with [ 18 F]-PSMA datasets despite no training with [ 18 F]-PSMA.
SIGNIFICANCE: LAFOV PET scatter can be accurately estimated from raw data using the proposed DLSE method.
PMID:39918764 | DOI:10.1007/s00259-025-07120-6
Applying deep learning for underwater broadband-source detection using a spherical array
J Acoust Soc Am. 2025 Feb 1;157(2):947-961. doi: 10.1121/10.0035787.
ABSTRACT
For improving passive detection of underwater broadband sources, a source-detection and direction-of-arrival-estimation method is developed herein based on a deep neural network (DNN) using a spherical array. Spherical Fourier transform is employed to convert the element pressure signals into spherical Fourier coefficients, which are used as inputs of the DNN. A Gaussian distribution with a spatial-spectrum-like form is adopted to design labels for the DNN. A physical model coupling underwater acoustic propagation and the spherical array is established to simulate array signals for DNN training. The introduction of white noise into the training data considerably enhances the detection capability of the DNN and effectively suppresses false estimation. The model's performance is evaluated based on its detection rate at a constant false alarm rate. Notably, the model does not rely on prior knowledge of the source's spectral features. Further, this study demonstrates that a DNN trained by one source can achieve multisource detection to a certain extent. The simulation and experimental processing results validate the broadband detection capability of the proposed method at varying signal-to-noise ratios.
PMID:39918577 | DOI:10.1121/10.0035787
Automated Description Generation of Cytologic Findings for Lung Cytological Images Using a Pretrained Vision Model and Dual Text Decoders: Preliminary Study
Cytopathology. 2025 Feb 7. doi: 10.1111/cyt.13474. Online ahead of print.
ABSTRACT
OBJECTIVE: Cytology plays a crucial role in lung cancer diagnosis. Pulmonary cytology involves cell morphological characterisation in the specimen and reporting the corresponding findings, which are extremely burdensome tasks. In this study, we propose a technique to generate cytologic findings from for cytologic images to assist in the reporting of pulmonary cytology.
METHODS: For this study, 801 patch images were retrieved using cytology specimens collected from 206 patients; the findings were assigned to each image as a dataset for generating cytologic findings. The proposed method consists of a vision model and dual text decoders. In the former, a convolutional neural network (CNN) is used to classify a given image as benign or malignant, and the features related to the image are extracted from the intermediate layer. Independent text decoders for benign and malignant cells are prepared for text generation, and the text decoder switches according to the CNN classification results. The text decoder is configured using a transformer that uses the features obtained from the CNN for generating findings.
RESULTS: The sensitivity and specificity were 100% and 96.4%, respectively, for automated benign and malignant case classification, and the saliency map indicated characteristic benign and malignant areas. The grammar and style of the generated texts were confirmed correct, achieving a BLEU-4 score of 0.828, reflecting high degree of agreement with the gold standard, outperforming existing LLM-based image-captioning methods and single-text-decoder ablation model.
CONCLUSION: Experimental results indicate that the proposed method is useful for pulmonary cytology classification and generation of cytologic findings.
PMID:39918342 | DOI:10.1111/cyt.13474
I-Brainer: Artificial intelligence/Internet of Things (AI/IoT)-Powered Detection of Brain Cancer
Curr Med Imaging. 2025 Feb 4. doi: 10.2174/0115734056333393250117164020. Online ahead of print.
ABSTRACT
BACKGROUND/OBJECTIVE: Brain tumour is characterized by its aggressive nature and low survival rate and thus regarded as one of the deadliest diseases. Thus, miss-diagnosis or miss-classification of brain tumour can lead to miss treatment or incorrect treatment and reduce survival chances. Therefore, there is need to develop a technique that can identify and detect brain tumour at early stages.
METHODS: Here, we proposed a framework titled I-Brainer which is an Artificial Intelligence/Internet of Things (AI/IoT)-powered classification of MRI. We employed a Br35H+SARTAJ brain MRI dataset which contain 7023 total images which include No tumour, pituitary, meningioma and glioma. In order to accurately classified MRI into 4-class, we developed LeNet model from scratch, implemented 2 pretrained models which include EfficientNet and ResNet-50 as well feature extraction of these models coupled with 2 Machine Learning classifiers k-Nearest Neighbours (KNN) and Support Vector Machines (SVM).
RESULT: Evaluation and comparison of the performance of 3 models has shown that EfficientNet+SVM achieved the best result in terms of AUC (99%) and ResNet-50-KNN ranked higher in terms of accuracy (94%) on testing dataset.
CONCLUSION: This framework can be harness by patients residing in remote areas and as confirmatory approach for medical experts.
PMID:39917913 | DOI:10.2174/0115734056333393250117164020
Design and structure of overlapping regions in PCA via deep learning
Synth Syst Biotechnol. 2024 Dec 27;10(2):442-451. doi: 10.1016/j.synbio.2024.12.007. eCollection 2025 Jun.
ABSTRACT
Polymerase cycling assembly (PCA) stands out as the predominant method in the synthesis of kilobase-length DNA fragments. The design of overlapping regions is the core factor affecting the success rate of synthesis. However, there still exists DNA sequences that are challenging to design and construct in the genome synthesis. Here we proposed a deep learning model based on extensive synthesis data to discern latent sequence representations in overlapping regions with an AUPR of 0.805. Utilizing the model, we developed the SmartCut algorithm aimed at designing oligonucleotides and enhancing the success rate of PCA experiments. This algorithm was successfully applied to sequences with diverse synthesis constraints, 80.4 % of which were synthesized in a single round. We further discovered structure differences represented by major groove width, stagger, slide, and centroid distance between overlapping and non-overlapping regions, which elucidated the model's reasonableness through the lens of physical chemistry. This comprehensive approach facilitates streamlined and efficient investigations into the genome synthesis.
PMID:39917768 | PMC:PMC11799973 | DOI:10.1016/j.synbio.2024.12.007
Advanced AI-assisted panoramic radiograph analysis for periodontal prognostication and alveolar bone loss detection
Front Dent Med. 2025 Jan 6;5:1509361. doi: 10.3389/fdmed.2024.1509361. eCollection 2024.
ABSTRACT
BACKGROUND: Periodontitis is a chronic inflammatory disease affecting the gingival tissues and supporting structures of the teeth, often leading to tooth loss. The condition begins with the accumulation of dental plaque, which initiates an immune response. Current radiographic methods for assessing alveolar bone loss are subjective, time-consuming, and labor-intensive. This study aims to develop an AI-driven model using Convolutional Neural Networks (CNNs) to accurately assess alveolar bone loss and provide individualized periodontal prognoses from panoramic radiographs.
METHODS: A total of 2,000 panoramic radiographs were collected using the same device, based on the periodontal diagnosis codes from the HOSxP Program. Image enhancement techniques were applied, and an AI model based on YOLOv8 was developed to segment teeth, identify the cemento-enamel junction (CEJ), and assess alveolar bone levels. The model quantified bone loss and classified prognoses for each tooth.
RESULTS: The teeth segmentation model achieved 97% accuracy, 90% sensitivity, 96% specificity, and an F1 score of 0.80. The CEJ and bone level segmentation model showed superior results with 98% accuracy, 100% sensitivity, 98% specificity, and an F1 score of 0.90. These findings confirm the models' effectiveness in analyzing panoramic radiographs for periodontal bone loss detection and prognostication.
CONCLUSION: This AI model offers a state-of-the-art approach for assessing alveolar bone loss and predicting individualized periodontal prognoses. It provides a faster, more accurate, and less labor-intensive alternative to current methods, demonstrating its potential for improving periodontal diagnosis and patient outcomes.
PMID:39917716 | PMC:PMC11797906 | DOI:10.3389/fdmed.2024.1509361
Artificial intelligence in dentistry and dental biomaterials
Front Dent Med. 2024 Dec 23;5:1525505. doi: 10.3389/fdmed.2024.1525505. eCollection 2024.
ABSTRACT
Artificial intelligence (AI) technology is being used in various fields and its use is increasingly expanding in dentistry. The key aspects of AI include machine learning (ML), deep learning (DL), and neural networks (NNs). The aim of this review is to present an overview of AI, its various aspects, and its application in biomedicine, dentistry, and dental biomaterials focusing on restorative dentistry and prosthodontics. AI-based systems can be a complementary tool in diagnosis and treatment planning, result prediction, and patient-centered care. AI software can be used to detect restorations, prosthetic crowns, periodontal bone loss, and root canal segmentation from the periapical radiographs. The integration of AI, digital imaging, and 3D printing can provide more precise, durable, and patient-oriented outcomes. AI can be also used for the automatic segmentation of panoramic radiographs showing normal anatomy of the oral and maxillofacial area. Recent advancement in AI in medical and dental sciences includes multimodal deep learning fusion, speech data detection, and neuromorphic computing. Hence, AI has helped dentists in diagnosis, planning, and aid in providing high-quality dental treatments in less time.
PMID:39917699 | PMC:PMC11797767 | DOI:10.3389/fdmed.2024.1525505
Transformer-based short-term traffic forecasting model considering traffic spatiotemporal correlation
Front Neurorobot. 2025 Jan 23;19:1527908. doi: 10.3389/fnbot.2025.1527908. eCollection 2025.
ABSTRACT
Traffic forecasting is crucial for a variety of applications, including route optimization, signal management, and travel time estimation. However, many existing prediction models struggle to accurately capture the spatiotemporal patterns in traffic data due to its inherent nonlinearity, high dimensionality, and complex dependencies. To address these challenges, a short-term traffic forecasting model, Trafficformer, is proposed based on the Transformer framework. The model first uses a multilayer perceptron to extract features from historical traffic data, then enhances spatial interactions through Transformer-based encoding. By incorporating road network topology, a spatial mask filters out noise and irrelevant interactions, improving prediction accuracy. Finally, traffic speed is predicted using another multilayer perceptron. In the experiments, Trafficformer is evaluated on the Seattle Loop Detector dataset. It is compared with six baseline methods, with Mean Absolute Error, Mean Absolute Percentage Error, and Root Mean Square Error used as metrics. The results show that Trafficformer not only has higher prediction accuracy, but also can effectively identify key sections, and has great potential in intelligent traffic control optimization and refined traffic resource allocation.
PMID:39917631 | PMC:PMC11799296 | DOI:10.3389/fnbot.2025.1527908
Artificial intelligence in the radiological diagnosis of cancer
Bioinformation. 2024 Sep 30;20(9):1512-1515. doi: 10.6026/9732063002001512. eCollection 2024.
ABSTRACT
Artificial intelligence (AI) is being used to diagnose deadly diseases such as cancer. The possible decrease in human error, fast diagnosis, and consistency of judgment are the key incentives for implementing these technologies. Therefore, it is of interest to assess the use of artificial intelligence in cancer diagnosis. Total 200 cancer cases were included with 100 cases each of Breast and lung cancer to evaluate with AI and conventional method by the radiologist. The cancer cases were identified with the application of AI-based machine learning techniques. The sensitivity and specificity check-up was used to assess the effectiveness of both approaches. The obtained data was statistically evaluated. AI has shown higher accuracy, sensitivity and specificity in cancer diagnosis compared to manual method of diagnosis by radiologist.
PMID:39917228 | PMC:PMC11795495 | DOI:10.6026/9732063002001512
WDRIV-Net: a weighted ensemble transfer learning to improve automatic type stratification of lumbar intervertebral disc bulge, prolapse, and herniation
Biomed Eng Online. 2025 Feb 6;24(1):11. doi: 10.1186/s12938-025-01341-4.
ABSTRACT
The degeneration of the intervertebral discs in the lumbar spine is the common cause of neurological and physical dysfunctions and chronic disability of patients, which can be stratified into single-(e.g., disc herniation, prolapse, or bulge) and comorbidity-type degeneration (e.g., simultaneous presence of two or more conditions), respectively. A sample of lumbar magnetic resonance imaging (MRI) images from multiple clinical hospitals in China was collected and used in the proposal assessment. We devised a weighted transfer learning framework WDRIV-Net by ensembling four pre-trained models including Densenet169, ResNet101, InceptionV3, and VGG19. The proposed approach was applied to the clinical data and achieved 96.25% accuracy, surpassing the benchmark ResNet101 (87.5%), DenseNet169 (82.5%), VGG19 (88.75%), InceptionV3 (93.75%), and other state-of-the-art (SOTA) ensemble deep learning models. Furthermore, improved performance was observed as well for the metric of the area under the curve (AUC), producing a ≥ 7% increase versus other SOTA ensemble learning, a ≥ 6% increase versus most-studied models, and a ≥ 2% increase versus the baselines. WDRIV-Net can serve as a guide in the initial and efficient type screening of complex degeneration of lumbar intervertebral discs (LID) and assist in the early-stage selection of clinically differentiated treatment options.
PMID:39915867 | DOI:10.1186/s12938-025-01341-4
Prevention and management of degenerative lumbar spine disorders through artificial intelligence-based decision support systems: a systematic review
BMC Musculoskelet Disord. 2025 Feb 7;26(1):126. doi: 10.1186/s12891-025-08356-x.
ABSTRACT
BACKGROUND: Low back pain is the leading cause of disability worldwide with a significant socioeconomic burden; artificial intelligence (AI) has proved to have a great potential in supporting clinical decisions at each stage of the healthcare process. In this article, we have systematically reviewed the available literature on the applications of AI-based Decision Support Systems (DSS) in the clinical prevention and management of Low Back Pain (LBP) due to lumbar degenerative spine disorders.
METHODS: A systematic review of Pubmed and Scopus databases was performed according to the PRISMA statement. Studies reporting the application of DSS to support the prevention and/or management of LBP due to lumbar degenerative diseases were included. The QUADAS-2 tool was utilized to assess the risk of bias in the included studies. The area under the curve (AUC) and accuracy were assessed for each study.
RESULTS: Twenty five articles met the inclusion criteria. Several different machine learning and deep learning algorithms were employed, and their predictive ability on clinical, demographic, psychosocial, and imaging data was assessed. The included studies mainly encompassed three tasks: clinical score definition, clinical assessment, and eligibility prediction and reached AUC scores of 0.93, 0.99 and 0.95, respectively.
CONCLUSIONS: AI-based DSS applications showed a high degree of accuracy in performing a wide set of different tasks. These findings lay the foundation for further research to improve the current understanding and encourage wider adoption of AI in clinical decision-making.
PMID:39915847 | DOI:10.1186/s12891-025-08356-x
Unlocking precision medicine: clinical applications of integrating health records, genetics, and immunology through artificial intelligence
J Biomed Sci. 2025 Feb 7;32(1):16. doi: 10.1186/s12929-024-01110-w.
ABSTRACT
Artificial intelligence (AI) has emerged as a transformative force in precision medicine, revolutionizing the integration and analysis of health records, genetics, and immunology data. This comprehensive review explores the clinical applications of AI-driven analytics in unlocking personalized insights for patients with autoimmune rheumatic diseases. Through the synergistic approach of integrating AI across diverse data sets, clinicians gain a holistic view of patient health and potential risks. Machine learning models excel at identifying high-risk patients, predicting disease activity, and optimizing therapeutic strategies based on clinical, genomic, and immunological profiles. Deep learning techniques have significantly advanced variant calling, pathogenicity prediction, splicing analysis, and MHC-peptide binding predictions in genetics. AI-enabled immunology data analysis, including dimensionality reduction, cell population identification, and sample classification, provides unprecedented insights into complex immune responses. The review highlights real-world examples of AI-driven precision medicine platforms and clinical decision support tools in rheumatology. Evaluation of outcomes demonstrates the clinical benefits and impact of these approaches in revolutionizing patient care. However, challenges such as data quality, privacy, and clinician trust must be navigated for successful implementation. The future of precision medicine lies in the continued research, development, and clinical integration of AI-driven strategies to unlock personalized patient care and drive innovation in rheumatology.
PMID:39915780 | DOI:10.1186/s12929-024-01110-w
Integrating manual annotation with deep transfer learning and radiomics for vertebral fracture analysis
BMC Med Imaging. 2025 Feb 6;25(1):41. doi: 10.1186/s12880-025-01573-9.
ABSTRACT
BACKGROUND: Vertebral compression fractures (VCFs) are prevalent in the elderly, often caused by osteoporosis or trauma. Differentiating acute from chronic VCFs is vital for treatment planning, but MRI, the gold standard, is inaccessible for some. However, CT, a more accessible alternative, lacks precision. This study aimed to enhance CT's diagnostic accuracy for VCFs using deep transfer learning (DTL) and radiomics.
METHODS: We retrospectively analyzed 218 VCF patients scanned with CT and MRI within 3 days from Oct 2022 to Feb 2024. MRI categorized VCFs. 3D regions of interest (ROIs) from CT scans underwent feature extraction and DTL modeling. Receiver operating characteristic (ROC) analysis evaluated models, with the best fused with radiomic features via LASSO. AUCs compared via Delong test, and clinical utility assessed by decision curve analysis (DCA).
RESULTS: Patients were split into training (175) and test (43) sets. Traditional radiomics with LR yielded AUCs of 0.973 (training) and 0.869 (test). Optimal DTL modeling improved to 0.992 (training) and 0.941 (test). Feature fusion further boosted AUCs to 1.000 (training) and 0.964 (test). DCA validated its clinical significance.
CONCLUSION: The feature fusion model enhances the differential diagnosis of acute and chronic VCFs, outperforming single-model approaches and offering a valuable decision-support tool for patients unable to undergo spinal MRI.
PMID:39915711 | DOI:10.1186/s12880-025-01573-9
DeepPrep: an accelerated, scalable and robust pipeline for neuroimaging preprocessing empowered by deep learning
Nat Methods. 2025 Feb 6. doi: 10.1038/s41592-025-02599-1. Online ahead of print.
ABSTRACT
Neuroimaging has entered the era of big data. However, the advancement of preprocessing pipelines falls behind the rapid expansion of data volume, causing substantial computational challenges. Here we present DeepPrep, a pipeline empowered by deep learning and a workflow manager. Evaluated on over 55,000 scans, DeepPrep demonstrates tenfold acceleration, scalability and robustness compared to the state-of-the-art pipeline, thereby meeting the scalability requirements of neuroimaging.
PMID:39915693 | DOI:10.1038/s41592-025-02599-1
Swarm learning with weak supervision enables automatic breast cancer detection in magnetic resonance imaging
Commun Med (Lond). 2025 Feb 6;5(1):38. doi: 10.1038/s43856-024-00722-5.
ABSTRACT
BACKGROUND: Over the next 5 years, new breast cancer screening guidelines recommending magnetic resonance imaging (MRI) for certain patients will significantly increase the volume of imaging data to be analyzed. While this increase poses challenges for radiologists, artificial intelligence (AI) offers potential solutions to manage this workload. However, the development of AI models is often hindered by manual annotation requirements and strict data-sharing regulations between institutions.
METHODS: In this study, we present an integrated pipeline combining weakly supervised learning-reducing the need for detailed annotations-with local AI model training via swarm learning (SL), which circumvents centralized data sharing. We utilized three datasets comprising 1372 female bilateral breast MRI exams from institutions in three countries: the United States (US), Switzerland, and the United Kingdom (UK) to train models. These models were then validated on two external datasets consisting of 649 bilateral breast MRI exams from Germany and Greece.
RESULTS: Upon systematically benchmarking various weakly supervised two-dimensional (2D) and three-dimensional (3D) deep learning (DL) methods, we find that the 3D-ResNet-101 demonstrates superior performance. By implementing a real-world SL setup across three international centers, we observe that these collaboratively trained models outperform those trained locally. Even with a smaller dataset, we demonstrate the practical feasibility of deploying SL internationally with on-site data processing, addressing challenges such as data privacy and annotation variability.
CONCLUSIONS: Combining weakly supervised learning with SL enhances inter-institutional collaboration, improving the utility of distributed datasets for medical AI training without requiring detailed annotations or centralized data sharing.
PMID:39915630 | DOI:10.1038/s43856-024-00722-5
Advanced artificial intelligence with federated learning framework for privacy-preserving cyberthreat detection in IoT-assisted sustainable smart cities
Sci Rep. 2025 Feb 6;15(1):4470. doi: 10.1038/s41598-025-88843-2.
ABSTRACT
With the fast growth of artificial intelligence (AI) and a novel generation of network technology, the Internet of Things (IoT) has become global. Malicious agents regularly utilize novel technical vulnerabilities to use IoT networks in essential industries, the military, defence systems, and medical diagnosis. The IoT has enabled well-known connectivity by connecting many services and objects. However, it has additionally made cloud and IoT frameworks vulnerable to cyberattacks, production cybersecurity major concerns, mainly for the growth of trustworthy IoT networks, particularly those empowering smart city systems. Federated Learning (FL) offers an encouraging solution to address these challenges by providing a privacy-preserving solution for investigating and detecting cyberattacks in IoT systems without negotiating data privacy. Nevertheless, the possibility of FL regarding IoT forensics remains mostly unexplored. Deep learning (DL) focused cyberthreat detection has developed as a powerful and effective approach to identifying abnormal patterns or behaviours in the data field. This manuscript presents an Advanced Artificial Intelligence with a Federated Learning Framework for Privacy-Preserving Cyberthreat Detection (AAIFLF-PPCD) approach in IoT-assisted sustainable smart cities. The AAIFLF-PPCD approach aims to ensure robust and scalable cyberthreat detection while preserving the privacy of IoT users in smart cities. Initially, the AAIFLF-PPCD model utilizes Harris Hawk optimization (HHO)-based feature selection to identify the most related features from the IoT data. Next, the stacked sparse auto-encoder (SSAE) classifier is employed for detecting cyberthreats. Eventually, the walrus optimization algorithm (WOA) is used for hyperparameter tuning to improve the parameters of the SSAE approach and achieve optimal performance. The simulated outcome of the AAIFLF-PPCD technique is evaluated using a benchmark dataset. The performance validation of the AAIFLF-PPCD technique exhibited a superior accuracy value of 99.47% over existing models under diverse measures.
PMID:39915579 | DOI:10.1038/s41598-025-88843-2
Application of deep learning algorithm for judicious use of anti-VEGF in diabetic macular edema
Sci Rep. 2025 Feb 7;15(1):4569. doi: 10.1038/s41598-025-87290-3.
ABSTRACT
Diabetic Macular Edema (DME) is a major complication of diabetic retinopathy characterized by fluid accumulation in the macula, leading to vision impairment. The standard treatment involves anti-VEGF (Vascular Endothelial Growth Factor) therapy, but approximately 36% of patients do not respond adequately, highlighting the need for more precise predictive models to guide treatment. This study aims to develop a Hybrid Deep Learning model to predict treatment responses in DME patients undergoing anti-VEGF therapy, thereby improving the accuracy of treatment planning and minimizing the unnecessary use of costly anti-VEGF agents. The model integrates both Optical Coherence Tomography (OCT) images and clinical data from 181 patients, including key parameters such as serum VEGFR-2 concentration and the duration of DME. The architecture combines convolutional neural networks (CNNs) for image data with multi-layer perceptron (MLP) for tabular clinical data, allowing for a comprehensive analysis of both data types. These pathways converge into a unified predictive framework designed to enhance the model's accuracy. This study utilized a Hybrid Deep Learning model that achieved an 85% accuracy, with additional metrics including precision, recall, and the area under the receiver operating characteristic curve (AUC-ROC) confirming its robustness and reliability. The findings suggest that the model accurately predicts patient responses to anti-VEGF therapy, paving the way for more personalized and targeted treatment strategies. This approach has the potential to enhance patient outcomes and minimize unnecessary administration of anti-VEGF agents, thereby optimizing therapeutic interventions in ophthalmology.
PMID:39915516 | DOI:10.1038/s41598-025-87290-3