Deep learning
A hybrid model for the classification of Autism Spectrum Disorder using Mu rhythm in EEG
Technol Health Care. 2024 Jul 15. doi: 10.3233/THC-240644. Online ahead of print.
ABSTRACT
BACKGROUND: Autism Spectrum Disorder (ASD) is a condition with social interaction, communication, and behavioral difficulties. Diagnostic methods mostly rely on subjective evaluations and can lack objectivity. In this research Machine learning (ML) and deep learning (DL) techniques are used to enhance ASD classification.
OBJECTIVE: This study focuses on improving ASD and TD classification accuracy with a minimal number of EEG channels. ML and DL models are used with EEG data, including Mu Rhythm from the Sensory Motor Cortex (SMC) for classification.
METHODS: Non-linear features in time and frequency domains are extracted and ML models are applied for classification. The EEG 1D data is transformed into images using Independent Component Analysis-Second Order Blind Identification (ICA-SOBI), Spectrogram, and Continuous Wavelet Transform (CWT).
RESULTS: Stacking Classifier employed with non-linear features yields precision, recall, F1-score, and accuracy rates of 78%, 79%, 78%, and 78% respectively. Including entropy and fuzzy entropy features further improves accuracy to 81.4%. In addition, DL models, employing SOBI, CWT, and spectrogram plots, achieve precision, recall, F1-score, and accuracy of 75%, 75%, 74%, and 75% respectively. The hybrid model, which combined deep learning features from spectrogram and CWT with machine learning, exhibits prominent improvement, attained precision, recall, F1-score, and accuracy of 94%, 94%, 94%, and 94% respectively. Incorporating entropy and fuzzy entropy features further improved the accuracy to 96.9%.
CONCLUSIONS: This study underscores the potential of ML and DL techniques in improving the classification of ASD and TD individuals, particularly when utilizing a minimal set of EEG channels.
PMID:39031413 | DOI:10.3233/THC-240644
Gastrointestinal tract disease detection via deep learning based structural and statistical features optimized hexa-classification model
Technol Health Care. 2024 Jul 9. doi: 10.3233/THC-240603. Online ahead of print.
ABSTRACT
BACKGROUND: Gastrointestinal tract (GIT) diseases impact the entire digestive system, spanning from the mouth to the anus. Wireless Capsule Endoscopy (WCE) stands out as an effective analytic instrument for Gastrointestinal tract diseases. Nevertheless, accurately identifying various lesion features, such as irregular sizes, shapes, colors, and textures, remains challenging in this field.
OBJECTIVE: Several computer vision algorithms have been introduced to tackle these challenges, but many relied on handcrafted features, resulting in inaccuracies in various instances.
METHODS: In this work, a novel Deep SS-Hexa model is proposed which is a combination two different deep learning structures for extracting two different features from the WCE images to detect various GIT ailment. The gathered images are denoised by weighted median filter to remove the noisy distortions and augment the images for enhancing the training data. The structural and statistical (SS) feature extraction process is sectioned into two phases for the analysis of distinct regions of gastrointestinal. In the first stage, statistical features of the image are retrieved using MobileNet with the support of SiLU activation function to retrieve the relevant features. In the second phase, the segmented intestine images are transformed into structural features to learn the local information. These SS features are parallelly fused for selecting the best relevant features with walrus optimization algorithm. Finally, Deep belief network (DBN) is used classified the GIT diseases into hexa classes namely normal, ulcer, pylorus, cecum, esophagitis and polyps on the basis of the selected features.
RESULTS: The proposed Deep SS-Hexa model attains an overall average accuracy of 99.16% in GIT disease detection based on KVASIR and KID datasets. The proposed Deep SS-Hexa model achieves high level of accuracy with minimal computational cost in the recognition of GIT illness.
CONCLUSIONS: The proposed Deep SS-Hexa Model progresses the overall accuracy range of 0.04%, 0.80% better than GastroVision, Genetic algorithm based on KVASIR dataset and 0.60%, 1.21% better than Modified U-Net, WCENet based on KID dataset respectively.
PMID:39031411 | DOI:10.3233/THC-240603
CNN-based glioma detection in MRI: A deep learning approach
Technol Health Care. 2024 May 23. doi: 10.3233/THC-240158. Online ahead of print.
ABSTRACT
BACKGROUND: More than a million people are affected by brain tumors each year; high-grade gliomas (HGGs) and low-grade gliomas (LGGs) present serious diagnostic and treatment hurdles, resulting in shortened life expectancies. Glioma segmentation is still a significant difficulty in clinical settings, despite improvements in Magnetic Resonance Imaging (MRI) and diagnostic tools. Convolutional neural networks (CNNs) have seen recent advancements that offer promise for increasing segmentation accuracy, addressing the pressing need for improved diagnostic and therapeutic approaches.
OBJECTIVE: The study intended to develop an automated glioma segmentation algorithm using CNN to accurately identify tumor components in MRI images. The goal was to match the accuracy of experienced radiologists with commercial instruments, hence improving diagnostic precision and quantification.
METHODS: 285 MRI scans of high-grade gliomas (HGGs) and low-grade gliomas (LGGs) were analyzed in the study. T1-weighted sequences were utilised for segmentation both pre-and post-contrast agent administration, along with T2-weighted sequences (with and without Fluid Attenuation by Inversion Recovery [FAIRE]). The segmentation performance was assessed with a U-Net network, renowned for its efficacy in medical image segmentation. DICE coefficients were computed for the tumour core with contrast enhancement, the entire tumour, and the tumour nucleus without contrast enhancement.
RESULTS: The U-Net network produced DICE values of 0.7331 for the tumour core with contrast enhancement, 0.8624 for the total tumour, and 0.7267 for the tumour nucleus without contrast enhancement. The results align with previous studies, demonstrating segmentation accuracy on par with professional radiologists and commercially accessible segmentation tools.
CONCLUSION: The study developed a CNN-based automated segmentation system for gliomas, achieving high accuracy in recognising glioma components in MRI images. The results confirm the ability of CNNs to enhance the accuracy of brain tumour diagnoses, suggesting a promising avenue for future research in medical imaging and diagnostics. This advancement is expected to improve diagnostic processes for clinicians and patients by providing more precise and quantitative results.
PMID:39031408 | DOI:10.3233/THC-240158
Mental fatigue recognition study based on 1D convolutional neural network and short-term ECG signals
Technol Health Care. 2024 Jun 22. doi: 10.3233/THC-240129. Online ahead of print.
ABSTRACT
BACKGROUND: Mental fatigue has become a non-negligible health problem in modern life, as well as one of the important causes of social transportation, production and life accidents.
OBJECTIVE: Fatigue detection based on traditional machine learning requires manual and tedious feature extraction and feature selection engineering, which is inefficient, poor in real-time, and the recognition accuracy needs to be improved. In order to recognize daily mental fatigue level more accurately and in real time, this paper proposes a mental fatigue recognition model based on 1D Convolutional Neural Network (1D-CNN), which inputs 1D raw ECG sequences of 5 s duration into the model, and can directly output the predicted fatigue level labels.
METHODS: The fatigue dataset was constructed by collecting the ECG signals of 22 subjects at three time periods: 9:00-11:00 a.m., 14:00-16:00 p.m., and 19:00-21:00 p.m., and then inputted into the 19-layer 1D-CNN model constructed in the present study for the classification of mental fatigue in three grades.
RESULTS: The results showed that the model was able to recognize the fatigue levels effectively, and its accuracy, precision, recall, and F1 score reached 98.44%, 98.47%, 98.41%, and 98.44%, respectively.
CONCLUSION: This study further improves the accuracy and real-time performance of recognizing multi-level mental fatigue based on electrocardiography, and provides theoretical support for real-time fatigue monitoring in daily life.
PMID:39031407 | DOI:10.3233/THC-240129
Identifying Synergistic Components of Botanical Fungicide Formulations Using Interpretable Graph Neural Networks
J Chem Inf Model. 2024 Jul 20. doi: 10.1021/acs.jcim.4c00128. Online ahead of print.
ABSTRACT
Botanical formulations are promising candidates for developing new biopesticides that can protect crops from pests and diseases while reducing harm to the environment. These biopesticides can be combined with permeation enhancer compounds to boost their efficacy against pests and fungal diseases. However, finding synergistic combinations of these compounds is challenging due to the large and complex chemical space. In this paper, we propose a novel deep learning method that can predict the synergy of botanical products and permeation enhancers based on in vitro assay data. Our method uses a weighted combination of component feature vectors to represent the input mixtures, which enables the model to handle a variable number of components and to interpret the contribution of each component to the synergy. We also employ an ensemble of interpretation methods to provide insights into the underlying mechanisms of synergy. We validate our method by testing the predicted synergistic combinations in wet-lab experiments and show that our method can discover novel and effective biopesticides that would otherwise be difficult to find. Our method is generalizable and applicable to other domains, where predicting mixtures of chemical compounds is important.
PMID:39031079 | DOI:10.1021/acs.jcim.4c00128
Advancements and Challenges in the Integration of Indium Arsenide and Van der Waals Heterostructures
Small. 2024 Jul 19:e2403129. doi: 10.1002/smll.202403129. Online ahead of print.
ABSTRACT
The strategic integration of low-dimensional InAs-based materials and emerging van der Waals systems is advancing in various scientific fields, including electronics, optics, and magnetics. With their unique properties, these InAs-based van der Waals materials and devices promise further miniaturization of semiconductor devices in line with Moore's Law. However, progress in this area lags behind other 2D materials like graphene and boron nitride. Challenges include synthesizing pure crystalline phase InAs nanostructures and single-atomic-layer 2D InAs films, both vital for advanced van der Waals heterostructures. Also, diverse surface state effects on InAs-based van der Waals devices complicate their performance evaluation. This review discusses the experimental advances in the van der Waals epitaxy of InAs-based materials and the working principles of InAs-based van der Waals devices. Theoretical achievements in understanding and guiding the design of InAs-based van der Waals systems are highlighted. Focusing on advancing novel selective area growth and remote epitaxy, exploring multi-functional applications, and incorporating deep learning into first-principles calculations are proposed. These initiatives aim to overcome existing bottlenecks and accelerate transformative advancements in integrating InAs and van der Waals heterostructures.
PMID:39030967 | DOI:10.1002/smll.202403129
Gtie-Rt: A Comprehensive Graph Learning Model for Predicting Drugs Targeting Metabolic Pathways in Human
J Bioinform Comput Biol. 2024 Jul 20:2450010. doi: 10.1142/S0219720024500100. Online ahead of print.
ABSTRACT
Drugs often target specific metabolic pathways to produce a therapeutic effect. However, these pathways are complex and interconnected, making it challenging to predict a drug's potential effects on an organism's overall metabolism. The mapping of drugs with targeting metabolic pathways in the organisms can provide a more complete understanding of the metabolic effects of a drug and help to identify potential drug-drug interactions. In this study, we proposed a machine learning hybrid model Graph Transformer Integrated Encoder (GTIE-RT) for mapping drugs to target metabolic pathways in human. The proposed model is a composite of a Graph Convolution Network (GCN) and transformer encoder for graph embedding and attention mechanism. The output of the transformer encoder is then fed into the Extremely Randomized Trees Classifier to predict target metabolic pathways. The evaluation of the GTIE-RT on drugs dataset demonstrates excellent performance metrics, including accuracy (>95%), recall (>92%), precision (>93%) and F1-score (>92%). Compared to other variants and machine learning methods, GTIE-RT consistently shows more reliable results.
PMID:39030668 | DOI:10.1142/S0219720024500100
STC-UNet: renal tumor segmentation based on enhanced feature extraction at different network levels
BMC Med Imaging. 2024 Jul 19;24(1):179. doi: 10.1186/s12880-024-01359-5.
ABSTRACT
Renal tumors are one of the common diseases of urology, and precise segmentation of these tumors plays a crucial role in aiding physicians to improve diagnostic accuracy and treatment effectiveness. Nevertheless, inherent challenges associated with renal tumors, such as indistinct boundaries, morphological variations, and uncertainties in size and location, segmenting renal tumors accurately remains a significant challenge in the field of medical image segmentation. With the development of deep learning, substantial achievements have been made in the domain of medical image segmentation. However, existing models lack specificity in extracting features of renal tumors across different network hierarchies, which results in insufficient extraction of renal tumor features and subsequently affects the accuracy of renal tumor segmentation. To address this issue, we propose the Selective Kernel, Vision Transformer, and Coordinate Attention Enhanced U-Net (STC-UNet). This model aims to enhance feature extraction, adapting to the distinctive characteristics of renal tumors across various network levels. Specifically, the Selective Kernel modules are introduced in the shallow layers of the U-Net, where detailed features are more abundant. By selectively employing convolutional kernels of different scales, the model enhances its capability to extract detailed features of renal tumors across multiple scales. Subsequently, in the deeper layers of the network, where feature maps are smaller yet contain rich semantic information, the Vision Transformer modules are integrated in a non-patch manner. These assist the model in capturing long-range contextual information globally. Their non-patch implementation facilitates the capture of fine-grained features, thereby achieving collaborative enhancement of global-local information and ultimately strengthening the model's extraction of semantic features of renal tumors. Finally, in the decoder segment, the Coordinate Attention modules embedding positional information are proposed aiming to enhance the model's feature recovery and tumor region localization capabilities. Our model is validated on the KiTS19 dataset, and experimental results indicate that compared to the baseline model, STC-UNet shows improvements of 1.60%, 2.02%, 2.27%, 1.18%, 1.52%, and 1.35% in IoU, Dice, Accuracy, Precision, Recall, and F1-score, respectively. Furthermore, the experimental results demonstrate that the proposed STC-UNet method surpasses other advanced algorithms in both visual effectiveness and objective evaluation metrics.
PMID:39030510 | DOI:10.1186/s12880-024-01359-5
Explainable lung cancer classification with ensemble transfer learning of VGG16, Resnet50 and InceptionV3 using grad-cam
BMC Med Imaging. 2024 Jul 19;24(1):176. doi: 10.1186/s12880-024-01345-x.
ABSTRACT
Medical imaging stands as a critical component in diagnosing various diseases, where traditional methods often rely on manual interpretation and conventional machine learning techniques. These approaches, while effective, come with inherent limitations such as subjectivity in interpretation and constraints in handling complex image features. This research paper proposes an integrated deep learning approach utilizing pre-trained models-VGG16, ResNet50, and InceptionV3-combined within a unified framework to improve diagnostic accuracy in medical imaging. The method focuses on lung cancer detection using images resized and converted to a uniform format to optimize performance and ensure consistency across datasets. Our proposed model leverages the strengths of each pre-trained network, achieving a high degree of feature extraction and robustness by freezing the early convolutional layers and fine-tuning the deeper layers. Additionally, techniques like SMOTE and Gaussian Blur are applied to address class imbalance, enhancing model training on underrepresented classes. The model's performance was validated on the IQ-OTH/NCCD lung cancer dataset, which was collected from the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases over a period of three months in fall 2019. The proposed model achieved an accuracy of 98.18%, with precision and recall rates notably high across all classes. This improvement highlights the potential of integrated deep learning systems in medical diagnostics, providing a more accurate, reliable, and efficient means of disease detection.
PMID:39030496 | DOI:10.1186/s12880-024-01345-x
Enhancing global maritime traffic network forecasting with gravity-inspired deep learning models
Sci Rep. 2024 Jul 19;14(1):16665. doi: 10.1038/s41598-024-67552-2.
ABSTRACT
Aquatic non-indigenous species (NIS) pose significant threats to biodiversity, disrupting ecosystems and inflicting substantial economic damages across agriculture, forestry, and fisheries. Due to the fast growth of global trade and transportation networks, NIS has been introduced and spread unintentionally in new environments. This study develops a new physics-informed model to forecast maritime shipping traffic between port regions worldwide. The predicted information provided by these models, in turn, is used as input for risk assessment of NIS spread through transportation networks to evaluate the capability of our solution. Inspired by the gravity model for international trades, our model considers various factors that influence the likelihood and impact of vessel activities, such as shipping flux density, distance between ports, trade flow, and centrality measures of transportation hubs. Accordingly, this paper introduces transformers to gravity models to rebuild the short- and long-term dependencies that make the risk analysis feasible. Thus, we introduce a physics-inspired framework that achieves an 89% binary accuracy for existing and non-existing trajectories and an 84.8% accuracy for the number of vessels flowing between key port areas, representing more than 10% improvement over the traditional deep-gravity model. Along these lines, this research contributes to a better understanding of NIS risk assessment. It allows policymakers, conservationists, and stakeholders to prioritize management actions by identifying high-risk invasion pathways. Besides, our model is versatile and can include new data sources, making it suitable for assessing international vessel traffic flow in a changing global landscape.
PMID:39030401 | DOI:10.1038/s41598-024-67552-2
Learning generalizable AI models for multi-center histopathology image classification
NPJ Precis Oncol. 2024 Jul 19;8(1):151. doi: 10.1038/s41698-024-00652-4.
ABSTRACT
Investigation of histopathology slides by pathologists is an indispensable component of the routine diagnosis of cancer. Artificial intelligence (AI) has the potential to enhance diagnostic accuracy, improve efficiency, and patient outcomes in clinical pathology. However, variations in tissue preparation, staining protocols, and histopathology slide digitization could result in over-fitting of deep learning models when trained on the data from only one center, thereby underscoring the necessity to generalize deep learning networks for multi-center use. Several techniques, including the use of grayscale images, color normalization techniques, and Adversarial Domain Adaptation (ADA) have been suggested to generalize deep learning algorithms, but there are limitations to their effectiveness and discriminability. Convolutional Neural Networks (CNNs) exhibit higher sensitivity to variations in the amplitude spectrum, whereas humans predominantly rely on phase-related components for object recognition. As such, we propose Adversarial fourIer-based Domain Adaptation (AIDA) which applies the advantages of a Fourier transform in adversarial domain adaptation. We conducted a comprehensive examination of subtype classification tasks in four cancers, incorporating cases from multiple medical centers. Specifically, the datasets included multi-center data for 1113 ovarian cancer cases, 247 pleural cancer cases, 422 bladder cancer cases, and 482 breast cancer cases. Our proposed approach significantly improved performance, achieving superior classification results in the target domain, surpassing the baseline, color augmentation and normalization techniques, and ADA. Furthermore, extensive pathologist reviews suggested that our proposed approach, AIDA, successfully identifies known histotype-specific features. This superior performance highlights AIDA's potential in addressing generalization challenges in deep learning models for multi-center histopathology datasets.
PMID:39030380 | DOI:10.1038/s41698-024-00652-4
Exploring deep learning strategies for intervertebral disc herniation detection on veterinary MRI
Sci Rep. 2024 Jul 19;14(1):16705. doi: 10.1038/s41598-024-67749-5.
ABSTRACT
Intervertebral Disc Herniation (IVDH) is a common spinal disease in dogs, significantly impacting their health, mobility, and overall well-being. This study initiates an effort to automate the detection and localization of IVDH lesions in veterinary MRI scans, utilizing advanced artificial intelligence (AI) methods. A comprehensive canine IVDH dataset, comprising T2-weighted sagittal MRI images from 213 pet dogs of various breeds, ages, and sizes, was compiled and utilized to train and test the IVDH detection models. The experimental results showed that traditional two-stage detection models reliably outperformed one-stage models, including the recent You Only Look Once X (YOLOX) detector. In terms of methodology, this study introduced a novel spinal localization module, successfully integrated into different object detection models to enhance IVDH detection, achieving an average precision (AP) of up to 75.32%. Additionally, transfer learning was explored to adapt the IVDH detection model for a smaller feline dataset. Overall, this study provides insights into advancing AI for veterinary care, identifying challenges and exploring potential strategies for future development in veterinary radiology.
PMID:39030338 | DOI:10.1038/s41598-024-67749-5
Absolute permeability estimation from microtomography rock images through deep learning super-resolution and adversarial fine tuning
Sci Rep. 2024 Jul 19;14(1):16704. doi: 10.1038/s41598-024-67367-1.
ABSTRACT
The carbon capture and storage (CCS) process has become one of the main technologies used for mitigating greenhouse gas emissions. The success of CCS projects relies on accurate subsurface reservoir petrophysical characterization, enabling efficient storage and captured CO 2 containment. In digital rock physics, X-ray microtomography ( μ -CT) is applied to characterize reservoir rocks, allowing a more assertive analysis of physical properties such as porosity and permeability, enabling better simulations of porous media flow. Estimating petrophysical properties through numeric simulations usually requires high-resolution images, which are expensive and time-inefficient to obtain with μ -CT. To address this, we propose using two deep learning models: a super-resolution model to enhance the quality of low-resolution images and a surrogate model that acts as a substitute for numerical simulations to estimate the petrophysical property of interest. A correction process inspired by generative adversarial network (GAN) adversarial training is applied. In this approach, the super-resolution model acts as a generator, creating high-resolution images, and the surrogate network acts as a discriminator. By adjusting the generator, images that correct the errors in the surrogate's estimations are produced. The proposed method was applied to the DeePore dataset. The results shows the proposed approach improved permeability estimation overall.
PMID:39030317 | DOI:10.1038/s41598-024-67367-1
Integrated image-based deep learning and language models for primary diabetes care
Nat Med. 2024 Jul 19. doi: 10.1038/s41591-024-03139-8. Online ahead of print.
ABSTRACT
Primary diabetes care and diabetic retinopathy (DR) screening persist as major public health challenges due to a shortage of trained primary care physicians (PCPs), particularly in low-resource settings. Here, to bridge the gaps, we developed an integrated image-language system (DeepDR-LLM), combining a large language model (LLM module) and image-based deep learning (DeepDR-Transformer), to provide individualized diabetes management recommendations to PCPs. In a retrospective evaluation, the LLM module demonstrated comparable performance to PCPs and endocrinology residents when tested in English and outperformed PCPs and had comparable performance to endocrinology residents in Chinese. For identifying referable DR, the average PCP's accuracy was 81.0% unassisted and 92.3% assisted by DeepDR-Transformer. Furthermore, we performed a single-center real-world prospective study, deploying DeepDR-LLM. We compared diabetes management adherence of patients under the unassisted PCP arm (n = 397) with those under the PCP+DeepDR-LLM arm (n = 372). Patients with newly diagnosed diabetes in the PCP+DeepDR-LLM arm showed better self-management behaviors throughout follow-up (P < 0.05). For patients with referral DR, those in the PCP+DeepDR-LLM arm were more likely to adhere to DR referrals (P < 0.01). Additionally, DeepDR-LLM deployment improved the quality and empathy level of management recommendations. Given its multifaceted performance, DeepDR-LLM holds promise as a digital solution for enhancing primary diabetes care and DR screening.
PMID:39030266 | DOI:10.1038/s41591-024-03139-8
Intelligent breast cancer diagnosis with two-stage using mammogram images
Sci Rep. 2024 Jul 19;14(1):16672. doi: 10.1038/s41598-024-65926-0.
ABSTRACT
Breast cancer (BC) significantly contributes to cancer-related mortality in women, underscoring the criticality of early detection for optimal patient outcomes. Mammography is a key tool for identifying and diagnosing breast abnormalities; however, accurately distinguishing malignant mass lesions remains challenging. To address this issue, we propose a novel deep learning approach for BC screening utilizing mammography images. Our proposed model comprises three distinct stages: data collection from established benchmark sources, image segmentation employing an Atrous Convolution-based Attentive and Adaptive Trans-Res-UNet (ACA-ATRUNet) architecture, and BC identification via an Atrous Convolution-based Attentive and Adaptive Multi-scale DenseNet (ACA-AMDN) model. The hyperparameters within the ACA-ATRUNet and ACA-AMDN models are optimized using the Modified Mussel Length-based Eurasian Oystercatcher Optimization (MML-EOO) algorithm. The performance is evaluated using a variety of metrics, and a comparative analysis against conventional methods is presented. Our experimental results reveal that the proposed BC detection framework attains superior precision rates in early disease detection, demonstrating its potential to enhance mammography-based screening methodologies.
PMID:39030248 | DOI:10.1038/s41598-024-65926-0
Automated PD-L1 status prediction in lung cancer with multi-modal PET/CT fusion
Sci Rep. 2024 Jul 19;14(1):16720. doi: 10.1038/s41598-024-66487-y.
ABSTRACT
Programmed death-ligand 1 (PD-L1) expressions play a crucial role in guiding therapeutic interventions such as the use of tyrosine kinase inhibitors (TKIs) and immune checkpoint inhibitors (ICIs) in lung cancer. Conventional determination of PD-L1 status includes careful surgical or biopsied tumor specimens. These specimens are gathered through invasive procedures, representing a risk of difficulties and potential challenges in getting reliable and representative tissue samples. Using a single center cohort of 189 patients, our objective was to evaluate various fusion methods that used non-invasive computed tomography (CT) and 18 F-FDG positron emission tomography (PET) images as inputs to various deep learning models to automatically predict PD-L1 in non-small cell lung cancer (NSCLC). We compared three different architectures (ResNet, DenseNet, and EfficientNet) and considered different input data (CT only, PET only, PET/CT early fusion, PET/CT late fusion without as well as with partially and fully shared weights to determine the best model performance. Models were assessed utilizing areas under the receiver operating characteristic curves (AUCs) considering their 95% confidence intervals (CI). The fusion of PET and CT images as input yielded better performance for PD-L1 classification. The different data fusion schemes systematically outperformed their individual counterparts when used as input of the various deep models. Furthermore, early fusion consistently outperformed late fusion, probably as a result of its capacity to capture more complicated patterns by merging PET and CT derived content at a lower level. When we looked more closely at the effects of weight sharing in late fusion architectures, we discovered that while it might boost model stability, it did not always result in better results. This suggests that although weight sharing could be beneficial when modality parameters are similar, the anatomical and metabolic information provided by CT and PET scans are too dissimilar to consistently lead to improved PD-L1 status predictions.
PMID:39030240 | DOI:10.1038/s41598-024-66487-y
Multi-degradation-adaptation network for fundus image enhancement with degradation representation learning
Med Image Anal. 2024 Jul 14;97:103273. doi: 10.1016/j.media.2024.103273. Online ahead of print.
ABSTRACT
Fundus image quality serves a crucial asset for medical diagnosis and applications. However, such images often suffer degradation during image acquisition where multiple types of degradation can occur in each image. Although recent deep learning based methods have shown promising results in image enhancement, they tend to focus on restoring one aspect of degradation and lack generalisability to multiple modes of degradation. We propose an adaptive image enhancement network that can simultaneously handle a mixture of different degradations. The main contribution of this work is to introduce our Multi-Degradation-Adaptive module which dynamically generates filters for different types of degradation. Moreover, we explore degradation representation learning and propose the degradation representation network and Multi-Degradation-Adaptive discriminator for our accompanying image enhancement network. Experimental results demonstrate that our method outperforms several existing state-of-the-art methods in fundus image enhancement. Code will be available at https://github.com/RuoyuGuo/MDA-Net.
PMID:39029157 | DOI:10.1016/j.media.2024.103273
Sequence of Morphological Changes Preceding Atrophy in Intermediate AMD Using Deep Learning
Invest Ophthalmol Vis Sci. 2024 Jul 1;65(8):30. doi: 10.1167/iovs.65.8.30.
ABSTRACT
PURPOSE: Investigating the sequence of morphological changes preceding outer plexiform layer (OPL) subsidence, a marker preceding geographic atrophy, in intermediate AMD (iAMD) using high-precision artificial intelligence (AI) quantifications on optical coherence tomography imaging.
METHODS: In this longitudinal observational study, individuals with bilateral iAMD participating in a multicenter clinical trial were screened for OPL subsidence and RPE and outer retinal atrophy. OPL subsidence was segmented on an A-scan basis in optical coherence tomography volumes, obtained 6-monthly with 36 months follow-up. AI-based quantification of photoreceptor (PR) and outer nuclear layer (ONL) thickness, drusen height and choroidal hypertransmission (HT) was performed. Changes were compared between topographic areas of OPL subsidence (AS), drusen (AD), and reference (AR).
RESULTS: Of 280 eyes of 140 individuals, OPL subsidence occurred in 53 eyes from 43 individuals. Thirty-six eyes developed RPE and outer retinal atrophy subsequently. In the cohort of 53 eyes showing OPL subsidence, PR and ONL thicknesses were significantly decreased in AS compared with AD and AR 12 and 18 months before OPL subsidence occurred, respectively (PR: 20 µm vs. 23 µm and 27 µm [P < 0.009]; ONL, 84 µm vs. 94 µm and 98 µm [P < 0.008]). Accelerated thinning of PR (0.6 µm/month; P < 0.001) and ONL (0.8 µm/month; P < 0.001) was observed in AS compared with AD and AR. Concomitant drusen regression and hypertransmission increase at the occurrence of OPL subsidence underline the atrophic progress in areas affected by OPL subsidence.
CONCLUSIONS: PR and ONL thinning are early subclinical features associated with subsequent OPL subsidence, an indicator of progression toward geographic atrophy. AI algorithms are able to predict and quantify morphological precursors of iAMD conversion and allow personalized risk stratification.
PMID:39028907 | DOI:10.1167/iovs.65.8.30
Data-driven rogue waves solutions for the focusing and variable coefficient nonlinear Schrodinger equations via deep learning
Chaos. 2024 Jul 1;34(7):073134. doi: 10.1063/5.0209068.
ABSTRACT
In this paper, we investigate the data-driven rogue waves solutions of the focusing and the variable coefficient nonlinear Schrödinger (NLS) equations by the deep learning method from initial and boundary conditions. Specifically, first- and second-order rogue wave solutions for the focusing NLS equation and three deformed rogue wave solutions for the variable coefficient NLS equation are solved using physics-informed memory networks (PIMNs). The effects of optimization algorithm, network structure, and mesh size on the solution accuracy are discussed. Numerical experiments clearly demonstrate that the PIMNs can capture the nonlinear features of rogue waves solutions very well. This is of great significance for revealing the dynamical behavior of the rogue waves solutions and advancing the application of deep learning in the field of solving partial differential equations.
PMID:39028903 | DOI:10.1063/5.0209068
Deep learning based detection and classification of fetal lip in ultrasound images
J Perinat Med. 2024 Jul 22. doi: 10.1515/jpm-2024-0122. Online ahead of print.
ABSTRACT
OBJECTIVES: Fetal cleft lip is a common congenital defect. Considering the delicacy and difficulty of observing fetal lips, we have utilized deep learning technology to develop a new model aimed at quickly and accurately assessing the development of fetal lips during prenatal examinations. This model can detect ultrasound images of the fetal lips and classify them, aiming to provide a more objective prediction for the development of fetal lips.
METHODS: This study included 632 pregnant women in their mid-pregnancy stage, who underwent ultrasound examinations of the fetal lips, collecting both normal and abnormal fetal lip ultrasound images. To improve the accuracy of the detection and classification of fetal lips, we proposed and validated the Yolov5-ECA model.
RESULTS: The experimental results show that, compared with the currently popular 10 models, our model achieved the best results in the detection and classification of fetal lips. In terms of the detection of fetal lips, the mAP@0.5 and mAP@0.5:0.95 were 0.920 and 0.630, respectively. In the classification of fetal lip ultrasound images, the accuracy reached 0.925.
CONCLUSIONS: The deep learning algorithm has accuracy consistent with manual evaluation in the detection and classification process of fetal lips. This automated recognition technology can provide a powerful tool for inexperienced young doctors, helping them to accurately conduct examinations and diagnoses of fetal lips.
PMID:39028804 | DOI:10.1515/jpm-2024-0122