Deep learning

Deep learning-based image quality assessment for optical coherence tomography macular scans: a multicentre study

Sat, 2024-07-20 06:00

Br J Ophthalmol. 2024 Jul 20:bjo-2023-323871. doi: 10.1136/bjo-2023-323871. Online ahead of print.

ABSTRACT

AIMS: To develop and externally test deep learning (DL) models for assessing the image quality of three-dimensional (3D) macular scans from Cirrus and Spectralis optical coherence tomography devices.

METHODS: We retrospectively collected two data sets including 2277 Cirrus 3D scans and 1557 Spectralis 3D scans, respectively, for training (70%), fine-tuning (10%) and internal validation (20%) from electronic medical and research records at The Chinese University of Hong Kong Eye Centre and the Hong Kong Eye Hospital. Scans with various eye diseases (eg, diabetic macular oedema, age-related macular degeneration, polypoidal choroidal vasculopathy and pathological myopia), and scans of normal eyes from adults and children were included. Two graders labelled each 3D scan as gradable or ungradable, according to standardised criteria. We used a 3D version of the residual network (ResNet)-18 for Cirrus 3D scans and a multiple-instance learning pipline with ResNet-18 for Spectralis 3D scans. Two deep learning (DL) models were further tested via three unseen Cirrus data sets from Singapore and five unseen Spectralis data sets from India, Australia and Hong Kong, respectively.

RESULTS: In the internal validation, the models achieved the area under curves (AUCs) of 0.930 (0.885-0.976) and 0.906 (0.863-0.948) for assessing the Cirrus 3D scans and Spectralis 3D scans, respectively. In the external testing, the models showed robust performance with AUCs ranging from 0.832 (0.730-0.934) to 0.930 (0.906-0.953) and 0.891 (0.836-0.945) to 0.962 (0.918-1.000), respectively.

CONCLUSIONS: Our models could be used for filtering out ungradable 3D scans and further incorporated with a disease-detection DL model, allowing a fully automated eye disease detection workflow.

PMID:39033014 | DOI:10.1136/bjo-2023-323871

Categories: Literature Watch

Advancements in computer vision and pathology: Unraveling the potential of artificial intelligence for precision diagnosis and beyond

Sat, 2024-07-20 06:00

Adv Cancer Res. 2024;161:431-478. doi: 10.1016/bs.acr.2024.05.006. Epub 2024 Jun 26.

ABSTRACT

The integration of computer vision into pathology through slide digitalization represents a transformative leap in the field's evolution. Traditional pathology methods, while reliable, are often time-consuming and susceptible to intra- and interobserver variability. In contrast, computer vision, empowered by artificial intelligence (AI) and machine learning (ML), promises revolutionary changes, offering consistent, reproducible, and objective results with ever-increasing speed and scalability. The applications of advanced algorithms and deep learning architectures like CNNs and U-Nets augment pathologists' diagnostic capabilities, opening new frontiers in automated image analysis. As these technologies mature and integrate into digital pathology workflows, they are poised to provide deeper insights into disease processes, quantify and standardize biomarkers, enhance patient outcomes, and automate routine tasks, reducing pathologists' workload. However, this transformative force calls for cross-disciplinary collaboration between pathologists, computer scientists, and industry innovators to drive research and development. While acknowledging its potential, this chapter addresses the limitations of AI in pathology, encompassing technical, practical, and ethical considerations during development and implementation.

PMID:39032956 | DOI:10.1016/bs.acr.2024.05.006

Categories: Literature Watch

Deep Learning-Assisted Two-Dimensional Transperineal Ultrasound for Analyzing Bladder Neck Motion in Women With Stress Urinary Incontinence

Sat, 2024-07-20 06:00

Am J Obstet Gynecol. 2024 Jul 18:S0002-9378(24)00764-6. doi: 10.1016/j.ajog.2024.07.021. Online ahead of print.

ABSTRACT

BACKGROUND: No universally recognized transperineal ultrasound parameters are available for evaluating stress urinary incontinence. The information captured by commonly used perineal ultrasound parameters is limited and insufficient for a comprehensive assessment of stress urinary incontinence. Although bladder neck motion plays a major role in stress urinary incontinence, objective and visual methods to evaluate its impact on stress urinary incontinence remain lacking.

OBJECTIVE: To use a deep learning-based system to evaluate bladder neck motion using two-dimensional transperineal ultrasound videos, exploring motion parameters for diagnosing and evaluating stress urinary incontinence. We hypothesized that bladder neck motion parameters are associated with stress urinary incontinence and are useful for stress urinary incontinence diagnosis and evaluation.

STUDY DESIGN: This retrospective study including 217 women involved the following parameters: maximum and average speeds of bladder neck descent, β angle, urethral rotation angle, and duration of the Valsalva maneuver. The fitted curves were derived to visualize bladder neck motion trajectories. Comparative analyses were conducted to assess these parameters between stress urinary incontinence and control groups. Logistic regression and receiver operating characteristic curve analyses were employed to evaluate the diagnostic performance of each motion parameter and their combinations for stress urinary incontinence.

RESULTS: Overall, 173 women were enrolled in this study (82, stress urinary incontinence group; 91, control group). No significant differences were observed in the maximum and average speeds of bladder neck descent and in the speed variance of bladder neck descent. The maximum and average speed of the β and urethral rotation angles were faster in the stress urinary incontinence group than in the control group (151.2 vs 109.0 mm/s, P=0.001; 6.0 vs 3.1 mm/s, P <0.001; 105.5 vs 69.6 mm/s, P <0.001; 10.1 vs 7.9 mm/s, P=0.011, respectively). The speed variance of the β and urethral rotation angles were higher in the stress urinary incontinence group (844.8 vs 336.4, P <0.001; 347.6 vs 131.1, P <0.001, respectively). The combination of the average speed of the β angle, maximum speed of the urethral rotation angle, and duration of the Valsalva maneuver demonstrated a strong diagnostic performance (area under the curve, 0.87). When 0.481*β anglea + 0.013*URAm + 0.483*Dval = 7.405, the diagnostic sensitivity was 70% and specificity was 92%, highlighting the significant role of bladder neck motion in stress urinary incontinence, particularly changes in the speed of the β and urethral rotation angles.

CONCLUSIONS: A system utilizing deep learning can describe the motion of the bladder neck in women with stress urinary incontinence during the Valsalva maneuver, making it possible to visualize and quantify bladder neck motion on transperineal ultrasound. The speeds of the β and urethral rotation angles and duration of the Valsalva maneuver were relatively reliable diagnostic parameters.

PMID:39032723 | DOI:10.1016/j.ajog.2024.07.021

Categories: Literature Watch

Advancing jasmine tea production: YOLOv7-based real-time jasmine flower detection

Sat, 2024-07-20 06:00

J Sci Food Agric. 2024 Jul 19. doi: 10.1002/jsfa.13752. Online ahead of print.

ABSTRACT

BACKGROUND: To produce jasmine tea of excellent quality, it is crucial to select jasmine flowers at their optimal growth stage during harvesting. However, achieving this goal remains a challenge due to environmental and manual factors. This study addresses this issue by classifying different jasmine flowers based on visual attributes using the YOLOv7 algorithm, one of the most advanced algorithms in convolutional neural networks.

RESULTS: The mean average precision (mAP value) for detecting jasmine flowers using this model is 0.948, and the accuracy for five different degrees of openness of jasmine flowers, namely small buds, buds, half-open, full-open and wiltered, is 87.7%, 90.3%, 89%, 93.9% and 86.4%, respectively. Meanwhile, other ways of processing the images in the dataset, such as blurring and changing the brightness, also increased the credibility of the algorithm.

CONCLUSION: This study shows that it is feasible to use deep learning algorithms for distinguishing jasmine flowers at different growth stages. This study can provide a reference for jasmine production estimation and for the development of intelligent and precise flower-picking applications to reduce flower waste and production costs. © 2024 Society of Chemical Industry.

PMID:39032018 | DOI:10.1002/jsfa.13752

Categories: Literature Watch

Interpretation of Artificial Intelligence Models in Healthcare: A Pictorial Guide for Clinicians

Sat, 2024-07-20 06:00

J Ultrasound Med. 2024 Jul 19. doi: 10.1002/jum.16524. Online ahead of print.

ABSTRACT

Artificial intelligence (AI) models can play a more effective role in managing patients with the explosion of digital health records available in the healthcare industry. Machine-learning (ML) and deep-learning (DL) techniques are two methods used to develop predictive models that serve to improve the clinical processes in the healthcare industry. These models are also implemented in medical imaging machines to empower them with an intelligent decision system to aid physicians in their decisions and increase the efficiency of their routine clinical practices. The physicians who are going to work with these machines need to have an insight into what happens in the background of the implemented models and how they work. More importantly, they need to be able to interpret their predictions, assess their performance, and compare them to find the one with the best performance and fewer errors. This review aims to provide an accessible overview of key evaluation metrics for physicians without AI expertise. In this review, we developed four real-world diagnostic AI models (two ML and two DL models) for breast cancer diagnosis using ultrasound images. Then, 23 of the most commonly used evaluation metrics were reviewed uncomplicatedly for physicians. Finally, all metrics were calculated and used practically to interpret and evaluate the outputs of the models. Accessible explanations and practical applications empower physicians to effectively interpret, evaluate, and optimize AI models to ensure safety and efficacy when integrated into clinical practice.

PMID:39032010 | DOI:10.1002/jum.16524

Categories: Literature Watch

A novel method combining deep learning with the Kennard-Stone algorithm for training dataset selection for image-based rice seed variety identification

Sat, 2024-07-20 06:00

J Sci Food Agric. 2024 Jun 21. doi: 10.1002/jsfa.13668. Online ahead of print.

ABSTRACT

BACKGROUND: Different varieties of rice vary in planting time, stress resistance, and other characteristics. With advances in rice-breeding technology, the number of rice varieties has increased significantly, making variety identification crucial for both trading and planting.

RESULTS: This study collected RGB images of 20 hybrid rice seed varieties. An enhanced deep super-resolution network (EDSR) was employed to enhance image resolution, and a variety classification model utilizing the high-resolution dataset demonstrated superior performance to that of the model using the low-resolution dataset. A novel training sample selection methodology was introduced integrating deep learning with the Kennard-Stone (KS) algorithm. Convolutional neural networks (CNN) and autoencoders served as supervised and unsupervised feature extractors, respectively. The extracted feature vectors were subsequently processed by the KS algorithm to select training samples. The proposed methodologies exhibited superior performance over the random selection approach in rice variety classification, with an approximately 10.08% improvement in overall classification accuracy. Furthermore, the impact of noise on the proposed methodology was investigated by introducing noise to the images, and the proposed methodologies maintained superior performance relative to the random selection approach on the noisy image dataset.

CONCLUSION: The experimental results indicate that both supervised and unsupervised learning models performed effectively as feature extractors, and the deep learning framework significantly influenced the selection of training set samples. This study presents a novel approach for training sample selection in classification tasks and suggests the potential for extending the proposed method to image datasets and other types of datasets. Further exploration of this potential is warranted. © 2024 Society of Chemical Industry.

PMID:39031773 | DOI:10.1002/jsfa.13668

Categories: Literature Watch

Development of an efficient novel method for coronary artery disease prediction using machine learning and deep learning techniques

Sat, 2024-07-20 06:00

Technol Health Care. 2024 Jul 15. doi: 10.3233/THC-240740. Online ahead of print.

ABSTRACT

BACKGROUND: Heart disease is a severe health issue that results in high fatality rates worldwide. Identifying cardiovascular diseases such as coronary artery disease (CAD) and heart attacks through repetitive clinical data analysis is a significant task. Detecting heart disease in its early stages can save lives. The most lethal cardiovascular condition is CAD, which develops over time due to plaque buildup in coronary arteries, causing incomplete blood flow obstruction. Machine Learning (ML) is progressively used in the medical sector to detect CAD disease.

OBJECTIVE: The primary aim of this work is to deliver a state-of-the-art approach to enhancing CAD prediction accuracy by using a DL algorithm in a classification context.

METHODS: A unique ML technique is proposed in this study to predict CAD disease accurately using a deep learning algorithm in a classification context. An ensemble voting classifier classification model is developed based on various methods such as Naïve Bayes (NB), Logistic Regression (LR), Decision Tree (DT), XGBoost, Random Forest (RF), Convolutional Neural Network (CNN), Support Vector Machine (SVM), K Nearest Neighbor (KNN), Bidirectional LSTM and Long Short-Term Memory (LSTM). The performance of the ensemble models and a novel model are compared in this study. The Alizadeh Sani dataset, which consists of a random sample of 216 cases with CAD, is used in this study. Synthetic Minority Over Sampling Technique (SMOTE) is used to address the issue of imbalanced datasets, and the Chi-square test is used for feature selection optimization. Performance is assessed using various assessment methodologies, such as confusion matrix, accuracy, recall, precision, f1-score, and auc-roc.

RESULTS: When a novel algorithm achieves the highest accuracy relative to other algorithms, it demonstrates its effectiveness in several ways, including superior performance, robustness, generalization capability, efficiency, innovative approaches, and benchmarking against baselines. These characteristics collectively contribute to establishing the novel algorithm as a promising solution for addressing the target problem in machine learning and related fields.

CONCLUSION: Implementing the novel model in this study significantly improved performance, achieving a prediction accuracy rate of 92% in the detection of CAD. These findings are competitive and on par with the top outcomes among other methods.

PMID:39031414 | DOI:10.3233/THC-240740

Categories: Literature Watch

A hybrid model for the classification of Autism Spectrum Disorder using Mu rhythm in EEG

Sat, 2024-07-20 06:00

Technol Health Care. 2024 Jul 15. doi: 10.3233/THC-240644. Online ahead of print.

ABSTRACT

BACKGROUND: Autism Spectrum Disorder (ASD) is a condition with social interaction, communication, and behavioral difficulties. Diagnostic methods mostly rely on subjective evaluations and can lack objectivity. In this research Machine learning (ML) and deep learning (DL) techniques are used to enhance ASD classification.

OBJECTIVE: This study focuses on improving ASD and TD classification accuracy with a minimal number of EEG channels. ML and DL models are used with EEG data, including Mu Rhythm from the Sensory Motor Cortex (SMC) for classification.

METHODS: Non-linear features in time and frequency domains are extracted and ML models are applied for classification. The EEG 1D data is transformed into images using Independent Component Analysis-Second Order Blind Identification (ICA-SOBI), Spectrogram, and Continuous Wavelet Transform (CWT).

RESULTS: Stacking Classifier employed with non-linear features yields precision, recall, F1-score, and accuracy rates of 78%, 79%, 78%, and 78% respectively. Including entropy and fuzzy entropy features further improves accuracy to 81.4%. In addition, DL models, employing SOBI, CWT, and spectrogram plots, achieve precision, recall, F1-score, and accuracy of 75%, 75%, 74%, and 75% respectively. The hybrid model, which combined deep learning features from spectrogram and CWT with machine learning, exhibits prominent improvement, attained precision, recall, F1-score, and accuracy of 94%, 94%, 94%, and 94% respectively. Incorporating entropy and fuzzy entropy features further improved the accuracy to 96.9%.

CONCLUSIONS: This study underscores the potential of ML and DL techniques in improving the classification of ASD and TD individuals, particularly when utilizing a minimal set of EEG channels.

PMID:39031413 | DOI:10.3233/THC-240644

Categories: Literature Watch

Gastrointestinal tract disease detection via deep learning based structural and statistical features optimized hexa-classification model

Sat, 2024-07-20 06:00

Technol Health Care. 2024 Jul 9. doi: 10.3233/THC-240603. Online ahead of print.

ABSTRACT

BACKGROUND: Gastrointestinal tract (GIT) diseases impact the entire digestive system, spanning from the mouth to the anus. Wireless Capsule Endoscopy (WCE) stands out as an effective analytic instrument for Gastrointestinal tract diseases. Nevertheless, accurately identifying various lesion features, such as irregular sizes, shapes, colors, and textures, remains challenging in this field.

OBJECTIVE: Several computer vision algorithms have been introduced to tackle these challenges, but many relied on handcrafted features, resulting in inaccuracies in various instances.

METHODS: In this work, a novel Deep SS-Hexa model is proposed which is a combination two different deep learning structures for extracting two different features from the WCE images to detect various GIT ailment. The gathered images are denoised by weighted median filter to remove the noisy distortions and augment the images for enhancing the training data. The structural and statistical (SS) feature extraction process is sectioned into two phases for the analysis of distinct regions of gastrointestinal. In the first stage, statistical features of the image are retrieved using MobileNet with the support of SiLU activation function to retrieve the relevant features. In the second phase, the segmented intestine images are transformed into structural features to learn the local information. These SS features are parallelly fused for selecting the best relevant features with walrus optimization algorithm. Finally, Deep belief network (DBN) is used classified the GIT diseases into hexa classes namely normal, ulcer, pylorus, cecum, esophagitis and polyps on the basis of the selected features.

RESULTS: The proposed Deep SS-Hexa model attains an overall average accuracy of 99.16% in GIT disease detection based on KVASIR and KID datasets. The proposed Deep SS-Hexa model achieves high level of accuracy with minimal computational cost in the recognition of GIT illness.

CONCLUSIONS: The proposed Deep SS-Hexa Model progresses the overall accuracy range of 0.04%, 0.80% better than GastroVision, Genetic algorithm based on KVASIR dataset and 0.60%, 1.21% better than Modified U-Net, WCENet based on KID dataset respectively.

PMID:39031411 | DOI:10.3233/THC-240603

Categories: Literature Watch

CNN-based glioma detection in MRI: A deep learning approach

Sat, 2024-07-20 06:00

Technol Health Care. 2024 May 23. doi: 10.3233/THC-240158. Online ahead of print.

ABSTRACT

BACKGROUND: More than a million people are affected by brain tumors each year; high-grade gliomas (HGGs) and low-grade gliomas (LGGs) present serious diagnostic and treatment hurdles, resulting in shortened life expectancies. Glioma segmentation is still a significant difficulty in clinical settings, despite improvements in Magnetic Resonance Imaging (MRI) and diagnostic tools. Convolutional neural networks (CNNs) have seen recent advancements that offer promise for increasing segmentation accuracy, addressing the pressing need for improved diagnostic and therapeutic approaches.

OBJECTIVE: The study intended to develop an automated glioma segmentation algorithm using CNN to accurately identify tumor components in MRI images. The goal was to match the accuracy of experienced radiologists with commercial instruments, hence improving diagnostic precision and quantification.

METHODS: 285 MRI scans of high-grade gliomas (HGGs) and low-grade gliomas (LGGs) were analyzed in the study. T1-weighted sequences were utilised for segmentation both pre-and post-contrast agent administration, along with T2-weighted sequences (with and without Fluid Attenuation by Inversion Recovery [FAIRE]). The segmentation performance was assessed with a U-Net network, renowned for its efficacy in medical image segmentation. DICE coefficients were computed for the tumour core with contrast enhancement, the entire tumour, and the tumour nucleus without contrast enhancement.

RESULTS: The U-Net network produced DICE values of 0.7331 for the tumour core with contrast enhancement, 0.8624 for the total tumour, and 0.7267 for the tumour nucleus without contrast enhancement. The results align with previous studies, demonstrating segmentation accuracy on par with professional radiologists and commercially accessible segmentation tools.

CONCLUSION: The study developed a CNN-based automated segmentation system for gliomas, achieving high accuracy in recognising glioma components in MRI images. The results confirm the ability of CNNs to enhance the accuracy of brain tumour diagnoses, suggesting a promising avenue for future research in medical imaging and diagnostics. This advancement is expected to improve diagnostic processes for clinicians and patients by providing more precise and quantitative results.

PMID:39031408 | DOI:10.3233/THC-240158

Categories: Literature Watch

Mental fatigue recognition study based on 1D convolutional neural network and short-term ECG signals

Sat, 2024-07-20 06:00

Technol Health Care. 2024 Jun 22. doi: 10.3233/THC-240129. Online ahead of print.

ABSTRACT

BACKGROUND: Mental fatigue has become a non-negligible health problem in modern life, as well as one of the important causes of social transportation, production and life accidents.

OBJECTIVE: Fatigue detection based on traditional machine learning requires manual and tedious feature extraction and feature selection engineering, which is inefficient, poor in real-time, and the recognition accuracy needs to be improved. In order to recognize daily mental fatigue level more accurately and in real time, this paper proposes a mental fatigue recognition model based on 1D Convolutional Neural Network (1D-CNN), which inputs 1D raw ECG sequences of 5 s duration into the model, and can directly output the predicted fatigue level labels.

METHODS: The fatigue dataset was constructed by collecting the ECG signals of 22 subjects at three time periods: 9:00-11:00 a.m., 14:00-16:00 p.m., and 19:00-21:00 p.m., and then inputted into the 19-layer 1D-CNN model constructed in the present study for the classification of mental fatigue in three grades.

RESULTS: The results showed that the model was able to recognize the fatigue levels effectively, and its accuracy, precision, recall, and F1 score reached 98.44%, 98.47%, 98.41%, and 98.44%, respectively.

CONCLUSION: This study further improves the accuracy and real-time performance of recognizing multi-level mental fatigue based on electrocardiography, and provides theoretical support for real-time fatigue monitoring in daily life.

PMID:39031407 | DOI:10.3233/THC-240129

Categories: Literature Watch

Identifying Synergistic Components of Botanical Fungicide Formulations Using Interpretable Graph Neural Networks

Sat, 2024-07-20 06:00

J Chem Inf Model. 2024 Jul 20. doi: 10.1021/acs.jcim.4c00128. Online ahead of print.

ABSTRACT

Botanical formulations are promising candidates for developing new biopesticides that can protect crops from pests and diseases while reducing harm to the environment. These biopesticides can be combined with permeation enhancer compounds to boost their efficacy against pests and fungal diseases. However, finding synergistic combinations of these compounds is challenging due to the large and complex chemical space. In this paper, we propose a novel deep learning method that can predict the synergy of botanical products and permeation enhancers based on in vitro assay data. Our method uses a weighted combination of component feature vectors to represent the input mixtures, which enables the model to handle a variable number of components and to interpret the contribution of each component to the synergy. We also employ an ensemble of interpretation methods to provide insights into the underlying mechanisms of synergy. We validate our method by testing the predicted synergistic combinations in wet-lab experiments and show that our method can discover novel and effective biopesticides that would otherwise be difficult to find. Our method is generalizable and applicable to other domains, where predicting mixtures of chemical compounds is important.

PMID:39031079 | DOI:10.1021/acs.jcim.4c00128

Categories: Literature Watch

Advancements and Challenges in the Integration of Indium Arsenide and Van der Waals Heterostructures

Sat, 2024-07-20 06:00

Small. 2024 Jul 19:e2403129. doi: 10.1002/smll.202403129. Online ahead of print.

ABSTRACT

The strategic integration of low-dimensional InAs-based materials and emerging van der Waals systems is advancing in various scientific fields, including electronics, optics, and magnetics. With their unique properties, these InAs-based van der Waals materials and devices promise further miniaturization of semiconductor devices in line with Moore's Law. However, progress in this area lags behind other 2D materials like graphene and boron nitride. Challenges include synthesizing pure crystalline phase InAs nanostructures and single-atomic-layer 2D InAs films, both vital for advanced van der Waals heterostructures. Also, diverse surface state effects on InAs-based van der Waals devices complicate their performance evaluation. This review discusses the experimental advances in the van der Waals epitaxy of InAs-based materials and the working principles of InAs-based van der Waals devices. Theoretical achievements in understanding and guiding the design of InAs-based van der Waals systems are highlighted. Focusing on advancing novel selective area growth and remote epitaxy, exploring multi-functional applications, and incorporating deep learning into first-principles calculations are proposed. These initiatives aim to overcome existing bottlenecks and accelerate transformative advancements in integrating InAs and van der Waals heterostructures.

PMID:39030967 | DOI:10.1002/smll.202403129

Categories: Literature Watch

Gtie-Rt: A Comprehensive Graph Learning Model for Predicting Drugs Targeting Metabolic Pathways in Human

Sat, 2024-07-20 06:00

J Bioinform Comput Biol. 2024 Jul 20:2450010. doi: 10.1142/S0219720024500100. Online ahead of print.

ABSTRACT

Drugs often target specific metabolic pathways to produce a therapeutic effect. However, these pathways are complex and interconnected, making it challenging to predict a drug's potential effects on an organism's overall metabolism. The mapping of drugs with targeting metabolic pathways in the organisms can provide a more complete understanding of the metabolic effects of a drug and help to identify potential drug-drug interactions. In this study, we proposed a machine learning hybrid model Graph Transformer Integrated Encoder (GTIE-RT) for mapping drugs to target metabolic pathways in human. The proposed model is a composite of a Graph Convolution Network (GCN) and transformer encoder for graph embedding and attention mechanism. The output of the transformer encoder is then fed into the Extremely Randomized Trees Classifier to predict target metabolic pathways. The evaluation of the GTIE-RT on drugs dataset demonstrates excellent performance metrics, including accuracy (>95%), recall (>92%), precision (>93%) and F1-score (>92%). Compared to other variants and machine learning methods, GTIE-RT consistently shows more reliable results.

PMID:39030668 | DOI:10.1142/S0219720024500100

Categories: Literature Watch

STC-UNet: renal tumor segmentation based on enhanced feature extraction at different network levels

Fri, 2024-07-19 06:00

BMC Med Imaging. 2024 Jul 19;24(1):179. doi: 10.1186/s12880-024-01359-5.

ABSTRACT

Renal tumors are one of the common diseases of urology, and precise segmentation of these tumors plays a crucial role in aiding physicians to improve diagnostic accuracy and treatment effectiveness. Nevertheless, inherent challenges associated with renal tumors, such as indistinct boundaries, morphological variations, and uncertainties in size and location, segmenting renal tumors accurately remains a significant challenge in the field of medical image segmentation. With the development of deep learning, substantial achievements have been made in the domain of medical image segmentation. However, existing models lack specificity in extracting features of renal tumors across different network hierarchies, which results in insufficient extraction of renal tumor features and subsequently affects the accuracy of renal tumor segmentation. To address this issue, we propose the Selective Kernel, Vision Transformer, and Coordinate Attention Enhanced U-Net (STC-UNet). This model aims to enhance feature extraction, adapting to the distinctive characteristics of renal tumors across various network levels. Specifically, the Selective Kernel modules are introduced in the shallow layers of the U-Net, where detailed features are more abundant. By selectively employing convolutional kernels of different scales, the model enhances its capability to extract detailed features of renal tumors across multiple scales. Subsequently, in the deeper layers of the network, where feature maps are smaller yet contain rich semantic information, the Vision Transformer modules are integrated in a non-patch manner. These assist the model in capturing long-range contextual information globally. Their non-patch implementation facilitates the capture of fine-grained features, thereby achieving collaborative enhancement of global-local information and ultimately strengthening the model's extraction of semantic features of renal tumors. Finally, in the decoder segment, the Coordinate Attention modules embedding positional information are proposed aiming to enhance the model's feature recovery and tumor region localization capabilities. Our model is validated on the KiTS19 dataset, and experimental results indicate that compared to the baseline model, STC-UNet shows improvements of 1.60%, 2.02%, 2.27%, 1.18%, 1.52%, and 1.35% in IoU, Dice, Accuracy, Precision, Recall, and F1-score, respectively. Furthermore, the experimental results demonstrate that the proposed STC-UNet method surpasses other advanced algorithms in both visual effectiveness and objective evaluation metrics.

PMID:39030510 | DOI:10.1186/s12880-024-01359-5

Categories: Literature Watch

Explainable lung cancer classification with ensemble transfer learning of VGG16, Resnet50 and InceptionV3 using grad-cam

Fri, 2024-07-19 06:00

BMC Med Imaging. 2024 Jul 19;24(1):176. doi: 10.1186/s12880-024-01345-x.

ABSTRACT

Medical imaging stands as a critical component in diagnosing various diseases, where traditional methods often rely on manual interpretation and conventional machine learning techniques. These approaches, while effective, come with inherent limitations such as subjectivity in interpretation and constraints in handling complex image features. This research paper proposes an integrated deep learning approach utilizing pre-trained models-VGG16, ResNet50, and InceptionV3-combined within a unified framework to improve diagnostic accuracy in medical imaging. The method focuses on lung cancer detection using images resized and converted to a uniform format to optimize performance and ensure consistency across datasets. Our proposed model leverages the strengths of each pre-trained network, achieving a high degree of feature extraction and robustness by freezing the early convolutional layers and fine-tuning the deeper layers. Additionally, techniques like SMOTE and Gaussian Blur are applied to address class imbalance, enhancing model training on underrepresented classes. The model's performance was validated on the IQ-OTH/NCCD lung cancer dataset, which was collected from the Iraq-Oncology Teaching Hospital/National Center for Cancer Diseases over a period of three months in fall 2019. The proposed model achieved an accuracy of 98.18%, with precision and recall rates notably high across all classes. This improvement highlights the potential of integrated deep learning systems in medical diagnostics, providing a more accurate, reliable, and efficient means of disease detection.

PMID:39030496 | DOI:10.1186/s12880-024-01345-x

Categories: Literature Watch

Enhancing global maritime traffic network forecasting with gravity-inspired deep learning models

Fri, 2024-07-19 06:00

Sci Rep. 2024 Jul 19;14(1):16665. doi: 10.1038/s41598-024-67552-2.

ABSTRACT

Aquatic non-indigenous species (NIS) pose significant threats to biodiversity, disrupting ecosystems and inflicting substantial economic damages across agriculture, forestry, and fisheries. Due to the fast growth of global trade and transportation networks, NIS has been introduced and spread unintentionally in new environments. This study develops a new physics-informed model to forecast maritime shipping traffic between port regions worldwide. The predicted information provided by these models, in turn, is used as input for risk assessment of NIS spread through transportation networks to evaluate the capability of our solution. Inspired by the gravity model for international trades, our model considers various factors that influence the likelihood and impact of vessel activities, such as shipping flux density, distance between ports, trade flow, and centrality measures of transportation hubs. Accordingly, this paper introduces transformers to gravity models to rebuild the short- and long-term dependencies that make the risk analysis feasible. Thus, we introduce a physics-inspired framework that achieves an 89% binary accuracy for existing and non-existing trajectories and an 84.8% accuracy for the number of vessels flowing between key port areas, representing more than 10% improvement over the traditional deep-gravity model. Along these lines, this research contributes to a better understanding of NIS risk assessment. It allows policymakers, conservationists, and stakeholders to prioritize management actions by identifying high-risk invasion pathways. Besides, our model is versatile and can include new data sources, making it suitable for assessing international vessel traffic flow in a changing global landscape.

PMID:39030401 | DOI:10.1038/s41598-024-67552-2

Categories: Literature Watch

Learning generalizable AI models for multi-center histopathology image classification

Fri, 2024-07-19 06:00

NPJ Precis Oncol. 2024 Jul 19;8(1):151. doi: 10.1038/s41698-024-00652-4.

ABSTRACT

Investigation of histopathology slides by pathologists is an indispensable component of the routine diagnosis of cancer. Artificial intelligence (AI) has the potential to enhance diagnostic accuracy, improve efficiency, and patient outcomes in clinical pathology. However, variations in tissue preparation, staining protocols, and histopathology slide digitization could result in over-fitting of deep learning models when trained on the data from only one center, thereby underscoring the necessity to generalize deep learning networks for multi-center use. Several techniques, including the use of grayscale images, color normalization techniques, and Adversarial Domain Adaptation (ADA) have been suggested to generalize deep learning algorithms, but there are limitations to their effectiveness and discriminability. Convolutional Neural Networks (CNNs) exhibit higher sensitivity to variations in the amplitude spectrum, whereas humans predominantly rely on phase-related components for object recognition. As such, we propose Adversarial fourIer-based Domain Adaptation (AIDA) which applies the advantages of a Fourier transform in adversarial domain adaptation. We conducted a comprehensive examination of subtype classification tasks in four cancers, incorporating cases from multiple medical centers. Specifically, the datasets included multi-center data for 1113 ovarian cancer cases, 247 pleural cancer cases, 422 bladder cancer cases, and 482 breast cancer cases. Our proposed approach significantly improved performance, achieving superior classification results in the target domain, surpassing the baseline, color augmentation and normalization techniques, and ADA. Furthermore, extensive pathologist reviews suggested that our proposed approach, AIDA, successfully identifies known histotype-specific features. This superior performance highlights AIDA's potential in addressing generalization challenges in deep learning models for multi-center histopathology datasets.

PMID:39030380 | DOI:10.1038/s41698-024-00652-4

Categories: Literature Watch

Exploring deep learning strategies for intervertebral disc herniation detection on veterinary MRI

Fri, 2024-07-19 06:00

Sci Rep. 2024 Jul 19;14(1):16705. doi: 10.1038/s41598-024-67749-5.

ABSTRACT

Intervertebral Disc Herniation (IVDH) is a common spinal disease in dogs, significantly impacting their health, mobility, and overall well-being. This study initiates an effort to automate the detection and localization of IVDH lesions in veterinary MRI scans, utilizing advanced artificial intelligence (AI) methods. A comprehensive canine IVDH dataset, comprising T2-weighted sagittal MRI images from 213 pet dogs of various breeds, ages, and sizes, was compiled and utilized to train and test the IVDH detection models. The experimental results showed that traditional two-stage detection models reliably outperformed one-stage models, including the recent You Only Look Once X (YOLOX) detector. In terms of methodology, this study introduced a novel spinal localization module, successfully integrated into different object detection models to enhance IVDH detection, achieving an average precision (AP) of up to 75.32%. Additionally, transfer learning was explored to adapt the IVDH detection model for a smaller feline dataset. Overall, this study provides insights into advancing AI for veterinary care, identifying challenges and exploring potential strategies for future development in veterinary radiology.

PMID:39030338 | DOI:10.1038/s41598-024-67749-5

Categories: Literature Watch

Absolute permeability estimation from microtomography rock images through deep learning super-resolution and adversarial fine tuning

Fri, 2024-07-19 06:00

Sci Rep. 2024 Jul 19;14(1):16704. doi: 10.1038/s41598-024-67367-1.

ABSTRACT

The carbon capture and storage (CCS) process has become one of the main technologies used for mitigating greenhouse gas emissions. The success of CCS projects relies on accurate subsurface reservoir petrophysical characterization, enabling efficient storage and captured CO 2 containment. In digital rock physics, X-ray microtomography ( μ -CT) is applied to characterize reservoir rocks, allowing a more assertive analysis of physical properties such as porosity and permeability, enabling better simulations of porous media flow. Estimating petrophysical properties through numeric simulations usually requires high-resolution images, which are expensive and time-inefficient to obtain with μ -CT. To address this, we propose using two deep learning models: a super-resolution model to enhance the quality of low-resolution images and a surrogate model that acts as a substitute for numerical simulations to estimate the petrophysical property of interest. A correction process inspired by generative adversarial network (GAN) adversarial training is applied. In this approach, the super-resolution model acts as a generator, creating high-resolution images, and the surrogate network acts as a discriminator. By adjusting the generator, images that correct the errors in the surrogate's estimations are produced. The proposed method was applied to the DeePore dataset. The results shows the proposed approach improved permeability estimation overall.

PMID:39030317 | DOI:10.1038/s41598-024-67367-1

Categories: Literature Watch

Pages