Deep learning

A CT deep learning reconstruction algorithm: Image quality evaluation for brain protocol at decreasing dose indexes in comparison with FBP and statistical iterative reconstruction algorithms

Thu, 2024-02-29 06:00

Phys Med. 2024 Feb 28;119:103319. doi: 10.1016/j.ejmp.2024.103319. Online ahead of print.

ABSTRACT

PURPOSE: To characterise the impact of Precise Image (PI) deep learning reconstruction algorithm on image quality, compared to filtered back-projection (FBP) and iDose4 iterative reconstruction for brain computed tomography (CT) phantom images.

METHODS: Catphan-600 phantom was acquired with an Incisive CT scanner using a dedicated brain protocol, at six different dose levels (volume computed tomography dose index (CTDIvol): 7/14/29/49/56/67 mGy). Images were reconstructed using FBP, levels 2/5 of iDose4, and PI algorithm (Sharper/Sharp/Standard/Smooth/Smoother). Image quality was assessed by evaluating CT numbers, image histograms, noise, image non-uniformity (NU), noise power spectrum, target transfer function, and detectability index.

RESULTS: The five PI levels did not significantly affect the mean CT number. For a given CTDIvol using Sharper-to-Smoother levels, the spatial resolution for all the investigated materials and the detectability index increased while the noise magnitude decreased, slightly affecting noise texture. For a fixed PI level increasing the CTDIvol the detectability index increased, the noise magnitude decreased. From 29 mGy, NU values converged within 1 Hounsfield Unit from each other without a substantial improvement at higher CTDIvol values.

CONCLUSIONS: The improved performances of intermediate PI levels in brain protocols compared to conventional algorithms seem to suggest a potential reduction of CTDIvol.

PMID:38422902 | DOI:10.1016/j.ejmp.2024.103319

Categories: Literature Watch

Semi-supervised iterative adaptive network for low-dose CT sinogram recovery

Thu, 2024-02-29 06:00

Phys Med Biol. 2024 Feb 29. doi: 10.1088/1361-6560/ad2ee7. Online ahead of print.

ABSTRACT

Concern has been expressed regarding the risk of carcinogenesis from medical computed tomography (CT) radiation. Lowering radiation in CT without appropriate modifications often leads to severe noise-induced artifacts in the images. The utilization of Deep Learning (DL) techniques has achieved promising reconstruction performance in low-dose CT (LDCT) imaging. However, most DL-based algorithms require the pre-collection of a large set of image pairs (low-dose/standard-dose) and the training of networks in an end-to-end supervised manner. Meanwhile, securing such a large volume of paired, well-registered training data in clinical practice is challenging. Moreover, these algorithms often overlook the potential to utilize the abundant information in a large collection of LDCT-only images/sinograms. In this paper, we introduce a Semi-supervised Iterative Adaptive Network (SIA-Net) for LDCT imaging, utilizing both labeled and unlabeled sinograms in a cohesive network framework, integrating supervised and unsupervised learning processes. Specifically, the supervised process captures critical features (i.e., noise distribution and tissue characteristics) latent in the paired sinograms, while the unsupervised process effectively learns these features in the unlabeled low-dose sinograms, employing a conventional weighted least-squares model with a regularization term. Furthermore, the SIA-Net method is designed to adaptively transfer the learned feature distribution from the supervised to the unsupervised process, thereby obtaining a high-fidelity sinogram through iterative adaptive learning. Finally, high-quality CT images can be reconstructed from the refined sinogram using the filtered back-projection algorithm. Experimental results on two clinical datasets indicate that the proposed SIA-Net method achieves competitive performance in terms of noise reduction and structure preservation in LDCT imaging, when compared to traditional supervised learning methods.

PMID:38422540 | DOI:10.1088/1361-6560/ad2ee7

Categories: Literature Watch

DeepD3, an open framework for automated quantification of dendritic spines

Thu, 2024-02-29 06:00

PLoS Comput Biol. 2024 Feb 29;20(2):e1011774. doi: 10.1371/journal.pcbi.1011774. eCollection 2024 Feb.

ABSTRACT

Dendritic spines are the seat of most excitatory synapses in the brain, and a cellular structure considered central to learning, memory, and activity-dependent plasticity. The quantification of dendritic spines from light microscopy data is usually performed by humans in a painstaking and error-prone process. We found that human-to-human variability is substantial (inter-rater reliability 82.2±6.4%), raising concerns about the reproducibility of experiments and the validity of using human-annotated 'ground truth' as an evaluation method for computational approaches of spine identification. To address this, we present DeepD3, an open deep learning-based framework to robustly quantify dendritic spines in microscopy data in a fully automated fashion. DeepD3's neural networks have been trained on data from different sources and experimental conditions, annotated and segmented by multiple experts and they offer precise quantification of dendrites and dendritic spines. Importantly, these networks were validated in a number of datasets on varying acquisition modalities, species, anatomical locations and fluorescent indicators. The entire DeepD3 open framework, including the fully segmented training data, a benchmark that multiple experts have annotated, and the DeepD3 model zoo is fully available, addressing the lack of openly available datasets of dendritic spines while offering a ready-to-use, flexible, transparent, and reproducible spine quantification method.

PMID:38422112 | DOI:10.1371/journal.pcbi.1011774

Categories: Literature Watch

Deep learning segmentation and registration-driven lung parenchymal volume and movement CT analysis in prone positioning

Thu, 2024-02-29 06:00

PLoS One. 2024 Feb 29;19(2):e0299366. doi: 10.1371/journal.pone.0299366. eCollection 2024.

ABSTRACT

PURPOSE: To conduct a volumetric and movement analysis of lung parenchyma in prone positioning using deep neural networks (DNNs).

METHOD: We included patients with suspected interstitial lung abnormalities or disease who underwent full-inspiratory supine and prone chest CT at a single institution between June 2021 and March 2022. A thoracic radiologist visually assessed the fibrosis extent in the total lung (using units of 10%) on supine CT. After preprocessing the images into 192×192×192 resolution, a DNN automatically segmented the whole lung and pulmonary lobes in prone and supine CT images. Affine registration matched the patient's center and location, and the DNN deformably registered prone and supine CT images to calculate the x-, y-, z-axis, and 3D pixel movements.

RESULTS: In total, 108 CT pairs had successful registration. Prone positioning significantly increased the left lower (90.2±69.5 mL, P = 0.000) and right lower lobar volumes (52.5±74.2 mL, P = 0.000). During deformable registration, the average maximum whole-lung pixel movements between the two positions were 1.5, 1.9, 1.6, and 2.8 cm in each axis and 3D plane. Compared to patients with <30% fibrosis, those with ≥30% fibrosis had smaller volume changes (P<0.001) and smaller pixel movements in all axes between the positions (P = 0.000-0.007). Forced vital capacity (FVC) correlated with the left lower lobar volume increase (Spearman correlation coefficient, 0.238) and the maximum whole-lung pixel movements in all axes (coefficients, 0.311 to 0.357).

CONCLUSIONS: Prone positioning led to the preferential expansion of the lower lobes, correlated with FVC, and lung fibrosis limited lung expansion during prone positioning.

PMID:38422097 | DOI:10.1371/journal.pone.0299366

Categories: Literature Watch

Can using a pre-trained deep learning model as the feature extractor in the bag-of-deep-visual-words model always improve image classification accuracy?

Thu, 2024-02-29 06:00

PLoS One. 2024 Feb 29;19(2):e0298228. doi: 10.1371/journal.pone.0298228. eCollection 2024.

ABSTRACT

This article investigates whether higher classification accuracy can always be achieved by utilizing a pre-trained deep learning model as the feature extractor in the Bag-of-Deep-Visual-Words (BoDVW) classification model, as opposed to directly using the new classification layer of the pre-trained model for classification. Considering the multiple factors related to the feature extractor -such as model architecture, fine-tuning strategy, number of training samples, feature extraction method, and feature encoding method-we investigate these factors through experiments and then provide detailed answers to the question. In our experiments, we use five feature encoding methods: hard-voting, soft-voting, locally constrained linear coding, super vector coding, and fisher vector (FV). We also employ two popular feature extraction methods: one (denoted as Ext-DFs(CP)) uses a convolutional or non-global pooling layer, and another (denoted as Ext-DFs(FC)) uses a fully-connected or global pooling layer. Three pre-trained models-VGGNet-16, ResNext-50(32×4d), and Swin-B-are utilized as feature extractors. Experimental results on six datasets (15-Scenes, TF-Flowers, MIT Indoor-67, COVID-19 CXR, NWPU-RESISC45, and Caltech-101) reveal that compared to using the pre-trained model with only the new classification layer re-trained for classification, employing it as the feature extractor in the BoDVW model improves the accuracy in 35 out of 36 experiments when using FV. With Ext-DFs(CP), the accuracy increases by 0.13% to 8.43% (averaged at 3.11%), and with Ext-DFs(FC), it increases by 1.06% to 14.63% (averaged at 5.66%). Furthermore, when all layers of the pre-trained model are fine-tuned and used as the feature extractor, the results vary depending on the methods used. If FV and Ext-DFs(FC) are used, the accuracy increases by 0.21% to 5.65% (averaged at 1.58%) in 14 out of 18 experiments. Our results suggest that while using a pre-trained deep learning model as the feature extractor does not always improve classification accuracy, it holds great potential as an accuracy improvement technique.

PMID:38422007 | DOI:10.1371/journal.pone.0298228

Categories: Literature Watch

Deep learning-based prediction of major arrhythmic events in dilated cardiomyopathy: A proof of concept study

Thu, 2024-02-29 06:00

PLoS One. 2024 Feb 29;19(2):e0297793. doi: 10.1371/journal.pone.0297793. eCollection 2024.

ABSTRACT

Prediction of major arrhythmic events (MAEs) in dilated cardiomyopathy represents an unmet clinical goal. Computational models and artificial intelligence (AI) are new technological tools that could offer a significant improvement in our ability to predict MAEs. In this proof-of-concept study, we propose a deep learning (DL)-based model, which we termed Deep ARrhythmic Prevention in dilated cardiomyopathy (DARP-D), built using multidimensional cardiac magnetic resonance data (cine videos and hypervideos and LGE images and hyperimages) and clinical covariates, aimed at predicting and tracking an individual patient's risk curve of MAEs (including sudden cardiac death, cardiac arrest due to ventricular fibrillation, sustained ventricular tachycardia lasting ≥30 s or causing haemodynamic collapse in <30 s, appropriate implantable cardiac defibrillator intervention) over time. The model was trained and validated in 70% of a sample of 154 patients with dilated cardiomyopathy and tested in the remaining 30%. DARP-D achieved a 95% CI in Harrell's C concordance indices of 0.12-0.68 on the test set. We demonstrate that our DL approach is feasible and represents a novelty in the field of arrhythmic risk prediction in dilated cardiomyopathy, able to analyze cardiac motion, tissue characteristics, and baseline covariates to predict an individual patient's risk curve of major arrhythmic events. However, the low number of patients, MAEs and epoch of training make the model a promising prototype but not ready for clinical usage. Further research is needed to improve, stabilize and validate the performance of the DARP-D to convert it from an AI experiment to a daily used tool.

PMID:38421987 | DOI:10.1371/journal.pone.0297793

Categories: Literature Watch

Randomness Regularization with Simple Consistency Training for Neural Networks

Thu, 2024-02-29 06:00

IEEE Trans Pattern Anal Mach Intell. 2024 Feb 29;PP. doi: 10.1109/TPAMI.2024.3370716. Online ahead of print.

ABSTRACT

Randomness is widely introduced in neural network training to simplify model optimization or avoid the over-fitting problem. Among them, dropout and its variations in different aspects (e.g., data, model structure) are prevalent in regularizing the training of deep neural networks. Though effective and performing well, the randomness introduced by these dropout-based methods causes nonnegligible inconsistency between training and inference. In this paper, we introduce a simple consistency training strategy to regularize such randomness, namely R-Drop, which forces two output distributions sampled by each type of randomness to be consistent. Specifically, R-Drop minimizes the bidirectional KL-divergence between two output distributions produced by dropout-based randomness for each training sample. Theoretical analysis reveals that R-Drop can reduce the above inconsistency by reducing the inconsistency among the sampled sub structures and bridging the gap between the loss calculated by the full model and sub structures. Experiments on 7 widely-used deep learning tasks ( 23 datasets in total) demonstrate that R-Drop is universally effective for different types of neural networks (i.e., feed-forward, recurrent, and graph neural networks) and different learning paradigms (supervised, parameter-efficient, and semi-supervised). In particular, it achieves state-of-the-art performances with the vanilla Transformer model on WMT14 English → German translation ( 30.91 BLEU) and WMT14 English → French translation ( 43.95 BLEU), even surpassing models trained with extra large-scale data and expert-designed advanced variants of Transformer models. Our code is available at GitHub https://github.com/dropreg/R-Drop.

PMID:38421846 | DOI:10.1109/TPAMI.2024.3370716

Categories: Literature Watch

Deep learning for enhanced prosthetic control: Real-time motor intent decoding for simultaneous control of artificial limbs

Thu, 2024-02-29 06:00

IEEE Trans Neural Syst Rehabil Eng. 2024 Feb 29;PP. doi: 10.1109/TNSRE.2024.3371896. Online ahead of print.

ABSTRACT

The development of advanced prosthetic devices that can be seamlessly used during an individual's daily life remains a significant challenge in the field of rehabilitation engineering. This study compares the performance of deep learning architectures to shallow networks in decoding motor intent for prosthetic control using electromyography (EMG) signals. Four neural network architectures, including a feedforward neural network with one hidden layer, a feedforward neural network with multiple hidden layers, a temporal convolutional network, and a convolutional neural network with squeeze-and-excitation operations were evaluated in real-time, human-in-the-loop experiments with able-bodied participants and an individual with an amputation. Our results demonstrate that deep learning architectures outperform shallow networks in decoding motor intent, with representation learning effectively extracting underlying motor control information from EMG signals. Furthermore, the observed performance improvements by using deep neural networks were consistent across both able-bodied and amputee participants. By employing deep neural networks instead of a shallow network, more reliable and precise control of a prosthesis can be achieved, which has the potential to significantly enhance prosthetic functionality and improve the quality of life for individuals with amputations.

PMID:38421839 | DOI:10.1109/TNSRE.2024.3371896

Categories: Literature Watch

Segmentation, feature extraction and classification of leukocytes leveraging neural networks, a comparative study

Thu, 2024-02-29 06:00

Cytometry A. 2024 Feb 29. doi: 10.1002/cyto.a.24832. Online ahead of print.

ABSTRACT

The gold standard of leukocyte differentiation is a manual examination of blood smears, which is not only time and labor intensive but also susceptible to human error. As to automatic classification, there is still no comparative study of cell segmentation, feature extraction, and cell classification, where a variety of machine and deep learning models are compared with home-developed approaches. In this study, both traditional machine learning of K-means clustering versus deep learning of U-Net, U-Net + ResNet18, and U-Net + ResNet34 were used for cell segmentation, producing segmentation accuracies of 94.36% versus 99.17% for the dataset of CellaVision and 93.20% versus 98.75% for the dataset of BCCD, confirming that deep learning produces higher performance than traditional machine learning in leukocyte classification. In addition, a series of deep-learning approaches, including AlexNet, VGG16, and ResNet18, was adopted to conduct feature extraction and cell classification of leukocytes, producing classification accuracies of 91.31%, 97.83%, and 100% of CellaVision as well as 81.18%, 91.64% and 97.82% of BCCD, confirming the capability of the increased deepness of neural networks in leukocyte classification. As to the demonstrations, this study further conducted cell-type classification of ALL-IDB2 and PCB-HBC datasets, producing high accuracies of 100% and 98.49% among all literature, validating the deep learning model used in this study.

PMID:38420862 | DOI:10.1002/cyto.a.24832

Categories: Literature Watch

Deep-Learning Uncovers certain CCM Isoforms as Transcription Factors

Thu, 2024-02-29 06:00

Front Biosci (Landmark Ed). 2024 Feb 21;29(2):75. doi: 10.31083/j.fbl2902075.

ABSTRACT

BACKGROUND: Cerebral Cavernous Malformations (CCMs) are brain vascular abnormalities associated with an increased risk of hemorrhagic strokes. Familial CCMs result from autosomal dominant inheritance involving three genes: KRIT1 (CCM1), MGC4607 (CCM2), and PDCD10 (CCM3). CCM1 and CCM3 form the CCM Signal Complex (CSC) by binding to CCM2. Both CCM1 and CCM2 exhibit cellular heterogeneity through multiple alternative spliced isoforms, where exons from the same gene combine in diverse ways, leading to varied mRNA transcripts. Additionally, both demonstrate nucleocytoplasmic shuttling between the nucleus and cytoplasm, suggesting their potential role in gene expression regulation as transcription factors (TFs). Due to the accumulated data indicating the cellular localization of CSC proteins in the nucleus and their interaction with progesterone receptors, which serve dual roles as both cellular signaling components and TFs, a question has arisen regarding whether CCMs could also function in both capacities like progesterone receptors.

METHODS: To investigate this potential, we employed our proprietary deep-learning (DL)-based algorithm, specifically utilizing a biased-Support Vector Machine (SVM) model, to explore the plausible cellular function of any of the CSC proteins, particularly focusing on CCM gene isoforms with nucleocytoplasmic shuttling, acting as TFs in gene expression regulation.

RESULTS: Through a comparative DL-based predictive analysis, we have effectively discerned a collective of 11 isoforms across all CCM proteins (CCM1-3). Additionally, we have substantiated the TF functionality of 8 isoforms derived from CCM1 and CCM2 proteins, marking the inaugural identification of CCM isoforms in the role of TFs.

CONCLUSIONS: This groundbreaking discovery directly challenges the prevailing paradigm, which predominantly emphasizes the involvement of CSC solely in endothelial cellular functions amid various potential cellular signal cascades during angiogenesis.

PMID:38420834 | DOI:10.31083/j.fbl2902075

Categories: Literature Watch

aiGeneR 1.0: An Artificial Intelligence Technique for the Revelation of Informative and Antibiotic Resistant Genes in <em>Escherichia coli</em>

Thu, 2024-02-29 06:00

Front Biosci (Landmark Ed). 2024 Feb 22;29(2):82. doi: 10.31083/j.fbl2902082.

ABSTRACT

BACKGROUND: There are several antibiotic resistance genes (ARG) for the Escherichia coli (E. coli) bacteria that cause urinary tract infections (UTI), and it is therefore important to identify these ARG. Artificial Intelligence (AI) has been used previously in the field of gene expression data, but never adopted for the detection and classification of bacterial ARG. We hypothesize, if the data is correctly conferred, right features are selected, and Deep Learning (DL) classification models are optimized, then (i) non-linear DL models would perform better than Machine Learning (ML) models, (ii) leads to higher accuracy, (iii) can identify the hub genes, and, (iv) can identify gene pathways accurately. We have therefore designed aiGeneR, the first of its kind system that uses DL-based models to identify ARG in E. coli in gene expression data.

METHODOLOGY: The aiGeneR consists of a tandem connection of quality control embedded with feature extraction and AI-based classification of ARG. We adopted a cross-validation approach to evaluate the performance of aiGeneR using accuracy, precision, recall, and F1-score. Further, we analyzed the effect of sample size ensuring generalization of models and compare against the power analysis. The aiGeneR was validated scientifically and biologically for hub genes and pathways. We benchmarked aiGeneR against two linear and two other non-linear AI models.

RESULTS: The aiGeneR identifies tetM (an ARG) and showed an accuracy of 93% with area under the curve (AUC) of 0.99 (p < 0.05). The mean accuracy of non-linear models was 22% higher compared to linear models. We scientifically and biologically validated the aiGeneR.

CONCLUSIONS: aiGeneR successfully detected the E. coli genes validating our four hypotheses.

PMID:38420832 | DOI:10.31083/j.fbl2902082

Categories: Literature Watch

Cross-modality Labeling Enables Noninvasive Capillary Quantification as a Sensitive Biomarker for Assessing Cardiovascular Risk

Thu, 2024-02-29 06:00

Ophthalmol Sci. 2023 Dec 5;4(3):100441. doi: 10.1016/j.xops.2023.100441. eCollection 2024 May-Jun.

ABSTRACT

PURPOSE: We aim to use fundus fluorescein angiography (FFA) to label the capillaries on color fundus (CF) photographs and train a deep learning model to quantify retinal capillaries noninvasively from CF and apply it to cardiovascular disease (CVD) risk assessment.

DESIGN: Cross-sectional and longitudinal study.

PARTICIPANTS: A total of 90732 pairs of CF-FFA images from 3893 participants for segmentation model development, and 49229 participants in the UK Biobank for association analysis.

METHODS: We matched the vessels extracted from FFA and CF, and used vessels from FFA as labels to train a deep learning model (RMHAS-FA) to segment retinal capillaries using CF. We tested the model's accuracy on a manually labeled internal test set (FundusCapi). For external validation, we tested the segmentation model on 7 vessel segmentation datasets, and investigated the clinical value of the segmented vessels in predicting CVD events in the UK Biobank.

MAIN OUTCOME MEASURES: Area under the receiver operating characteristic curve (AUC), accuracy, sensitivity, and specificity for segmentation. Hazard ratio (HR; 95% confidence interval [CI]) for Cox regression analysis.

RESULTS: On the FundusCapi dataset, the segmentation performance was AUC = 0.95, accuracy = 0.94, sensitivity = 0.90, and specificity = 0.93. Smaller vessel skeleton density had a stronger correlation with CVD risk factors and incidence (P < 0.01). Reduced density of small vessel skeletons was strongly associated with an increased risk of CVD incidence and mortality for women (HR [95% CI] = 0.91 [0.84-0.98] and 0.68 [0.54-0.86], respectively).

CONCLUSIONS: Using paired CF-FFA images, we automated the laborious manual labeling process and enabled noninvasive capillary quantification from CF, supporting its potential as a sensitive screening method for identifying individuals at high risk of future CVD events.

FINANCIAL DISCLOSURES: Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.

PMID:38420613 | PMC:PMC10899028 | DOI:10.1016/j.xops.2023.100441

Categories: Literature Watch

Computational pathology: A survey review and the way forward

Thu, 2024-02-29 06:00

J Pathol Inform. 2024 Jan 14;15:100357. doi: 10.1016/j.jpi.2023.100357. eCollection 2024 Dec.

ABSTRACT

Computational Pathology (CPath) is an interdisciplinary science that augments developments of computational approaches to analyze and model medical histopathology images. The main objective for CPath is to develop infrastructure and workflows of digital diagnostics as an assistive CAD system for clinical pathology, facilitating transformational changes in the diagnosis and treatment of cancer that are mainly address by CPath tools. With evergrowing developments in deep learning and computer vision algorithms, and the ease of the data flow from digital pathology, currently CPath is witnessing a paradigm shift. Despite the sheer volume of engineering and scientific works being introduced for cancer image analysis, there is still a considerable gap of adopting and integrating these algorithms in clinical practice. This raises a significant question regarding the direction and trends that are undertaken in CPath. In this article we provide a comprehensive review of more than 800 papers to address the challenges faced in problem design all-the-way to the application and implementation viewpoints. We have catalogued each paper into a model-card by examining the key works and challenges faced to layout the current landscape in CPath. We hope this helps the community to locate relevant works and facilitate understanding of the field's future directions. In a nutshell, we oversee the CPath developments in cycle of stages which are required to be cohesively linked together to address the challenges associated with such multidisciplinary science. We overview this cycle from different perspectives of data-centric, model-centric, and application-centric problems. We finally sketch remaining challenges and provide directions for future technical developments and clinical integration of CPath. For updated information on this survey review paper and accessing to the original model cards repository, please refer to GitHub. Updated version of this draft can also be found from arXiv.

PMID:38420608 | PMC:PMC10900832 | DOI:10.1016/j.jpi.2023.100357

Categories: Literature Watch

An interpretable deep learning model for identifying the morphological characteristics of dMMR/MSI-H gastric cancer

Thu, 2024-02-29 06:00

iScience. 2024 Feb 16;27(3):109243. doi: 10.1016/j.isci.2024.109243. eCollection 2024 Mar 15.

ABSTRACT

Accurate tumor diagnosis by pathologists relies on identifying specific morphological characteristics. However, summarizing these unique morphological features in tumor classifications can be challenging. Although deep learning models have been extensively studied for tumor classification, their indirect and subjective interpretation obstructs pathologists from comprehending the model and discerning the morphological features accountable for classifications. In this study, we introduce a new approach utilizing Style Generative Adversarial Networks, which enables a direct interpretation of deep learning models to detect significant morphological characteristics within datasets representing patients with deficient mismatch repair/microsatellite instability-high gastric cancer. Our approach effectively identifies distinct morphological features crucial for tumor classification, offering valuable insights for pathologists to enhance diagnostic accuracy and foster professional growth.

PMID:38420592 | PMC:PMC10901137 | DOI:10.1016/j.isci.2024.109243

Categories: Literature Watch

Image synthesis of apparel stitching defects using deep convolutional generative adversarial networks

Thu, 2024-02-29 06:00

Heliyon. 2024 Feb 15;10(4):e26466. doi: 10.1016/j.heliyon.2024.e26466. eCollection 2024 Feb 29.

ABSTRACT

In industrial manufacturing, the detection of stitching defects in fabric has become a pivotal stage in ensuring product quality. Deep learning-based fabric defect detection models have demonstrated remarkable accuracy, but they often require a vast amount of training data. Unfortunately, practical production lines typically lack a sufficient quantity of apparel stitching defect images due to limited research-industry collaboration and privacy concerns. To address this challenge, this study introduces an innovative approach based on DCGAN (Deep Convolutional Generative Adversarial Network), enabling the automatic generation of stitching defects in fabric. The evaluation encompasses both quantitative and qualitative assessments, supported by extensive comparative experiments. For validation of results, ten industrial experts marked 80% accuracy of the generated images. Moreover, Fréchet Inception Distance also inferred promising results. The outcomes, marked by high accuracy rate, underscore the effectiveness of proposed defect generation model. It demonstrates the ability to produce realistic stitching defective data, bridging the gap caused by data scarcity in practical industrial settings.

PMID:38420437 | PMC:PMC10900799 | DOI:10.1016/j.heliyon.2024.e26466

Categories: Literature Watch

A brief review and scientometric analysis on ensemble learning methods for handling COVID-19

Thu, 2024-02-29 06:00

Heliyon. 2024 Feb 20;10(4):e26694. doi: 10.1016/j.heliyon.2024.e26694. eCollection 2024 Feb 29.

ABSTRACT

Numerous efforts and research have been conducted worldwide to combat the coronavirus disease 2019 (COVID-19) pandemic. In this regard, some researchers have focused on deep and machine-learning approaches to discover more about this disease. There have been many articles on using ensemble learning methods for COVID-19 detection. Still, there seems to be no scientometric analysis or a brief review of these researches. Hence, a combined method of scientometric analysis and brief review was used to study the published articles that employed an ensemble learning approach to detect COVID-19. This research used both methods to overcome their limitations, leading to enhanced and reliable outcomes. The related articles were retrieved from the Scopus database. Then a two-step procedure was employed. A concise review of the collected articles was conducted. Then they underwent scientometric and bibliometric analyses. The findings revealed that convolutional neural network (CNN) is the mostly employed algorithm, while support vector machine (SVM), random forest, Resnet, DenseNet, and visual geometry group (VGG) were also frequently used. Additionally, China has had a significant presence in the numerous top-ranking categories of this field of research. Both study phases yielded valuable results and rankings.

PMID:38420425 | PMC:PMC10901105 | DOI:10.1016/j.heliyon.2024.e26694

Categories: Literature Watch

Lung disease recognition methods using audio-based analysis with machine learning

Thu, 2024-02-29 06:00

Heliyon. 2024 Feb 17;10(4):e26218. doi: 10.1016/j.heliyon.2024.e26218. eCollection 2024 Feb 29.

ABSTRACT

The use of computer-based automated approaches and improvements in lung sound recording techniques have made lung sound-based diagnostics even better and devoid of subjectivity errors. Using a computer to evaluate lung sound features more thoroughly with the use of analyzing changes in lung sound behavior, recording measurements, suppressing the presence of noise contaminations, and graphical representations are all made possible by computer-based lung sound analysis. This paper starts with a discussion of the need for this research area, providing an overview of the field and the motivations behind it. Following that, it details the survey methodology used in this work. It presents a discussion on the elements of sound-based lung disease classification using machine learning algorithms. This includes commonly prior considered datasets, feature extraction techniques, pre-processing methods, artifact removal methods, lung-heart sound separation, deep learning algorithms, and wavelet transform of lung audio signals. The study introduces studies that review lung screening including a summary table of these references and discusses the literature gaps in the existing studies. It is concluded that the use of sound-based machine learning in the classification of respiratory diseases has promising results. While we believe this material will prove valuable to physicians and researchers exploring sound-signal-based machine learning, large-scale investigations remain essential to solidify the findings and foster wider adoption within the medical community.

PMID:38420389 | PMC:PMC10900411 | DOI:10.1016/j.heliyon.2024.e26218

Categories: Literature Watch

Learning multi-site harmonization of magnetic resonance images without traveling human phantoms

Thu, 2024-02-29 06:00

Commun Eng. 2024;3:6. doi: 10.1038/s44172-023-00140-w. Epub 2024 Jan 5.

ABSTRACT

Harmonization improves Magn. Reson. Imaging (MRI) data consistency and is central to effective integration of diverse imaging data acquired across multiple sites. Recent deep learning techniques for harmonization are predominantly supervised in nature and hence require imaging data of the same human subjects to be acquired at multiple sites. Data collection as such requires the human subjects to travel across sites and is hence challenging, costly, and impractical, more so when sufficient sample size is needed for reliable network training. Here we show how harmonization can be achieved with a deep neural network that does not rely on traveling human phantom data. Our method disentangles site-specific appearance information and site-invariant anatomical information from images acquired at multiple sites and then employs the disentangled information to generate the image of each subject for any target site. We demonstrate with more than 6,000 multi-site T1- and T2-weighted images that our method is remarkably effective in generating images with realistic site-specific appearances without altering anatomical details. Our method allows retrospective harmonization of data in a wide range of existing modern large-scale imaging studies, conducted via different scanners and protocols, without additional data collection.

PMID:38420332 | PMC:PMC10898625 | DOI:10.1038/s44172-023-00140-w

Categories: Literature Watch

VGG19 demonstrates the highest accuracy rate in a nine-class wound classification task among various deep learning networks: a pilot study

Wed, 2024-02-28 06:00

Wounds. 2024 Jan;36(1):8-14.

ABSTRACT

BACKGROUND: Current literature suggests relatively low accuracy of multi-class wound classification tasks using deep learning networks. Solutions are needed to address the increasing diagnostic burden of wounds on wound care professionals and to aid non-wound care professionals in wound management.

OBJECTIVE: To develop a reliable, accurate 9-class classification system to aid wound care professionals and perhaps eventually, patients and non-wound care professionals, in managing wounds.

METHODS: A total of 8173 training data images and 904 test data images were classified into 9 categories: operation wound, laceration, abrasion, skin defect, infected wound, necrosis, diabetic foot ulcer, chronic ulcer, and wound dehiscence. Six deep learning networks, based on VGG16, VGG19, EfficientNet-B0, EfficientNet-B5, RepVGG-A0, and RepVGG-B0, were established, trained, and tested on the same images. For each network the accuracy rate, defined as the sum of true positive and true negative values divided by the total number, was analyzed.

RESULTS: The overall accuracy varied from 74.0% to 82.4%. Of all the networks, VGG19 achieved the highest accuracy, at 82.4%. This result is comparable to those reported in previous studies.

CONCLUSION: These findings indicate the potential for VGG19 to be the basis for a more comprehensive and detailed AI-based wound diagnostic system. Eventually, such systems also may aid patients and non-wound care professionals in diagnosing and treating wounds.

PMID:38417818

Categories: Literature Watch

Towards development of functional climate-driven early warning systems for climate-sensitive infectious disease: Statistical models and recommendations

Wed, 2024-02-28 06:00

Environ Res. 2024 Feb 26:118568. doi: 10.1016/j.envres.2024.118568. Online ahead of print.

ABSTRACT

Climate, weather and environmental change have significantly influenced patterns of infectious disease transmission, necessitating the development of early warning systems to anticipate potential impacts and respond in a timely and effective way. Statistical modelling plays a pivotal role in understanding the intricate relationships between climatic factors and infectious disease transmission. For example, time series regression modelling and spatial cluster analysis have been employed to identify risk factors and predict spatial and temporal patterns of infectious diseases. Recently advanced spatio-temporal models and machine learning offer an increasingly robust framework for modelling uncertainty, which is essential in climate-driven disease surveillance due to the dynamic and multifaceted nature of the data. Moreover, Artificial Intelligence (AI) techniques, including deep learning and neural networks, excel in capturing intricate patterns and hidden relationships within climate and environmental data sets. Web-based data has emerged as a powerful complement to other datasets encompassing climate variables and disease occurrences. However, given the complexity and non-linearity of climate-disease interactions, advanced techniques are required to integrate and analyse these diverse data to obtain more accurate predictions of impending outbreaks, epidemics or pandemics. This article presents an overview of an approach to creating climate-driven early warning systems with a focus on statistical model suitability and selection, along with recommendations for utilizing spatio-temporal and machine learning techniques. By addressing the limitations and embracing the recommendations for future research, we could enhance preparedness and response strategies, ultimately contributing to the safeguarding of public health in the face of evolving climate challenges.

PMID:38417659 | DOI:10.1016/j.envres.2024.118568

Categories: Literature Watch

Pages