Deep learning

Radio-opaque contrast agents for liver cancer targeting with KIM during radiation therapy (ROCK-RT): an observational feasibility study

Tue, 2024-10-08 06:00

Radiat Oncol. 2024 Oct 8;19(1):139. doi: 10.1186/s13014-024-02524-4.

ABSTRACT

BACKGROUND: This observational study aims to establish the feasibility of using x-ray images of radio-opaque chemoembolisation deposits in patients as a method for real-time image-guided radiation therapy of hepatocellular carcinoma.

METHODS: This study will recruit 50 hepatocellular carcinoma patients who have had or will have stereotactic ablative radiation therapy and have had transarterial chemoembolisation with a radio-opaque agent. X-ray and computed tomography images of the patients will be analysed retrospectively. Additionally, a deep learning method for real-time motion tracking will be developed. We hypothesise that: (i) deep learning software can be developed that will successfully track the contrast agent mass on two thirds of cone beam computed tomography (CBCT) projection and intra-treatment images (ii), the mean and standard deviation (mm) difference in the location of the mass between ground truth and deep learning detection are ≤ 2 mm and ≤ 3 mm respectively and (iii) statistical modelling of study data will predict tracking success in 85% of trial participants.

DISCUSSION: Developing a real-time tracking method will enable increased targeting accuracy, without the need for additional invasive procedures to implant fiducial markers.

TRIAL REGISTRATION: Registered to ClinicalTrials.gov (NCT05169177) 12th October 2021.

PMID:39380004 | DOI:10.1186/s13014-024-02524-4

Categories: Literature Watch

Prediction of homologous recombination deficiency from routine histology with attention-based multiple instance learning in nine different tumor types

Tue, 2024-10-08 06:00

BMC Biol. 2024 Oct 8;22(1):225. doi: 10.1186/s12915-024-02022-9.

ABSTRACT

BACKGROUND: Homologous recombination deficiency (HRD) is recognized as a pan-cancer predictive biomarker that potentially indicates who could benefit from treatment with PARP inhibitors (PARPi). Despite its clinical significance, HRD testing is highly complex. Here, we investigated in a proof-of-concept study whether Deep Learning (DL) can predict HRD status solely based on routine hematoxylin & eosin (H&E) histology images across nine different cancer types.

METHODS: We developed a deep learning pipeline with attention-weighted multiple instance learning (attMIL) to predict HRD status from histology images. As part of our approach, we calculated a genomic scar HRD score by combining loss of heterozygosity (LOH), telomeric allelic imbalance (TAI), and large-scale state transitions (LST) from whole genome sequencing (WGS) data of n = 5209 patients across two independent cohorts. The model's effectiveness was evaluated using the area under the receiver operating characteristic curve (AUROC), focusing on its accuracy in predicting genomic HRD against a clinically recognized cutoff value.

RESULTS: Our study demonstrated the predictability of genomic HRD status in endometrial, pancreatic, and lung cancers reaching cross-validated AUROCs of 0.79, 0.58, and 0.66, respectively. These predictions generalized well to an external cohort, with AUROCs of 0.93, 0.81, and 0.73. Moreover, a breast cancer-trained image-based HRD classifier yielded an AUROC of 0.78 in the internal validation cohort and was able to predict HRD in endometrial, prostate, and pancreatic cancer with AUROCs of 0.87, 0.84, and 0.67, indicating that a shared HRD-like phenotype occurs across these tumor entities.

CONCLUSIONS: This study establishes that HRD can be directly predicted from H&E slides using attMIL, demonstrating its applicability across nine different tumor types.

PMID:39379982 | DOI:10.1186/s12915-024-02022-9

Categories: Literature Watch

Crossfeat: a transformer-based cross-feature learning model for predicting drug side effect frequency

Tue, 2024-10-08 06:00

BMC Bioinformatics. 2024 Oct 8;25(1):324. doi: 10.1186/s12859-024-05915-2.

ABSTRACT

BACKGROUND: Safe drug treatment requires an understanding of the potential side effects. Identifying the frequency of drug side effects can reduce the risks associated with drug use. However, existing computational methods for predicting drug side effect frequencies heavily depend on known drug side effect frequency information. Consequently, these methods face challenges when predicting the side effect frequencies of new drugs. Although a few methods can predict the side effect frequencies of new drugs, they exhibit unreliable performance owing to the exclusion of drug-side effect relationships.

RESULTS: This study proposed CrossFeat, a model based on convolutional neural network-transformer architecture with cross-feature learning that can predict the occurrence and frequency of drug side effects for new drugs, even in the absence of information regarding drug-side effect relationships. CrossFeat facilitates the concurrent learning of drugs and side effect information within its transformer architecture. This simultaneous exchange of information enables drugs to learn about their associated side effects, while side effects concurrently acquire information about the respective drugs. Such bidirectional learning allows for the comprehensive integration of drug and side effect knowledge. Our five-fold cross-validation experiments demonstrated that CrossFeat outperforms existing studies in predicting side effect frequencies for new drugs without prior knowledge.

CONCLUSIONS: Our model offers a promising approach for predicting the drug side effect frequencies, particularly for new drugs where prior information is limited. CrossFeat's superior performance in cross-validation experiments, along with evidence from case studies and ablation experiments, highlights its effectiveness.

PMID:39379821 | DOI:10.1186/s12859-024-05915-2

Categories: Literature Watch

Automatic Acne Severity Grading with a Small and Imbalanced Data Set of Low-Resolution Images

Tue, 2024-10-08 06:00

Dermatol Ther (Heidelb). 2024 Oct 8. doi: 10.1007/s13555-024-01283-0. Online ahead of print.

ABSTRACT

INTRODUCTION: Developing automatic acne vulgaris grading systems based on machine learning is an expensive endeavor in terms of data acquisition. A machine learning practitioner will need to gather high-resolution pictures from a considerable number of different patients, with a well-balanced distribution between acne severity grades and potentially very tedious labeling. We developed a deep learning model to grade acne severity with respect to the Investigator's Global Assessment (IGA) scale that can be trained on low-resolution images, with pictures from a small number of different patients, a strongly imbalanced severity grade distribution and minimal labeling.

METHODS: A total of 1374 triplets of images (frontal and lateral views) from 391 different patients suffering from acne labeled with the IGA severity grade by an expert dermatologist were used to train and validate a deep learning model that predicts the IGA severity grade.

RESULTS: On the test set we obtained 66.67% accuracy with an equivalent performance for all grades despite the highly imbalanced severity grade distribution of our database. Importantly, we obtained performance on par with more tedious methods in terms of data acquisition which have the same simple labeling as ours but require either a more balanced severity grade distribution or large numbers of high-resolution images.

CONCLUSIONS: Our deep learning model demonstrated promising accuracy despite the limited data set on which it was trained, indicating its potential for further development both as an assistance tool for medical practitioners and as a way to provide patients with an immediately available and standardized acne grading tool.

TRIAL REGISTRATION: chinadrugtrials.org.cn identifier CTR20211314.

PMID:39379778 | DOI:10.1007/s13555-024-01283-0

Categories: Literature Watch

Deep learning model using planar whole-body bone scintigraphy for diagnosis of skull base invasion in patients with nasopharyngeal carcinoma

Tue, 2024-10-08 06:00

J Cancer Res Clin Oncol. 2024 Oct 9;150(10):449. doi: 10.1007/s00432-024-05969-y.

ABSTRACT

PURPOSE: This study assesses the reliability of deep learning models based on planar whole-body bone scintigraphy for diagnosing Skull base invasion (SBI) in nasopharyngeal carcinoma (NPC) patients.

METHODS: In this multicenter study, a deep learning model was developed using data from one center with a 7:3 allocation to training and internal test sets, to diagnose SBI in patients newly diagnosed with NPC using planar whole-body bone scintigraphy. Patients were diagnosed based on a composite reference standard incorporating radiologic and follow-up data. Ten different convolutional neural network (CNN) models were applied to both whole-image and partial-image input modes to determine the optimal model for each analysis. Model performance was assessed using the area under the receiver operating characteristic curve (AUC), calibration, decision curve analysis (DCA), and compared with expert assessments by two nuclear medicine physicians.

RESULTS: The best-performing model using partial-body input achieved AUCs of 0.80 (95% CI: 0.73, 0.86) in the internal test set, 0.84 (95% CI: 0.77, 0.91) in the external cohort, and 0.78 (95% CI: 0.73, 0.83) in the treatment test cohort. Calibration curves and DCA confirmed the models' excellent discrimination, calibration, and potential clinical utility across internal and external datasets. The AUCs of both nuclear medicine physicians were lower than those of the best-performing deep learning model in external test set (AUC: 0.75 vs. 0.77 vs. 0.84).

CONCLUSION: Deep learning models utilizing partial-body input from planar whole-body bone scintigraphy demonstrate high discriminatory power for diagnosing SBI in NPC patients, surpassing experienced nuclear medicine physicians.

PMID:39379746 | DOI:10.1007/s00432-024-05969-y

Categories: Literature Watch

Novel deep learning radiomics nomogram-based multiparametric MRI for predicting the lymph node metastasis in rectal cancer: A dual-center study

Tue, 2024-10-08 06:00

J Cancer Res Clin Oncol. 2024 Oct 9;150(10):450. doi: 10.1007/s00432-024-05986-x.

ABSTRACT

PURPOSE: To develop and evaluate a nomogram that integrates clinical parameters with deep learning radiomics (DLR) extracted from Magnetic Resonance Imaging (MRI) data to enhance the predictive accuracy for preoperative lymph node (LN) metastasis in rectal cancer.

METHODS: A retrospective analysis was conducted on 356 patients diagnosed with rectal cancer. Of these, 286 patients were allocated to the training set, and 70 patients comprised the external validation cohort. Preprocessed T2-weighted and diffusion-weighted imaging performed preoperatively facilitated the extraction of DLR features. Five machine learning algorithms-k-nearest neighbor, light gradient boosting machine, logistic regression, random forest, and support vector machine-were utilized to develop DLR models. The most effective algorithm was identified and used to establish a clinical DLR (CDLR) nomogram specifically designed to predict LN metastasis in rectal cancer. The performance of the nomogram was evaluated using receiver operating characteristic curve analysis.

RESULTS: The logistic regression classifier demonstrated significant predictive accuracy using the DLR signature, achieving an Area Under the Curve (AUC) of 0.919 in the training cohort and 0.778 in the external validation cohort. The integrated CDLR nomogram exhibited robust predictive performance across both datasets, with AUC values of 0.921 in the training cohort and 0.818 in the external validation cohort. Notably, it outperformed both the clinical model, which had AUC values of 0.770 and 0.723 in the training and external validation cohorts, respectively, and the stand-alone DLR model.

CONCLUSION: The nomogram derived from multiparametric MRI data, referred to as the CDLR model, demonstrates strong predictive efficacy in forecasting LN metastasis in rectal cancer.

PMID:39379733 | DOI:10.1007/s00432-024-05986-x

Categories: Literature Watch

Understanding Episode Hardness in Few-Shot Learning

Tue, 2024-10-08 06:00

IEEE Trans Pattern Anal Mach Intell. 2024 Oct 8;PP. doi: 10.1109/TPAMI.2024.3476075. Online ahead of print.

ABSTRACT

Achieving generalization for deep learning models has usually suffered from the bottleneck of annotated sample scarcity. As a common way of tackling this issue, few-shot learning focuses on "episodes", i.e. sampled tasks that help the model acquire generalizable knowledge onto unseen categories - better the episodes, the higher a model's generalisability. Despite extensive research, the characteristics of episodes and their potential effects are relatively less explored. A recent paper discussed that different episodes exhibit different prediction difficulties, and coined a new metric "hardness" to quantify episodes, which however is too wide-range for an arbitrary dataset and thus remains impractical for realistic applications. In this paper therefore, we for the first time conduct an algebraic analysis of the critical factors influencing episode hardness supported by experimental demonstrations, that reveal episode hardness to largely depend on classes within an episode, and importantly propose an efficient pre-sampling hardness assessment technique named Inverse-Fisher Discriminant Ratio (IFDR). This enables sampling hard episodes at the class level via class-level (cl) sampling scheme that drastically decreases quantification cost. Delving deeper, we also develop a variant called class-pair-level (cpl) sampling, which further reduces the sampling cost while guaranteeing the sampled distribution. Finally, comprehensive experiments conducted on benchmark datasets verify the efficacy of our proposed method. Codes are available at: https://github.com/PRIS-CV/class-level-sampling.

PMID:39378258 | DOI:10.1109/TPAMI.2024.3476075

Categories: Literature Watch

GraKerformer: A Transformer With Graph Kernel for Unsupervised Graph Representation Learning

Tue, 2024-10-08 06:00

IEEE Trans Cybern. 2024 Oct 8;PP. doi: 10.1109/TCYB.2024.3465213. Online ahead of print.

ABSTRACT

While highly influential in deep learning, especially in natural language processing, the Transformer model has not exhibited competitive performance in unsupervised graph representation learning (UGRL). Conventional approaches, which focus on local substructures on the graph, offer simplicity but often fall short in encapsulating comprehensive structural information of the graph. This deficiency leads to suboptimal generalization performance. To address this, we proposed the GraKerformer model, a variant of the standard Transformer architecture, to mitigate the shortfall in structural information representation and enhance the performance in UGRL. By leveraging the shortest-path graph kernel (SPGK) to weight attention scores and combining graph neural networks, the GraKerformer effectively encodes the nuanced structural information of graphs. We conducted evaluations on the benchmark datasets for graph classification to validate the superior performance of our approach.

PMID:39378254 | DOI:10.1109/TCYB.2024.3465213

Categories: Literature Watch

Multi-scale Spatio-temporal Memory Network for Lightweight Video denoising

Tue, 2024-10-08 06:00

IEEE Trans Image Process. 2024 Oct 8;PP. doi: 10.1109/TIP.2024.3444315. Online ahead of print.

ABSTRACT

Deep learning-based video denoising methods have achieved great performance improvements in recent years. However, the expensive computational cost arising from sophisticated network design has severely limited their applications in real-world scenarios. To address this practical weakness, we propose a multiscale spatio-temporal memory network for fast video denoising, named MSTMN, aiming at striking an improved trade-off between cost and performance. To develop an efficient and effective algorithm for video denoising, we exploit a multiscale representation based on the Gaussian-Laplacian pyramid decomposition so that the reference frame can be restored in a coarse-to-fine manner. Guided by a model-based optimization approach, we design an effective variance estimation module, an alignment error estimation module and an adaptive fusion module for each scale of the pyramid representation. For the fusion module, we employ a reconstruction recurrence strategy to incorporate local temporal information. Moreover, we propose a memory enhancement module to exploit the global spatio-temporal information. Meanwhile, the similarity computation of the spatio-temporal memory network enables the proposed network to adaptively search the valuable information at the patch level, which avoids computationally expensive motion estimation and compensation operations. Experimental results on real-world raw video datasets have demonstrated that the proposed lightweight network outperforms current state-of-the-art fast video denoising algorithms such as FastDVDnet, EMVD, and ReMoNet with fewer computational costs.

PMID:39378250 | DOI:10.1109/TIP.2024.3444315

Categories: Literature Watch

The cardiologist in the age of artificial intelligence: what is left for us?

Tue, 2024-10-08 06:00

Cardiovasc Res. 2024 Oct 1:cvae171. doi: 10.1093/cvr/cvae171. Online ahead of print.

NO ABSTRACT

PMID:39378220 | DOI:10.1093/cvr/cvae171

Categories: Literature Watch

SPECT-MPI iterative denoising during the reconstruction process using a two-phase learned convolutional neural network

Tue, 2024-10-08 06:00

EJNMMI Phys. 2024 Oct 8;11(1):82. doi: 10.1186/s40658-024-00687-3.

ABSTRACT

PURPOSE: The problem of image denoising in single-photon emission computed tomography (SPECT) myocardial perfusion imaging (MPI) is a fundamental challenge. Although various image processing techniques have been presented, they may degrade the contrast of denoised images. The proposed idea in this study is to use a deep neural network as the denoising procedure during the iterative reconstruction process rather than the post-reconstruction phase. This method could decrease the background coefficient of variation (COV_bkg) of the final reconstructed image, which represents the amount of random noise, while improving the contrast-to-noise ratio (CNR).

METHODS: In this study, a generative adversarial network is used, where its generator is trained by a two-phase approach. In the first phase, the network is trained by a confined image region around the heart in transverse view. The second phase improves the network's generalization by tuning the network weights with the full image size as the input. The network was trained and tested by a dataset of 247 patients who underwent two immediate serially high- and low-noise SPECT-MPI.

RESULTS: Quantitative results show that compared to post-reconstruction low pass filtering and post-reconstruction deep denoising methods, our proposed method can decline the COV_bkg of the images by up to 10.28% and 12.52% and enhance the CNR by up to 54.54% and 45.82%, respectively.

CONCLUSION: The iterative deep denoising method outperforms 2D low-pass Gaussian filtering with an 8.4-mm FWHM and post-reconstruction deep denoising approaches.

PMID:39378001 | DOI:10.1186/s40658-024-00687-3

Categories: Literature Watch

Detectability of Hypoattenuating Liver Lesions with Deep Learning CT Reconstruction: A Phantom and Patient Study

Tue, 2024-10-08 06:00

Radiology. 2024 Oct;313(1):e232749. doi: 10.1148/radiol.232749.

ABSTRACT

Background CT deep learning image reconstruction (DLIR) improves image quality by reducing noise compared with adaptive statistical iterative reconstruction-V (ASIR-V). However, objective assessment of low-contrast lesion detectability is lacking. Purpose To investigate low-contrast detectability of hypoattenuating liver lesions on CT scans reconstructed with DLIR compared with CT scans reconstructed with ASIR-V in a patient and a phantom study. Materials and Methods This single-center retrospective study included patients undergoing portal venous phase abdominal CT between February and May 2021 and a low-contrast-resolution phantom scanned with the same protocol. Four reconstructions (ASIR-V at 40% strength [ASIR-V 40] and DLIR at three strengths) were generated. Five radiologists qualitatively assessed the images using the five-point Likert scale for image quality, lesion diagnostic confidence, conspicuity, and small lesion (≤1 cm) visibility. Up to two key lesions per patient, confirmed at histopathologic testing or at prior or follow-up imaging studies, were included. Lesion-to-background contrast-to-noise ratio was calculated. Interreader variability was analyzed. Intergroup qualitative and quantitative metrics were compared between DLIR and ASIR-V 40 using proportional odds logistic regression models. Results Eighty-six liver lesions (mean size, 15 mm ± 9.5 [SD]) in 50 patients (median age, 62 years [IQR, 57-73 years]; 27 [54%] female patients) were included. Differences were not detected for various qualitative low-contrast detectability metrics between ASIR-V 40 and DLIR (P > .05). Quantitatively, medium-strength DLIR and high-strength DLIR yielded higher lesion-to-background contrast-to-noise ratios than ASIR-V 40 (medium-strength DLIR vs ASIR-V 40: odds ratio [OR], 1.96 [95% CI: 1.65, 2.33]; high-strength DLIR vs ASIR-V 40: OR, 5.36 [95% CI: 3.68, 7.82]; P < .001). Low-contrast lesion attenuation was reduced by 2.8-3.6 HU with DLIR. Interreader agreement was moderate to very good for the qualitative metrics. Subgroup analysis based on lesion size of larger than 1 cm and 1 cm or smaller yielded similar results (P > .05). Qualitatively, phantom study results were similar to those in patients (P > .05). Conclusion The detectability of low-contrast liver lesions was similar on CT scans reconstructed with low-, medium-, and high-strength DLIR and ASIR-V 40 in both patient and phantom studies. Lesion-to-background contrast-to-noise ratios were higher for DLIR medium- and high-strength reconstructions compared with ASIR-V 40. © RSNA, 2024 Supplemental material is available for this article.

PMID:39377679 | DOI:10.1148/radiol.232749

Categories: Literature Watch

Boosting Deep Learning for Interpretable Brain MRI Lesion Detection through the Integration of Radiology Report Information

Tue, 2024-10-08 06:00

Radiol Artif Intell. 2024 Oct 9:e230520. doi: 10.1148/ryai.230520. Online ahead of print.

ABSTRACT

"Just Accepted" papers have undergone full peer review and have been accepted for publication in Radiology: Artificial Intelligence. This article will undergo copyediting, layout, and proof review before it is published in its final version. Please note that during production of the final copyedited article, errors may be discovered which could affect the content. Purpose To guide a deep learning (DL) model's attention toward brain lesion MRI characteristics by incorporating radiology report-derived textual features to achieve interpretable lesion detection. Materials and Methods In this retrospective study, 35282 brain MRI scans (January 2018-June 2023) and corresponding radiology reports from center 1 were used for training, validation, and internal testing. 2655 brain MRI scans (January 2022-December 2022) from centers 2-5 were reserved for external testing. Textual features were extracted from radiology reports to guide a DL model (ReportGuidedNet) focusing on lesion characteristics. Another DL model (PlainNet) without textual features was developed for comparative analysis. Both models diagnosed 15 conditions, including 14 diseases and normal brains. Performance of each model was assessed by calculating macro-and microaveraged area under the receiver operating characteristic curves (ma-AUC, mi-AUC). Attention maps, visualizing model attention, were assessed with a 5-point Likert scale. Results ReportGuidedNet outperformed PlainNet for all diagnoses on both internal (ma-AUC: 0.93 [95% CI: 0.91- 0.95] versus 0.85 [95% CI: 0.81-0.88]; mi-AUC: 0.93 [95% CI: 0.90-0.95] versus 0.89 [95% CI: 0.83-0.92]) and external (ma-AUC: 0.91 [95% CI: 0.88-0.93] versus 0.75 [95% CI: 0.72-0.79]; mi-AUC: 0.90 [95% CI: 0.87-0.92] versus 0.76 [95% CI: 0.72-0.80]) testing sets. The performance difference between internal and external testing sets was smaller for ReportGuidedNet than for PlainNet (Δma-AUC: 0.03 versus 0.10; Δmi-AUC: 0.02 versus 0.13). The Likert scale score of ReportGuidedNet was higher than that of PlainNet (mean ± SD: 2.50 ± 1.09 versus 1.32 ± 1.20; P < .001). Conclusion The integration of radiology report textual features improved the DL model's ability to detect brain lesions, enhancing interpretability and generalizability. Published under a CC BY 4.0 license.

PMID:39377669 | DOI:10.1148/ryai.230520

Categories: Literature Watch

Enhancing pap smear image classification: integrating transfer learning and attention mechanisms for improved detection of cervical abnormalities

Tue, 2024-10-08 06:00

Biomed Phys Eng Express. 2024 Sep 30;10(6). doi: 10.1088/2057-1976/ad7bc0.

ABSTRACT

Cervical cancer remains a major global health challenge, accounting for significant morbidity and mortality among women. Early detection through screening, such as Pap smear tests, is crucial for effective treatment and improved patient outcomes. However, traditional manual analysis of Pap smear images is labor-intensive, subject to human error, and requires extensive expertise. To address these challenges, automated approaches using deep learning techniques have been increasingly explored, offering the potential for enhanced diagnostic accuracy and efficiency. This research focuses on improving cervical cancer detection from Pap smear images using advanced deep-learning techniques. Specifically, we aim to enhance classification performance by leveraging Transfer Learning (TL) combined with an attention mechanism, supplemented by effective preprocessing techniques. Our preprocessing pipeline includes image normalization, resizing, and the application of Histogram of Oriented Gradients (HOG), all of which contribute to better feature extraction and improved model performance. The dataset used in this study is the Mendeley Liquid-Based Cytology (LBC) dataset, which provides a comprehensive collection of cervical cytology images annotated by expert cytopathologists. Initial experiments with the ResNet model on raw data yielded an accuracy of 63.95%. However, by applying our preprocessing techniques and integrating an attention mechanism, the accuracy of the ResNet model increased dramatically to 96.74%. Further, the Xception model, known for its superior feature extraction capabilities, achieved the best performance with an accuracy of 98.95%, along with high precision (0.97), recall (0.99), and F1-Score (0.98) on preprocessed data with an attention mechanism. These results underscore the effectiveness of combining preprocessing techniques, TL, and attention mechanisms to significantly enhance the performance of automated cervical cancer detection systems. Our findings demonstrate the potential of these advanced techniques to provide reliable, accurate, and efficient diagnostic tools, which could greatly benefit clinical practice and improve patient outcomes in cervical cancer screening.

PMID:39377445 | DOI:10.1088/2057-1976/ad7bc0

Categories: Literature Watch

Trap colour strongly affects the ability of deep learning models to recognize insect species in images of sticky traps

Tue, 2024-10-08 06:00

Pest Manag Sci. 2024 Oct 8. doi: 10.1002/ps.8464. Online ahead of print.

ABSTRACT

BACKGROUND: The use of computer vision and deep learning models to automatically classify insect species on sticky traps has proven to be a cost- and time-efficient approach to pest monitoring. As different species are attracted to different colours, the variety of sticky trap colours poses a challenge to the performance of the models. However, the effectiveness of deep learning in classifying pests on different coloured sticky traps has not yet been sufficiently explored. In this study, we aim to investigate the influence of sticky trap colour and imaging devices on the performance of deep learning models in classifying pests on sticky traps.

RESULTS: Our results show that using the MobileNetV2 architecture with transparent sticky traps as training data, the model predicted the pest species on transparent sticky traps with an accuracy of at least 0.95 and on other sticky trap colours with at least 0.85 of the F1 score. Using a generalised linear model (GLM) and a Boruta feature selection algorithm, we also showed that the colour and architecture of the sticky traps significantly influenced the performance of the model.

CONCLUSION: Our results support the development of an automatic classification of pests on a sticky trap, which should focus on colour and deep learning architecture to achieve good results. Future studies could aim to incorporate the trap system into pest monitoring, providing more accurate and cost-effective results in a pest management programme. © 2024 The Author(s). Pest Management Science published by John Wiley & Sons Ltd on behalf of Society of Chemical Industry.

PMID:39377441 | DOI:10.1002/ps.8464

Categories: Literature Watch

Correction to: Growing ecosystem of deep learning methods for modeling protein-protein interactions

Tue, 2024-10-08 06:00

Protein Eng Des Sel. 2024 Jan 29;37:gzae016. doi: 10.1093/protein/gzae016.

NO ABSTRACT

PMID:39377372 | DOI:10.1093/protein/gzae016

Categories: Literature Watch

Development and validation of a machine learning-based model for post-sepsis frailty

Tue, 2024-10-08 06:00

ERJ Open Res. 2024 Oct 7;10(5):00166-2024. doi: 10.1183/23120541.00166-2024. eCollection 2024 Sep.

ABSTRACT

BACKGROUND: The development of post-sepsis frailty is a common and significant problem, but it is a challenge to predict.

METHODS: Data for deep learning were extracted from a national multicentre prospective observational cohort of patients with sepsis in Korea between September 2019 and December 2021. The primary outcome was frailty at survival discharge, defined as a clinical frailty score on the Clinical Frailty Scale ≥5. We developed a deep learning model for predicting frailty after sepsis by 10 variables routinely collected at the recognition of sepsis. With cross-validation, we trained and tuned six machine learning models, including four conventional and two neural network models. Moreover, we computed the importance of each predictor variable in the model. We measured the performance of these models using a temporal validation data set.

RESULTS: A total of 8518 patients were included in the analysis; 5463 (64.1%) were frail, and 3055 (35.9%) were non-frail at discharge. The Extreme Gradient Boosting (XGB) achieved the highest area under the receiver operating characteristic curve (AUC) (0.8175) and accuracy (0.7414). To confirm the generalisation performance of artificial intelligence in predicting frailty at discharge, we conducted external validation with the COVID-19 data set. The XGB still showed a good performance with an AUC of 0.7668. The machine learning model could predict frailty despite the disparity in data distribution.

CONCLUSION: The machine learning-based model developed for predicting frailty after sepsis achieved high performance with limited baseline clinical parameters.

PMID:39377092 | PMC:PMC11456972 | DOI:10.1183/23120541.00166-2024

Categories: Literature Watch

Multi-modal remote perception learning for object sensory data

Tue, 2024-10-08 06:00

Front Neurorobot. 2024 Sep 19;18:1427786. doi: 10.3389/fnbot.2024.1427786. eCollection 2024.

ABSTRACT

INTRODUCTION: When it comes to interpreting visual input, intelligent systems make use of contextual scene learning, which significantly improves both resilience and context awareness. The management of enormous amounts of data is a driving force behind the growing interest in computational frameworks, particularly in the context of autonomous cars.

METHOD: The purpose of this study is to introduce a novel approach known as Deep Fused Networks (DFN), which improves contextual scene comprehension by merging multi-object detection and semantic analysis.

RESULTS: To enhance accuracy and comprehension in complex situations, DFN makes use of a combination of deep learning and fusion techniques. With a minimum gain of 6.4% in accuracy for the SUN-RGB-D dataset and 3.6% for the NYU-Dv2 dataset.

DISCUSSION: Findings demonstrate considerable enhancements in object detection and semantic analysis when compared to the methodologies that are currently being utilized.

PMID:39377028 | PMC:PMC11457376 | DOI:10.3389/fnbot.2024.1427786

Categories: Literature Watch

Integrating tabular data through image conversion for enhanced diagnosis: A novel intelligent decision support system for stratifying obstructive sleep apnoea patients using convolutional neural networks

Tue, 2024-10-08 06:00

Digit Health. 2024 Oct 3;10:20552076241272632. doi: 10.1177/20552076241272632. eCollection 2024 Jan-Dec.

ABSTRACT

OBJECTIVE: High-dimensional databases make it difficult to apply traditional learning algorithms to biomedical applications. Recent developments in computer technology have introduced deep learning (DL) as a potential solution to these difficulties. This study presents a novel intelligent decision support system based on a novel interpretation of data formalisation from tabular data in DL techniques. Once defined, it is used to diagnose the severity of obstructive sleep apnoea, distinguishing between moderate to severe and mild/no cases.

METHODS: The study uses a complete database extract from electronic health records of 2472 patients, including anthropometric data, habits, medications, comorbidities, and patient-reported symptoms. The novelty of this methodology lies in the initial processing of the patients' data, which is formalised into images. These images are then used as input to train a convolutional neural network (CNN), which acts as the inference engine of the system.

RESULTS: The initial tests of the system were performed on a set of 247 samples from the Pulmonary Department of the Álvaro Cunqueiro Hospital in Vigo (Galicia, Spain), with an AUC value of ≈ 0.8.

CONCLUSIONS: This study demonstrates the benefits of an intelligent decision support system based on a novel data formalisation approach that allows the use of advanced DL techniques starting from tabular data. In this way, the ability of CNNs to recognise complex patterns using visual elements such as gradients and contrasts can be exploited. This approach effectively addresses the challenges of analysing large amounts of tabular data and reduces common problems such as bias and variance, resulting in improved diagnostic accuracy.

PMID:39376943 | PMC:PMC11457234 | DOI:10.1177/20552076241272632

Categories: Literature Watch

Analyzing Racial Differences in Imaging Joint Replacement Registries Using Generative Artificial Intelligence: Advancing Orthopaedic Data Equity

Tue, 2024-10-08 06:00

Arthroplast Today. 2024 Sep 23;29:101503. doi: 10.1016/j.artd.2024.101503. eCollection 2024 Oct.

ABSTRACT

BACKGROUND: Discrepancies in medical data sets can perpetuate bias, especially when training deep learning models, potentially leading to biased outcomes in clinical applications. Understanding these biases is crucial for the development of equitable healthcare technologies. This study employs generative deep learning technology to explore and understand radiographic differences based on race among patients undergoing total hip arthroplasty.

METHODS: Utilizing a large institutional registry, we retrospectively analyzed pelvic radiographs from total hip arthroplasty patients, characterized by demographics and image features. Denoising diffusion probabilistic models generated radiographs conditioned on demographic and imaging characteristics. Fréchet Inception Distance assessed the generated image quality, showing the diversity and realism of the generated images. Sixty transition videos were generated that showed transforming White pelvises to their closest African American counterparts and vice versa while controlling for patients' sex, age, and body mass index. Two expert surgeons and 2 radiologists carefully studied these videos to understand the systematic differences that are present in the 2 races' radiographs.

RESULTS: Our data set included 480,407 pelvic radiographs, with a predominance of White patients over African Americans. The generative denoising diffusion probabilistic model created high-quality images and reached an Fréchet Inception Distance of 6.8. Experts identified 6 characteristics differentiating races, including interacetabular distance, osteoarthritis degree, obturator foramina shape, femoral neck-shaft angle, pelvic ring shape, and femoral cortical thickness.

CONCLUSIONS: This study demonstrates the potential of generative models for understanding disparities in medical imaging data sets. By visualizing race-based differences, this method aids in identifying bias in downstream tasks, fostering the development of fairer healthcare practices.

PMID:39376670 | PMC:PMC11456877 | DOI:10.1016/j.artd.2024.101503

Categories: Literature Watch

Pages