Deep learning

Uncertainty quantification for probabilistic machine learning in earth observation using conformal prediction

Sat, 2024-07-13 06:00

Sci Rep. 2024 Jul 13;14(1):16166. doi: 10.1038/s41598-024-65954-w.

ABSTRACT

Machine learning is increasingly applied to Earth Observation (EO) data to obtain datasets that contribute towards international accords. However, these datasets contain inherent uncertainty that needs to be quantified reliably to avoid negative consequences. In response to the increased need to report uncertainty, we bring attention to the promise of conformal prediction within the domain of EO. Unlike previous uncertainty quantification methods, conformal prediction offers statistically valid prediction regions while concurrently supporting any machine learning model and data distribution. To support the need for conformal prediction, we reviewed EO datasets and found that only 22.5% of the datasets incorporated a degree of uncertainty information, with unreliable methods prevalent. Current open implementations require moving large amounts of EO data to the algorithms. We introduced Google Earth Engine native modules that bring conformal prediction to the data and compute, facilitating the integration of uncertainty quantification into existing traditional and deep learning modelling workflows. To demonstrate the versatility and scalability of these tools we apply them to valued EO applications spanning local to global extents, regression, and classification tasks. Subsequently, we discuss the opportunities arising from the use of conformal prediction in EO. We anticipate that accessible and easy-to-use tools, such as those provided here, will drive wider adoption of rigorous uncertainty quantification in EO, thereby enhancing the reliability of downstream uses such as operational monitoring and decision-making.

PMID:39003341 | DOI:10.1038/s41598-024-65954-w

Categories: Literature Watch

Oral mucosal lesions triage via YOLOv7 models

Sat, 2024-07-13 06:00

J Formos Med Assoc. 2024 Jul 12:S0929-6646(24)00313-9. doi: 10.1016/j.jfma.2024.07.010. Online ahead of print.

ABSTRACT

BACKGROUND/PURPOSE: The global incidence of lip and oral cavity cancer continues to rise, necessitating improved early detection methods. This study leverages the capabilities of computer vision and deep learning to enhance the early detection and classification of oral mucosal lesions.

METHODS: A dataset initially consisting of 6903 white-light macroscopic images collected from 2006 to 2013 was expanded to over 50,000 images to train the YOLOv7 deep learning model. Lesions were categorized into three referral grades: benign (green), potentially malignant (yellow), and malignant (red), facilitating efficient triage.

RESULTS: The YOLOv7 models, particularly the YOLOv7-E6, demonstrated high precision and recall across all lesion categories. The YOLOv7-D6 model excelled at identifying malignant lesions with notable precision, recall, and F1 scores. Enhancements, including the integration of coordinate attention in the YOLOv7-D6-CA model, significantly improved the accuracy of lesion classification.

CONCLUSION: The study underscores the robust comparison of various YOLOv7 model configurations in the classification to triage oral lesions. The overall results highlight the potential of deep learning models to contribute to the early detection of oral cancers, offering valuable tools for both clinical settings and remote screening applications.

PMID:39003230 | DOI:10.1016/j.jfma.2024.07.010

Categories: Literature Watch

Predictive value of MRI-based deep learning model for lymphovascular invasion status in node-negative invasive breast cancer

Sat, 2024-07-13 06:00

Sci Rep. 2024 Jul 13;14(1):16204. doi: 10.1038/s41598-024-67217-0.

ABSTRACT

To retrospectively assess the effectiveness of deep learning (DL) model, based on breast magnetic resonance imaging (MRI), in predicting preoperative lymphovascular invasion (LVI) status in patients diagnosed with invasive breast cancer who have negative axillary lymph nodes (LNs). Data was gathered from 280 patients, including 148 with LVI-positive and 141 with LVI-negative lesions. These patients had undergone preoperative breast MRI and were histopathologically confirmed to have invasive breast cancer without axillary LN metastasis. The cohort was randomly split into training and validation groups in a 7:3 ratio. Radiomics features for each lesion were extracted from the first post-contrast dynamic contrast-enhanced (DCE)-MRI. The Least Absolute Shrinkage and Selection Operator (LASSO) regression method and logistic regression analyses were employed to identify significant radiomic features and clinicoradiological variables. These models were established using four machine learning (ML) algorithms and one DL algorithm. The predictive performance of the models (radiomics, clinicoradiological, and combination) was assessed through discrimination and compared using the DeLong test. Four clinicoradiological parameters and 10 radiomic features were selected by LASSO for model development. The Multilayer Perceptron (MLP) model, constructed using both radiomic and clinicoradiological features, demonstrated excellent performance in predicting LVI, achieving a high area under the curve (AUC) of 0.835 for validation. The DL model (MLP-radiomic) achieved the highest accuracy (AUC = 0.896), followed by DL model (MLP-combination) with an AUC of 0.835. Both DL models were significantly superior to the ML model (RF-clinical) with an AUC of 0.720. The DL model (MLP), which integrates radiomic features with clinicoradiological information, effectively aids in the preoperative determination of LVI status in patients with invasive breast cancer and negative axillary LNs. This is beneficial for making informed clinical decisions.

PMID:39003325 | DOI:10.1038/s41598-024-67217-0

Categories: Literature Watch

Automatic 3D reconstruction of vertebrae from orthogonal bi-planar radiographs

Sat, 2024-07-13 06:00

Sci Rep. 2024 Jul 13;14(1):16165. doi: 10.1038/s41598-024-65795-7.

ABSTRACT

When conducting spine-related diagnosis and surgery, the three-dimensional (3D) upright posture of the spine under natural weight bearing is of significant clinical value for physicians to analyze the force on the spine. However, existing medical imaging technologies cannot meet current requirements of medical service. On the one hand, the mainstream 3D volumetric imaging modalities (e.g. CT and MRI) require patients to lie down during the imaging process. On the other hand, the imaging modalities conducted in an upright posture (e.g. radiograph) can only realize 2D projections, which lose the valid information of spinal anatomy and curvature. Developments of deep learning-based 3D reconstruction methods bring potential to overcome the limitations of the existing medical imaging technologies. To deal with the limitations of current medical imaging technologies as is described above, in this paper, we propose a novel deep learning framework, ReVerteR, which can realize automatic 3D Reconstruction of Vertebrae from orthogonal bi-planar Radiographs. With the utilization of self-attention mechanism and specially designed loss function combining Dice, Hausdorff, Focal, and MSE, ReVerteR can alleviate the sample-imbalance problem during the reconstruction process and realize the fusion of the centroid annotation and the focused vertebra. Furthermore, aiming at automatic and customized 3D spinal reconstruction in real-world scenarios, we extend ReVerteR to a clinical deployment-oriented framework, and develop an interactive interface with all functions in the framework integrated so as to enhance human-computer interaction during clinical decision-making. Extensive experiments and visualization conducted on our constructed datasets based on two benchmark datasets of spinal CT, VerSe 2019 and VerSe 2020, demonstrate the effectiveness of our proposed ReVerteR. In this paper, we propose an automatic 3D reconstruction method of vertebrae based on orthogonal bi-planar radiographs. With the 3D upright posture of the spine under natural weight bearing effectively constructed, our proposed method is expected to better support doctors make clinical decision during spine-related diagnosis and surgery.

PMID:39003269 | DOI:10.1038/s41598-024-65795-7

Categories: Literature Watch

Preoperative Contrast-Enhanced CT-Based Deep Learning Radiomics Model for Distinguishing Retroperitoneal Lipomas and Well-Differentiated Liposarcomas

Sat, 2024-07-13 06:00

Acad Radiol. 2024 Jul 12:S1076-6332(24)00422-7. doi: 10.1016/j.acra.2024.06.035. Online ahead of print.

ABSTRACT

RATIONALE AND OBJECTIVES: To assess the efficacy of a preoperative contrast-enhanced CT (CECT)-based deep learning radiomics nomogram (DLRN) for predicting murine double minute 2 (MDM2) gene amplification as a means of distinguishing between retroperitoneal well-differentiated liposarcomas (WDLPS) and lipomas.

METHODS: This retrospective multi-center study included 167 patients (training/external test cohort, 104/63) with MDM2-positive WDLPS or MDM2-negative lipomas. Clinical data and CECT features were independently measured and analyzed by two radiologists. A clinico-radiological model, radiomics signature (RS), deep learning and radiomics signature (DLRS), and a DLRN incorporating radiomics and deep learning features were developed to differentiate between WDLPS and lipoma. The model utility was evaluated based on the area under the receiver operating characteristic curve (AUC), accuracy, calibration curve, and decision curve analysis (DCA).

RESULTS: The DLRN showed good performance for distinguishing retroperitoneal lipomas and WDLPS in the training (AUC, 0.981; accuracy, 0.933) and external validation group (AUC, 0.861; accuracy, 0.810). The DeLong test revealed the DLRN was noticeably better than clinico-radiological and RS models (training: 0.981 vs. 0.890 vs. 0.751; validation: 0.861 vs. 0.724 vs. 0.700; both P < 0.05); however, no discernible difference in performance was seen between the DLRN and DLRS (training: 0.981 vs. 0.969; validation: 0.861 vs. 0.837; both P > 0.05). The calibration curve analysis and DCA demonstrated that the nomogram exhibited good calibration and offered substantial clinical advantages.

CONCLUSION: The DLRN exhibited strong predictive capability in predicting WDLPS and retroperitoneal lipomas preoperatively, making it a promising imaging biomarker that can facilitate personalized management and precision medicine.

PMID:39003228 | DOI:10.1016/j.acra.2024.06.035

Categories: Literature Watch

Deep survival analysis for interpretable Time-Varying prediction of preeclampsia risk

Sat, 2024-07-13 06:00

J Biomed Inform. 2024 Jul 11:104688. doi: 10.1016/j.jbi.2024.104688. Online ahead of print.

ABSTRACT

OBJECTIVE: Survival analysis is widely utilized in healthcare to predict the timing of disease onset. Traditional methods of survival analysis are usually based on Cox Proportional Hazards model and assume proportional risk for all subjects. However, this assumption is rarely true for most diseases, as the underlying factors have complex, non-linear, and time-varying relationships. This concern is especially relevant for pregnancy, where the risk for pregnancy-related complications, such as preeclampsia, varies across gestation. Recently, deep learning survival models have shown promise in addressing the limitations of classical models, as the novel models allow for non-proportional risk handling, capturing nonlinear relationships, and navigating complex temporal dynamics.

METHODS: We present a methodology to model the temporal risk of preeclampsia during pregnancy and investigate the associated clinical risk factors. We utilized a retrospective dataset including 66,425 pregnant individuals who delivered in two tertiary care centers from 2015 to 2023. We modeled the preeclampsia risk by modifying DeepHit, a deep survival model, which leverages neural network architecture to capture time-varying relationships between covariates in pregnancy. We applied time series k-means clustering to DeepHit's normalized output and investigated interpretability using Shapley values.

RESULTS: We demonstrate that DeepHit can effectively handle high-dimensional data and evolving risk hazards over time with performance similar to the Cox Proportional Hazards model, achieving an area under the curve (AUC) of 0.78 for both models. The deep survival model outperformed traditional methodology by identifying time-varied risk trajectories for preeclampsia, providing insights for early and individualized intervention. K-means clustering resulted in patients delineating into low-risk, early-onset, and late-onset preeclampsia groups-notably, each of those has distinct risk factors.

CONCLUSION: This work demonstrates a novel application of deep survival analysis in time-varying prediction of preeclampsia risk. Our results highlight the advantage of deep survival models compared to Cox Proportional Hazards models in providing personalized risk trajectory and demonstrating the potential of deep survival models to generate interpretable and meaningful clinical applications in medicine.

PMID:39002866 | DOI:10.1016/j.jbi.2024.104688

Categories: Literature Watch

Identification and diagnosis of schizophrenia based on multichannel EEG and CNN deep learning model

Sat, 2024-07-13 06:00

Schizophr Res. 2024 Jul 12;271:28-35. doi: 10.1016/j.schres.2024.07.015. Online ahead of print.

ABSTRACT

This paper proposes a high-accuracy EEG-based schizophrenia (SZ) detection approach. Unlike comparable literature studies employing conventional machine learning algorithms, our method autonomously extracts the necessary features for network training from EEG recordings. The proposed model is a ten-layered CNN that contains a max pooling layer, a Global Average Pooling layer, four convolution layers, two dropout layers for overfitting prevention, and two fully connected layers. The efficiency of the suggested method was assessed using the ten-fold-cross validation technique and the EEG records of 14 healthy subjects and 14 SZ patients. The obtained mean accuracy score was 99.18 %. To confirm the high mean accuracy attained, we tested the model on unseen data with a near-perfect accuracy score (almost 100 %). In addition, the results we obtained outperform numerous other comparable works.

PMID:39002527 | DOI:10.1016/j.schres.2024.07.015

Categories: Literature Watch

Driving risk identification of urban arterial and collector roads based on multi-scale data

Sat, 2024-07-13 06:00

Accid Anal Prev. 2024 Jul 12;206:107712. doi: 10.1016/j.aap.2024.107712. Online ahead of print.

ABSTRACT

Urban arterial and collector roads, while interconnected within the urban transportation network, serve distinct purposes, leading to different driving risk profiles. Investigating these differences using advanced methods is of paramount significance. This study aims to achieve this by primarily collecting and processing relevant vehicle trajectory data alongside driver-vehicle-road-environment data. A comprehensive risk assessment matrix is constructed to assess driving risks, incorporating multiple conflict and traffic flow indicators with statistically temporal stability. The Entropy weight-TOPSIS method and the K-means algorithm are employed to determine the risk scores and levels of the target arterial and collector roads. Using risk levels as the outcome variables and multi-scale features as the explanatory variables, random parameters models with heterogeneity in means and variances are developed to identify the determinants of driving risks at different levels. Likelihood ratio tests and comparisons of out-of-sample and within-sample prediction are conducted. Results reveal significant statistical differences in the risk profiles between arterial and collector roads. The marginal effects of significant parameters are then calculated separately for arterial and collector roads, indicating that several factors have different impacts on the probability of risk levels for arterial and collector roads, such as the number of movable elements in road landscape pictures, the standard deviation of the vehicle's lateral acceleration, the average standard deviation of speed for all vehicles on the road segment, and the number of one-way lanes on the road segment. Some practical implications are provided based on the findings. Future research can be implemented by expanding the collected data to different regions and cities over longer periods.

PMID:39002352 | DOI:10.1016/j.aap.2024.107712

Categories: Literature Watch

deepbet: Fast brain extraction of T1-weighted MRI using Convolutional Neural Networks

Sat, 2024-07-13 06:00

Comput Biol Med. 2024 Jul 12;179:108845. doi: 10.1016/j.compbiomed.2024.108845. Online ahead of print.

ABSTRACT

BACKGROUND: Brain extraction in magnetic resonance imaging (MRI) data is an important segmentation step in many neuroimaging preprocessing pipelines. Image segmentation is one of the research fields in which deep learning had the biggest impact in recent years. Consequently, traditional brain extraction methods are now being replaced by deep learning-based methods.

METHOD: Here, we used a unique dataset compilation comprising 7837 T1-weighted (T1w) MR images from 191 different OpenNeuro datasets in combination with advanced deep learning methods to build a fast, high-precision brain extraction tool called deepbet.

RESULTS: deepbet sets a novel state-of-the-art performance during cross-dataset validation with a median Dice score (DSC) of 99.0 on unseen datasets, outperforming the current best performing deep learning (DSC=97.9) and classic (DSC=96.5) methods. While current methods are more sensitive to outliers, deepbet achieves a Dice score of >97.4 across all 7837 images from 191 different datasets. This robustness was additionally tested in 5 external datasets, which included challenging clinical MR images. During visual exploration of each method's output which resulted in the lowest Dice score, major errors could be found for all of the tested tools except deepbet. Finally, deepbet uses a compute efficient variant of the UNet architecture, which accelerates brain extraction by a factor of ≈10 compared to current methods, enabling the processing of one image in ≈2 s on low level hardware.

CONCLUSIONS: In conclusion, deepbet demonstrates superior performance and reliability in brain extraction across a wide range of T1w MR images of adults, outperforming existing top tools. Its high minimal Dice score and minimal objective errors, even in challenging conditions, validate deepbet as a highly dependable tool for accurate brain extraction. deepbet can be conveniently installed via "pip install deepbet" and is publicly accessible at https://github.com/wwu-mmll/deepbet.

PMID:39002314 | DOI:10.1016/j.compbiomed.2024.108845

Categories: Literature Watch

Predicting recovery following stroke: Deep learning, multimodal data and feature selection using explainable AI

Sat, 2024-07-13 06:00

Neuroimage Clin. 2024 Jul 2;43:103638. doi: 10.1016/j.nicl.2024.103638. Online ahead of print.

ABSTRACT

Machine learning offers great potential for automated prediction of post-stroke symptoms and their response to rehabilitation. Major challenges for this endeavour include the very high dimensionality of neuroimaging data, the relatively small size of the datasets available for learning and interpreting the predictive features, as well as, how to effectively combine neuroimaging and tabular data (e.g. demographic information and clinical characteristics). This paper evaluates several solutions based on two strategies. The first is to use 2D images that summarise MRI scans. The second is to select key features that improve classification accuracy. Additionally, we introduce the novel approach of training a convolutional neural network (CNN) on images that combine regions-of-interests (ROIs) extracted from MRIs, with symbolic representations of tabular data. We evaluate a series of CNN architectures (both 2D and a 3D) that are trained on different representations of MRI and tabular data, to predict whether a composite measure of post-stroke spoken picture description ability is in the aphasic or non-aphasic range. MRI and tabular data were acquired from 758 English speaking stroke survivors who participated in the PLORAS study. Each participant was assigned to one of five different groups that were matched for initial severity of symptoms, recovery time, left lesion size and the months or years post-stroke that spoken description scores were collected. Training and validation were carried out on the first four groups. The fifth (lock-box/test set) group was used to test how well model accuracy generalises to new (unseen) data. The classification accuracy for a baseline logistic regression was 0.678 based on lesion size alone, rising to 0.757 and 0.813 when initial symptom severity and recovery time were successively added. The highest classification accuracy (0.854), area under the curve (0.899) and F1 score (0.901) were observed when 8 regions of interest were extracted from each MRI scan and combined with lesion size, initial severity and recovery time in a 2D Residual Neural Network (ResNet). This was also the best model when data were limited to the 286 participants with moderate or severe initial aphasia (with area under curve = 0.865), a group that would be considered more difficult to classify. Our findings demonstrate how imaging and tabular data can be combined to achieve high post-stroke classification accuracy, even when the dataset is small in machine learning terms. We conclude by proposing how the current models could be improved to achieve even higher levels of accuracy using images from hospital scanners.

PMID:39002223 | DOI:10.1016/j.nicl.2024.103638

Categories: Literature Watch

Needle tracking in low-resolution ultrasound volumes using deep learning

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03234-8. Online ahead of print.

ABSTRACT

PURPOSE: Clinical needle insertion into tissue, commonly assisted by 2D ultrasound imaging for real-time navigation, faces the challenge of precise needle and probe alignment to reduce out-of-plane movement. Recent studies investigate 3D ultrasound imaging together with deep learning to overcome this problem, focusing on acquiring high-resolution images to create optimal conditions for needle tip detection. However, high-resolution also requires a lot of time for image acquisition and processing, which limits the real-time capability. Therefore, we aim to maximize the US volume rate with the trade-off of low image resolution. We propose a deep learning approach to directly extract the 3D needle tip position from sparsely sampled US volumes.

METHODS: We design an experimental setup with a robot inserting a needle into water and chicken liver tissue. In contrast to manual annotation, we assess the needle tip position from the known robot pose. During insertion, we acquire a large data set of low-resolution volumes using a 16 × 16 element matrix transducer with a volume rate of 4 Hz. We compare the performance of our deep learning approach with conventional needle segmentation.

RESULTS: Our experiments in water and liver show that deep learning outperforms the conventional approach while achieving sub-millimeter accuracy. We achieve mean position errors of 0.54 mm in water and 1.54 mm in liver for deep learning.

CONCLUSION: Our study underlines the strengths of deep learning to predict the 3D needle positions from low-resolution ultrasound volumes. This is an important milestone for real-time needle navigation, simplifying the alignment of needle and ultrasound probe and enabling a 3D motion analysis.

PMID:39002100 | DOI:10.1007/s11548-024-03234-8

Categories: Literature Watch

3D mobile regression vision transformer for collateral imaging in acute ischemic stroke

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03229-5. Online ahead of print.

ABSTRACT

PURPOSE: The accurate and timely assessment of the collateral perfusion status is crucial in the diagnosis and treatment of patients with acute ischemic stroke. Previous works have shown that collateral imaging, derived from CT angiography, MR perfusion, and MR angiography, aids in evaluating the collateral status. However, such methods are time-consuming and/or sub-optimal due to the nature of manual processing and heuristics. Recently, deep learning approaches have shown to be promising for generating collateral imaging. These, however, suffer from the computational complexity and cost.

METHODS: In this study, we propose a mobile, lightweight deep regression neural network for collateral imaging in acute ischemic stroke, leveraging dynamic susceptibility contrast MR perfusion (DSC-MRP). Built based upon lightweight convolution and Transformer architectures, the proposed model manages the balance between the model complexity and performance.

RESULTS: We evaluated the performance of the proposed model in generating the five-phase collateral maps, including arterial, capillary, early venous, late venous, and delayed phases, using DSC-MRP from 952 patients. In comparison with various deep learning models, the proposed method was superior to the competitors with similar complexity and was comparable to the competitors of high complexity.

CONCLUSION: The results suggest that the proposed model is able to facilitate rapid and precise assessment of the collateral status of patients with acute ischemic stroke, leading to improved patient care and outcome.

PMID:39002099 | DOI:10.1007/s11548-024-03229-5

Categories: Literature Watch

Domain adaptation using AdaBN and AdaIN for high-resolution IVD mesh reconstruction from clinical MRI

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03233-9. Online ahead of print.

ABSTRACT

PURPOSE: Deep learning has firmly established its dominance in medical imaging applications. However, careful consideration must be exercised when transitioning a trained source model to adapt to an entirely distinct environment that deviates significantly from the training set. The majority of the efforts to mitigate this issue have predominantly focused on classification and segmentation tasks. In this work, we perform a domain adaptation of a trained source model to reconstruct high-resolution intervertebral disc meshes from low-resolution MRI.

METHODS: To address the outlined challenges, we use MRI2Mesh as the shape reconstruction network. It incorporates three major modules: image encoder, mesh deformation, and cross-level feature fusion. This feature fusion module is used to encapsulate local and global disc features. We evaluate two major domain adaptation techniques: adaptive batch normalization (AdaBN) and adaptive instance normalization (AdaIN) for the task of shape reconstruction.

RESULTS: Experiments conducted on distinct datasets, including data from different populations, machines, and test sites demonstrate the effectiveness of MRI2Mesh for domain adaptation. MRI2Mesh achieved up to a 14% decrease in Hausdorff distance (HD) and a 19% decrease in the point-to-surface (P2S) metric for both AdaBN and AdaIN experiments, indicating improved performance.

CONCLUSION: MRI2Mesh has demonstrated consistent superiority to the state-of-the-art Voxel2Mesh network across a diverse range of datasets, populations, and scanning protocols, highlighting its versatility. Additionally, AdaBN has emerged as a robust method compared to the AdaIN technique. Further experiments show that MRI2Mesh, when combined with AdaBN, holds immense promise for enhancing the precision of anatomical shape reconstruction in domain adaptation.

PMID:39002098 | DOI:10.1007/s11548-024-03233-9

Categories: Literature Watch

The use of artificial intelligence in musculoskeletal ultrasound: a systematic review of the literature

Sat, 2024-07-13 06:00

Radiol Med. 2024 Jul 13. doi: 10.1007/s11547-024-01856-1. Online ahead of print.

ABSTRACT

PURPOSE: To systematically review the use of artificial intelligence (AI) in musculoskeletal (MSK) ultrasound (US) with an emphasis on AI algorithm categories and validation strategies.

MATERIAL AND METHODS: An electronic literature search was conducted for articles published up to January 2024. Inclusion criteria were the use of AI in MSK US, involvement of humans, English language, and ethics committee approval.

RESULTS: Out of 269 identified papers, 16 studies published between 2020 and 2023 were included. The research was aimed at predicting diagnosis and/or segmentation in a total of 11 (69%) out of 16 studies. A total of 11 (69%) studies used deep learning (DL)-based algorithms, three (19%) studies employed conventional machine learning (ML)-based algorithms, and two (12%) studies employed both conventional ML- and DL-based algorithms. Six (38%) studies used cross-validation techniques with K-fold cross-validation being the most frequently employed (n = 4, 25%). Clinical validation with separate internal test datasets was reported in nine (56%) papers. No external clinical validation was reported.

CONCLUSION: AI is a topic of increasing interest in MSK US research. In future studies, attention should be paid to the use of validation strategies, particularly regarding independent clinical validation performed on external datasets.

PMID:39001961 | DOI:10.1007/s11547-024-01856-1

Categories: Literature Watch

Deep learning model based on endoscopic images predicting treatment response in locally advanced rectal cancer undergo neoadjuvant chemoradiotherapy: a multicenter study

Sat, 2024-07-13 06:00

J Cancer Res Clin Oncol. 2024 Jul 13;150(7):350. doi: 10.1007/s00432-024-05876-2.

ABSTRACT

PURPOSE: Neoadjuvant chemoradiotherapy has been the standard practice for patients with locally advanced rectal cancer. However, the treatment response varies greatly among individuals, how to select the optimal candidates for neoadjuvant chemoradiotherapy is crucial. This study aimed to develop an endoscopic image-based deep learning model for predicting the response to neoadjuvant chemoradiotherapy in locally advanced rectal cancer.

METHODS: In this multicenter observational study, pre-treatment endoscopic images of patients from two Chinese medical centers were retrospectively obtained and a deep learning-based tumor regression model was constructed. Treatment response was evaluated based on the tumor regression grade and was defined as good response and non-good response. The prediction performance of the deep learning model was evaluated in the internal and external test sets. The main outcome was the accuracy of the treatment prediction model, measured by the AUC and accuracy.

RESULTS: This deep learning model achieved favorable prediction performance. In the internal test set, the AUC and accuracy were 0.867 (95% CI: 0.847-0.941) and 0.836 (95% CI: 0.818-0.896), respectively. The prediction performance was fully validated in the external test set, and the model had an AUC of 0.758 (95% CI: 0.724-0.834) and an accuracy of 0.807 (95% CI: 0.774-0.843).

CONCLUSION: The deep learning model based on endoscopic images demonstrated exceptional predictive power for neoadjuvant treatment response, highlighting its potential for guiding personalized therapy.

PMID:39001926 | DOI:10.1007/s00432-024-05876-2

Categories: Literature Watch

Artificial intelligence based diagnosis of sulcus: assesment of videostroboscopy via deep learning

Sat, 2024-07-13 06:00

Eur Arch Otorhinolaryngol. 2024 Jul 13. doi: 10.1007/s00405-024-08801-y. Online ahead of print.

ABSTRACT

PURPOSE: To develop a convolutional neural network (CNN)-based model for classifying videostroboscopic images of patients with sulcus, benign vocal fold (VF) lesions, and healthy VFs to improve clinicians' accuracy in diagnosis during videostroboscopies when evaluating sulcus.

MATERIALS AND METHODS: Videostroboscopies of 433 individuals who were diagnosed with sulcus (91), who were diagnosed with benign VF diseases (i.e., polyp, nodule, papilloma, cyst, or pseudocyst [311]), or who were healthy (33) were analyzed. After extracting 91,159 frames from videostroboscopies, a CNN-based model was created and tested. The healthy and sulcus groups underwent binary classification. In the second phase of the study, benign VF lesions were added to the training set, and multiclassification was executed across all groups. The proposed CNN-based model results were compared with five laryngology experts' assessments.

RESULTS: In the binary classification phase, the CNN-based model achieved 98% accuracy, 98% recall, 97% precision, and a 97% F1 score for classifying sulcus and healthy VFs. During the multiclassification phase, when evaluated on a subset of frames encompassing all included groups, the CNN-based model demonstrated greater accuracy when compared with that of the five laryngologists (%76 versus 72%, 68%, 72%, 63%, and 72%).

CONCLUSION: The utilization of a CNN-based model serves as a significant aid in the diagnosis of sulcus, a VF disease that presents notable challenges in the diagnostic process. Further research could be undertaken to assess the practicality of implementing this approach in real-time application in clinical practice.

PMID:39001913 | DOI:10.1007/s00405-024-08801-y

Categories: Literature Watch

Three-Dimensional Label-Free Observing of the Self-Assembled Nanoparticles inside a Single Cell at Nanoscale Resolution

Sat, 2024-07-13 06:00

ACS Nano. 2024 Jul 13. doi: 10.1021/acsnano.4c06095. Online ahead of print.

ABSTRACT

Understanding the intracellular behavior of nanoparticles (NPs) plays a key role in optimizing the self-assembly performance of nanomedicine. However, conducting the 3D, label-free, quantitative observation of self-assembled NPs within intact single cells remains a substantial challenge in complicated intracellular environments. Here, we propose a deep learning combined synchrotron radiation hard X-ray nanotomography approach to visualize the self-assembled ultrasmall iron oxide (USIO) NPs in a single cell. The method allows us to explore comprehensive information on NPs, such as their distribution, morphology, location, and interaction with cell organelles, and provides quantitative analysis of the heterogeneous size and morphologies of USIO NPs under diverse conditions. This label-free, in situ method provides a tool for precise characterization of intracellular self-assembled NPs to improve the evaluation and design of a bioresponsive nanomedicine.

PMID:39001860 | DOI:10.1021/acsnano.4c06095

Categories: Literature Watch

Current status and future directions in artificial intelligence for nuclear cardiology

Sat, 2024-07-13 06:00

Expert Rev Cardiovasc Ther. 2024 Jul 13. doi: 10.1080/14779072.2024.2380764. Online ahead of print.

ABSTRACT

INTRODUCTION: Myocardial perfusion imaging (MPI) is one of the most commonly ordered cardiac imaging tests. Accurate motion correction, image registration, and reconstruction is critical for high-quality imaging, but this can be technically challenging and traditionally has relied on expert manual processing. With accurate processing, there is a rich variety of clinical, stress, functional, and anatomic data that can be integrated to guide patient management.

AREAS COVERED: Pubmed and Google Scholar were reviewed for articles related to artificial intelligence in nuclear cardiology published between 2020 and 2024. We will outline the prominent roles for artificial intelligence (AI) solutions to provide motion correction, image registration, and reconstruction. We will review the role for AI in extracting anatomic data for hybrid MPI which is otherwise neglected. Lastly, we will discuss AI methods to integrate the wealth of data to improve disease diagnosis or risk stratification.

EXPERT OPINION: There is growing evidence that AI will transform the performance of MPI by automating and improving on aspects of image acquisition and reconstruction. Physicians and researchers will need to understand the potential strengths of AI in order to benefit from the full clinical utility of MPI.

PMID:39001698 | DOI:10.1080/14779072.2024.2380764

Categories: Literature Watch

Enhancing Immunotherapy Response Prediction in Metastatic Lung Adenocarcinoma: Leveraging Shallow and Deep Learning with CT-Based Radiomics across Single and Multiple Tumor Sites

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jul 8;16(13):2491. doi: 10.3390/cancers16132491.

ABSTRACT

This study aimed to evaluate the potential of pre-treatment CT-based radiomics features (RFs) derived from single and multiple tumor sites, and state-of-the-art machine-learning survival algorithms, in predicting progression-free survival (PFS) for patients with metastatic lung adenocarcinoma (MLUAD) receiving first-line treatment including immune checkpoint inhibitors (CPIs). To do so, all adults with newly diagnosed MLUAD, pre-treatment contrast-enhanced CT scan, and performance status ≤ 2 who were treated at our cancer center with first-line CPI between November 2016 and November 2022 were included. RFs were extracted from all measurable lesions with a volume ≥ 1 cm3 on the CT scan. To capture intra- and inter-tumor heterogeneity, RFs from the largest tumor of each patient, as well as lowest, highest, and average RF values over all lesions per patient were collected. Intra-patient inter-tumor heterogeneity metrics were calculated to measure the similarity between each patient lesions. After filtering predictors with univariable Cox p < 0.100 and analyzing their correlations, five survival machine-learning algorithms (stepwise Cox regression [SCR], LASSO Cox regression, random survival forests, gradient boosted machine [GBM], and deep learning [Deepsurv]) were trained in 100-times repeated 5-fold cross-validation (rCV) to predict PFS on three inputs: (i) clinicopathological variables, (ii) all radiomics-based and clinicopathological (full input), and (iii) uncorrelated radiomics-based and clinicopathological variables (uncorrelated input). The Models' performances were evaluated using the concordance index (c-index). Overall, 140 patients were included (median age: 62.5 years, 36.4% women). In rCV, the highest c-index was reached with Deepsurv (c-index = 0.631, 95%CI = 0.625-0.647), followed by GBM (c-index = 0.603, 95%CI = 0.557-0.646), significantly outperforming standard SCR whatever its input (c-index range: 0.560-0.570, all p < 0.0001). Thus, single- and multi-site pre-treatment radiomics data provide valuable prognostic information for predicting PFS in MLUAD patients undergoing first-line CPI treatment when analyzed with advanced machine-learning survival algorithms.

PMID:39001553 | DOI:10.3390/cancers16132491

Categories: Literature Watch

Integrating Omics Data and AI for Cancer Diagnosis and Prognosis

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jul 3;16(13):2448. doi: 10.3390/cancers16132448.

ABSTRACT

Cancer is one of the leading causes of death, making timely diagnosis and prognosis very important. Utilization of AI (artificial intelligence) enables providers to organize and process patient data in a way that can lead to better overall outcomes. This review paper aims to look at the varying uses of AI for diagnosis and prognosis and clinical utility. PubMed and EBSCO databases were utilized for finding publications from 1 January 2020 to 22 December 2023. Articles were collected using key search terms such as "artificial intelligence" and "machine learning." Included in the collection were studies of the application of AI in determining cancer diagnosis and prognosis using multi-omics data, radiomics, pathomics, and clinical and laboratory data. The resulting 89 studies were categorized into eight sections based on the type of data utilized and then further subdivided into two subsections focusing on cancer diagnosis and prognosis, respectively. Eight studies integrated more than one form of omics, namely genomics, transcriptomics, epigenomics, and proteomics. Incorporating AI into cancer diagnosis and prognosis alongside omics and clinical data represents a significant advancement. Given the considerable potential of AI in this domain, ongoing prospective studies are essential to enhance algorithm interpretability and to ensure safe clinical integration.

PMID:39001510 | DOI:10.3390/cancers16132448

Categories: Literature Watch

Pages