Deep learning

Driving risk identification of urban arterial and collector roads based on multi-scale data

Sat, 2024-07-13 06:00

Accid Anal Prev. 2024 Jul 12;206:107712. doi: 10.1016/j.aap.2024.107712. Online ahead of print.

ABSTRACT

Urban arterial and collector roads, while interconnected within the urban transportation network, serve distinct purposes, leading to different driving risk profiles. Investigating these differences using advanced methods is of paramount significance. This study aims to achieve this by primarily collecting and processing relevant vehicle trajectory data alongside driver-vehicle-road-environment data. A comprehensive risk assessment matrix is constructed to assess driving risks, incorporating multiple conflict and traffic flow indicators with statistically temporal stability. The Entropy weight-TOPSIS method and the K-means algorithm are employed to determine the risk scores and levels of the target arterial and collector roads. Using risk levels as the outcome variables and multi-scale features as the explanatory variables, random parameters models with heterogeneity in means and variances are developed to identify the determinants of driving risks at different levels. Likelihood ratio tests and comparisons of out-of-sample and within-sample prediction are conducted. Results reveal significant statistical differences in the risk profiles between arterial and collector roads. The marginal effects of significant parameters are then calculated separately for arterial and collector roads, indicating that several factors have different impacts on the probability of risk levels for arterial and collector roads, such as the number of movable elements in road landscape pictures, the standard deviation of the vehicle's lateral acceleration, the average standard deviation of speed for all vehicles on the road segment, and the number of one-way lanes on the road segment. Some practical implications are provided based on the findings. Future research can be implemented by expanding the collected data to different regions and cities over longer periods.

PMID:39002352 | DOI:10.1016/j.aap.2024.107712

Categories: Literature Watch

deepbet: Fast brain extraction of T1-weighted MRI using Convolutional Neural Networks

Sat, 2024-07-13 06:00

Comput Biol Med. 2024 Jul 12;179:108845. doi: 10.1016/j.compbiomed.2024.108845. Online ahead of print.

ABSTRACT

BACKGROUND: Brain extraction in magnetic resonance imaging (MRI) data is an important segmentation step in many neuroimaging preprocessing pipelines. Image segmentation is one of the research fields in which deep learning had the biggest impact in recent years. Consequently, traditional brain extraction methods are now being replaced by deep learning-based methods.

METHOD: Here, we used a unique dataset compilation comprising 7837 T1-weighted (T1w) MR images from 191 different OpenNeuro datasets in combination with advanced deep learning methods to build a fast, high-precision brain extraction tool called deepbet.

RESULTS: deepbet sets a novel state-of-the-art performance during cross-dataset validation with a median Dice score (DSC) of 99.0 on unseen datasets, outperforming the current best performing deep learning (DSC=97.9) and classic (DSC=96.5) methods. While current methods are more sensitive to outliers, deepbet achieves a Dice score of >97.4 across all 7837 images from 191 different datasets. This robustness was additionally tested in 5 external datasets, which included challenging clinical MR images. During visual exploration of each method's output which resulted in the lowest Dice score, major errors could be found for all of the tested tools except deepbet. Finally, deepbet uses a compute efficient variant of the UNet architecture, which accelerates brain extraction by a factor of ≈10 compared to current methods, enabling the processing of one image in ≈2 s on low level hardware.

CONCLUSIONS: In conclusion, deepbet demonstrates superior performance and reliability in brain extraction across a wide range of T1w MR images of adults, outperforming existing top tools. Its high minimal Dice score and minimal objective errors, even in challenging conditions, validate deepbet as a highly dependable tool for accurate brain extraction. deepbet can be conveniently installed via "pip install deepbet" and is publicly accessible at https://github.com/wwu-mmll/deepbet.

PMID:39002314 | DOI:10.1016/j.compbiomed.2024.108845

Categories: Literature Watch

Predicting recovery following stroke: Deep learning, multimodal data and feature selection using explainable AI

Sat, 2024-07-13 06:00

Neuroimage Clin. 2024 Jul 2;43:103638. doi: 10.1016/j.nicl.2024.103638. Online ahead of print.

ABSTRACT

Machine learning offers great potential for automated prediction of post-stroke symptoms and their response to rehabilitation. Major challenges for this endeavour include the very high dimensionality of neuroimaging data, the relatively small size of the datasets available for learning and interpreting the predictive features, as well as, how to effectively combine neuroimaging and tabular data (e.g. demographic information and clinical characteristics). This paper evaluates several solutions based on two strategies. The first is to use 2D images that summarise MRI scans. The second is to select key features that improve classification accuracy. Additionally, we introduce the novel approach of training a convolutional neural network (CNN) on images that combine regions-of-interests (ROIs) extracted from MRIs, with symbolic representations of tabular data. We evaluate a series of CNN architectures (both 2D and a 3D) that are trained on different representations of MRI and tabular data, to predict whether a composite measure of post-stroke spoken picture description ability is in the aphasic or non-aphasic range. MRI and tabular data were acquired from 758 English speaking stroke survivors who participated in the PLORAS study. Each participant was assigned to one of five different groups that were matched for initial severity of symptoms, recovery time, left lesion size and the months or years post-stroke that spoken description scores were collected. Training and validation were carried out on the first four groups. The fifth (lock-box/test set) group was used to test how well model accuracy generalises to new (unseen) data. The classification accuracy for a baseline logistic regression was 0.678 based on lesion size alone, rising to 0.757 and 0.813 when initial symptom severity and recovery time were successively added. The highest classification accuracy (0.854), area under the curve (0.899) and F1 score (0.901) were observed when 8 regions of interest were extracted from each MRI scan and combined with lesion size, initial severity and recovery time in a 2D Residual Neural Network (ResNet). This was also the best model when data were limited to the 286 participants with moderate or severe initial aphasia (with area under curve = 0.865), a group that would be considered more difficult to classify. Our findings demonstrate how imaging and tabular data can be combined to achieve high post-stroke classification accuracy, even when the dataset is small in machine learning terms. We conclude by proposing how the current models could be improved to achieve even higher levels of accuracy using images from hospital scanners.

PMID:39002223 | DOI:10.1016/j.nicl.2024.103638

Categories: Literature Watch

Needle tracking in low-resolution ultrasound volumes using deep learning

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03234-8. Online ahead of print.

ABSTRACT

PURPOSE: Clinical needle insertion into tissue, commonly assisted by 2D ultrasound imaging for real-time navigation, faces the challenge of precise needle and probe alignment to reduce out-of-plane movement. Recent studies investigate 3D ultrasound imaging together with deep learning to overcome this problem, focusing on acquiring high-resolution images to create optimal conditions for needle tip detection. However, high-resolution also requires a lot of time for image acquisition and processing, which limits the real-time capability. Therefore, we aim to maximize the US volume rate with the trade-off of low image resolution. We propose a deep learning approach to directly extract the 3D needle tip position from sparsely sampled US volumes.

METHODS: We design an experimental setup with a robot inserting a needle into water and chicken liver tissue. In contrast to manual annotation, we assess the needle tip position from the known robot pose. During insertion, we acquire a large data set of low-resolution volumes using a 16 × 16 element matrix transducer with a volume rate of 4 Hz. We compare the performance of our deep learning approach with conventional needle segmentation.

RESULTS: Our experiments in water and liver show that deep learning outperforms the conventional approach while achieving sub-millimeter accuracy. We achieve mean position errors of 0.54 mm in water and 1.54 mm in liver for deep learning.

CONCLUSION: Our study underlines the strengths of deep learning to predict the 3D needle positions from low-resolution ultrasound volumes. This is an important milestone for real-time needle navigation, simplifying the alignment of needle and ultrasound probe and enabling a 3D motion analysis.

PMID:39002100 | DOI:10.1007/s11548-024-03234-8

Categories: Literature Watch

3D mobile regression vision transformer for collateral imaging in acute ischemic stroke

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03229-5. Online ahead of print.

ABSTRACT

PURPOSE: The accurate and timely assessment of the collateral perfusion status is crucial in the diagnosis and treatment of patients with acute ischemic stroke. Previous works have shown that collateral imaging, derived from CT angiography, MR perfusion, and MR angiography, aids in evaluating the collateral status. However, such methods are time-consuming and/or sub-optimal due to the nature of manual processing and heuristics. Recently, deep learning approaches have shown to be promising for generating collateral imaging. These, however, suffer from the computational complexity and cost.

METHODS: In this study, we propose a mobile, lightweight deep regression neural network for collateral imaging in acute ischemic stroke, leveraging dynamic susceptibility contrast MR perfusion (DSC-MRP). Built based upon lightweight convolution and Transformer architectures, the proposed model manages the balance between the model complexity and performance.

RESULTS: We evaluated the performance of the proposed model in generating the five-phase collateral maps, including arterial, capillary, early venous, late venous, and delayed phases, using DSC-MRP from 952 patients. In comparison with various deep learning models, the proposed method was superior to the competitors with similar complexity and was comparable to the competitors of high complexity.

CONCLUSION: The results suggest that the proposed model is able to facilitate rapid and precise assessment of the collateral status of patients with acute ischemic stroke, leading to improved patient care and outcome.

PMID:39002099 | DOI:10.1007/s11548-024-03229-5

Categories: Literature Watch

Domain adaptation using AdaBN and AdaIN for high-resolution IVD mesh reconstruction from clinical MRI

Sat, 2024-07-13 06:00

Int J Comput Assist Radiol Surg. 2024 Jul 13. doi: 10.1007/s11548-024-03233-9. Online ahead of print.

ABSTRACT

PURPOSE: Deep learning has firmly established its dominance in medical imaging applications. However, careful consideration must be exercised when transitioning a trained source model to adapt to an entirely distinct environment that deviates significantly from the training set. The majority of the efforts to mitigate this issue have predominantly focused on classification and segmentation tasks. In this work, we perform a domain adaptation of a trained source model to reconstruct high-resolution intervertebral disc meshes from low-resolution MRI.

METHODS: To address the outlined challenges, we use MRI2Mesh as the shape reconstruction network. It incorporates three major modules: image encoder, mesh deformation, and cross-level feature fusion. This feature fusion module is used to encapsulate local and global disc features. We evaluate two major domain adaptation techniques: adaptive batch normalization (AdaBN) and adaptive instance normalization (AdaIN) for the task of shape reconstruction.

RESULTS: Experiments conducted on distinct datasets, including data from different populations, machines, and test sites demonstrate the effectiveness of MRI2Mesh for domain adaptation. MRI2Mesh achieved up to a 14% decrease in Hausdorff distance (HD) and a 19% decrease in the point-to-surface (P2S) metric for both AdaBN and AdaIN experiments, indicating improved performance.

CONCLUSION: MRI2Mesh has demonstrated consistent superiority to the state-of-the-art Voxel2Mesh network across a diverse range of datasets, populations, and scanning protocols, highlighting its versatility. Additionally, AdaBN has emerged as a robust method compared to the AdaIN technique. Further experiments show that MRI2Mesh, when combined with AdaBN, holds immense promise for enhancing the precision of anatomical shape reconstruction in domain adaptation.

PMID:39002098 | DOI:10.1007/s11548-024-03233-9

Categories: Literature Watch

The use of artificial intelligence in musculoskeletal ultrasound: a systematic review of the literature

Sat, 2024-07-13 06:00

Radiol Med. 2024 Jul 13. doi: 10.1007/s11547-024-01856-1. Online ahead of print.

ABSTRACT

PURPOSE: To systematically review the use of artificial intelligence (AI) in musculoskeletal (MSK) ultrasound (US) with an emphasis on AI algorithm categories and validation strategies.

MATERIAL AND METHODS: An electronic literature search was conducted for articles published up to January 2024. Inclusion criteria were the use of AI in MSK US, involvement of humans, English language, and ethics committee approval.

RESULTS: Out of 269 identified papers, 16 studies published between 2020 and 2023 were included. The research was aimed at predicting diagnosis and/or segmentation in a total of 11 (69%) out of 16 studies. A total of 11 (69%) studies used deep learning (DL)-based algorithms, three (19%) studies employed conventional machine learning (ML)-based algorithms, and two (12%) studies employed both conventional ML- and DL-based algorithms. Six (38%) studies used cross-validation techniques with K-fold cross-validation being the most frequently employed (n = 4, 25%). Clinical validation with separate internal test datasets was reported in nine (56%) papers. No external clinical validation was reported.

CONCLUSION: AI is a topic of increasing interest in MSK US research. In future studies, attention should be paid to the use of validation strategies, particularly regarding independent clinical validation performed on external datasets.

PMID:39001961 | DOI:10.1007/s11547-024-01856-1

Categories: Literature Watch

Deep learning model based on endoscopic images predicting treatment response in locally advanced rectal cancer undergo neoadjuvant chemoradiotherapy: a multicenter study

Sat, 2024-07-13 06:00

J Cancer Res Clin Oncol. 2024 Jul 13;150(7):350. doi: 10.1007/s00432-024-05876-2.

ABSTRACT

PURPOSE: Neoadjuvant chemoradiotherapy has been the standard practice for patients with locally advanced rectal cancer. However, the treatment response varies greatly among individuals, how to select the optimal candidates for neoadjuvant chemoradiotherapy is crucial. This study aimed to develop an endoscopic image-based deep learning model for predicting the response to neoadjuvant chemoradiotherapy in locally advanced rectal cancer.

METHODS: In this multicenter observational study, pre-treatment endoscopic images of patients from two Chinese medical centers were retrospectively obtained and a deep learning-based tumor regression model was constructed. Treatment response was evaluated based on the tumor regression grade and was defined as good response and non-good response. The prediction performance of the deep learning model was evaluated in the internal and external test sets. The main outcome was the accuracy of the treatment prediction model, measured by the AUC and accuracy.

RESULTS: This deep learning model achieved favorable prediction performance. In the internal test set, the AUC and accuracy were 0.867 (95% CI: 0.847-0.941) and 0.836 (95% CI: 0.818-0.896), respectively. The prediction performance was fully validated in the external test set, and the model had an AUC of 0.758 (95% CI: 0.724-0.834) and an accuracy of 0.807 (95% CI: 0.774-0.843).

CONCLUSION: The deep learning model based on endoscopic images demonstrated exceptional predictive power for neoadjuvant treatment response, highlighting its potential for guiding personalized therapy.

PMID:39001926 | DOI:10.1007/s00432-024-05876-2

Categories: Literature Watch

Artificial intelligence based diagnosis of sulcus: assesment of videostroboscopy via deep learning

Sat, 2024-07-13 06:00

Eur Arch Otorhinolaryngol. 2024 Jul 13. doi: 10.1007/s00405-024-08801-y. Online ahead of print.

ABSTRACT

PURPOSE: To develop a convolutional neural network (CNN)-based model for classifying videostroboscopic images of patients with sulcus, benign vocal fold (VF) lesions, and healthy VFs to improve clinicians' accuracy in diagnosis during videostroboscopies when evaluating sulcus.

MATERIALS AND METHODS: Videostroboscopies of 433 individuals who were diagnosed with sulcus (91), who were diagnosed with benign VF diseases (i.e., polyp, nodule, papilloma, cyst, or pseudocyst [311]), or who were healthy (33) were analyzed. After extracting 91,159 frames from videostroboscopies, a CNN-based model was created and tested. The healthy and sulcus groups underwent binary classification. In the second phase of the study, benign VF lesions were added to the training set, and multiclassification was executed across all groups. The proposed CNN-based model results were compared with five laryngology experts' assessments.

RESULTS: In the binary classification phase, the CNN-based model achieved 98% accuracy, 98% recall, 97% precision, and a 97% F1 score for classifying sulcus and healthy VFs. During the multiclassification phase, when evaluated on a subset of frames encompassing all included groups, the CNN-based model demonstrated greater accuracy when compared with that of the five laryngologists (%76 versus 72%, 68%, 72%, 63%, and 72%).

CONCLUSION: The utilization of a CNN-based model serves as a significant aid in the diagnosis of sulcus, a VF disease that presents notable challenges in the diagnostic process. Further research could be undertaken to assess the practicality of implementing this approach in real-time application in clinical practice.

PMID:39001913 | DOI:10.1007/s00405-024-08801-y

Categories: Literature Watch

Three-Dimensional Label-Free Observing of the Self-Assembled Nanoparticles inside a Single Cell at Nanoscale Resolution

Sat, 2024-07-13 06:00

ACS Nano. 2024 Jul 13. doi: 10.1021/acsnano.4c06095. Online ahead of print.

ABSTRACT

Understanding the intracellular behavior of nanoparticles (NPs) plays a key role in optimizing the self-assembly performance of nanomedicine. However, conducting the 3D, label-free, quantitative observation of self-assembled NPs within intact single cells remains a substantial challenge in complicated intracellular environments. Here, we propose a deep learning combined synchrotron radiation hard X-ray nanotomography approach to visualize the self-assembled ultrasmall iron oxide (USIO) NPs in a single cell. The method allows us to explore comprehensive information on NPs, such as their distribution, morphology, location, and interaction with cell organelles, and provides quantitative analysis of the heterogeneous size and morphologies of USIO NPs under diverse conditions. This label-free, in situ method provides a tool for precise characterization of intracellular self-assembled NPs to improve the evaluation and design of a bioresponsive nanomedicine.

PMID:39001860 | DOI:10.1021/acsnano.4c06095

Categories: Literature Watch

Current status and future directions in artificial intelligence for nuclear cardiology

Sat, 2024-07-13 06:00

Expert Rev Cardiovasc Ther. 2024 Jul 13. doi: 10.1080/14779072.2024.2380764. Online ahead of print.

ABSTRACT

INTRODUCTION: Myocardial perfusion imaging (MPI) is one of the most commonly ordered cardiac imaging tests. Accurate motion correction, image registration, and reconstruction is critical for high-quality imaging, but this can be technically challenging and traditionally has relied on expert manual processing. With accurate processing, there is a rich variety of clinical, stress, functional, and anatomic data that can be integrated to guide patient management.

AREAS COVERED: Pubmed and Google Scholar were reviewed for articles related to artificial intelligence in nuclear cardiology published between 2020 and 2024. We will outline the prominent roles for artificial intelligence (AI) solutions to provide motion correction, image registration, and reconstruction. We will review the role for AI in extracting anatomic data for hybrid MPI which is otherwise neglected. Lastly, we will discuss AI methods to integrate the wealth of data to improve disease diagnosis or risk stratification.

EXPERT OPINION: There is growing evidence that AI will transform the performance of MPI by automating and improving on aspects of image acquisition and reconstruction. Physicians and researchers will need to understand the potential strengths of AI in order to benefit from the full clinical utility of MPI.

PMID:39001698 | DOI:10.1080/14779072.2024.2380764

Categories: Literature Watch

Enhancing Immunotherapy Response Prediction in Metastatic Lung Adenocarcinoma: Leveraging Shallow and Deep Learning with CT-Based Radiomics across Single and Multiple Tumor Sites

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jul 8;16(13):2491. doi: 10.3390/cancers16132491.

ABSTRACT

This study aimed to evaluate the potential of pre-treatment CT-based radiomics features (RFs) derived from single and multiple tumor sites, and state-of-the-art machine-learning survival algorithms, in predicting progression-free survival (PFS) for patients with metastatic lung adenocarcinoma (MLUAD) receiving first-line treatment including immune checkpoint inhibitors (CPIs). To do so, all adults with newly diagnosed MLUAD, pre-treatment contrast-enhanced CT scan, and performance status ≤ 2 who were treated at our cancer center with first-line CPI between November 2016 and November 2022 were included. RFs were extracted from all measurable lesions with a volume ≥ 1 cm3 on the CT scan. To capture intra- and inter-tumor heterogeneity, RFs from the largest tumor of each patient, as well as lowest, highest, and average RF values over all lesions per patient were collected. Intra-patient inter-tumor heterogeneity metrics were calculated to measure the similarity between each patient lesions. After filtering predictors with univariable Cox p < 0.100 and analyzing their correlations, five survival machine-learning algorithms (stepwise Cox regression [SCR], LASSO Cox regression, random survival forests, gradient boosted machine [GBM], and deep learning [Deepsurv]) were trained in 100-times repeated 5-fold cross-validation (rCV) to predict PFS on three inputs: (i) clinicopathological variables, (ii) all radiomics-based and clinicopathological (full input), and (iii) uncorrelated radiomics-based and clinicopathological variables (uncorrelated input). The Models' performances were evaluated using the concordance index (c-index). Overall, 140 patients were included (median age: 62.5 years, 36.4% women). In rCV, the highest c-index was reached with Deepsurv (c-index = 0.631, 95%CI = 0.625-0.647), followed by GBM (c-index = 0.603, 95%CI = 0.557-0.646), significantly outperforming standard SCR whatever its input (c-index range: 0.560-0.570, all p < 0.0001). Thus, single- and multi-site pre-treatment radiomics data provide valuable prognostic information for predicting PFS in MLUAD patients undergoing first-line CPI treatment when analyzed with advanced machine-learning survival algorithms.

PMID:39001553 | DOI:10.3390/cancers16132491

Categories: Literature Watch

Integrating Omics Data and AI for Cancer Diagnosis and Prognosis

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jul 3;16(13):2448. doi: 10.3390/cancers16132448.

ABSTRACT

Cancer is one of the leading causes of death, making timely diagnosis and prognosis very important. Utilization of AI (artificial intelligence) enables providers to organize and process patient data in a way that can lead to better overall outcomes. This review paper aims to look at the varying uses of AI for diagnosis and prognosis and clinical utility. PubMed and EBSCO databases were utilized for finding publications from 1 January 2020 to 22 December 2023. Articles were collected using key search terms such as "artificial intelligence" and "machine learning." Included in the collection were studies of the application of AI in determining cancer diagnosis and prognosis using multi-omics data, radiomics, pathomics, and clinical and laboratory data. The resulting 89 studies were categorized into eight sections based on the type of data utilized and then further subdivided into two subsections focusing on cancer diagnosis and prognosis, respectively. Eight studies integrated more than one form of omics, namely genomics, transcriptomics, epigenomics, and proteomics. Incorporating AI into cancer diagnosis and prognosis alongside omics and clinical data represents a significant advancement. Given the considerable potential of AI in this domain, ongoing prospective studies are essential to enhance algorithm interpretability and to ensure safe clinical integration.

PMID:39001510 | DOI:10.3390/cancers16132448

Categories: Literature Watch

Deep Learning Histology for Prediction of Lymph Node Metastases and Tumor Regression after Neoadjuvant FLOT Therapy of Gastroesophageal Adenocarcinoma

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jul 3;16(13):2445. doi: 10.3390/cancers16132445.

ABSTRACT

BACKGROUND: The aim of this study was to establish a deep learning prediction model for neoadjuvant FLOT chemotherapy response. The neural network utilized clinical data and visual information from whole-slide images (WSIs) of therapy-naïve gastroesophageal cancer biopsies.

METHODS: This study included 78 patients from the University Hospital of Cologne and 59 patients from the University Hospital of Heidelberg used as external validation.

RESULTS: After surgical resection, 33 patients from Cologne (42.3%) were ypN0 and 45 patients (57.7%) were ypN+, while 23 patients from Heidelberg (39.0%) were ypN0 and 36 patients (61.0%) were ypN+ (p = 0.695). The neural network had an accuracy of 92.1% to predict lymph node metastasis and the area under the curve (AUC) was 0.726. A total of 43 patients from Cologne (55.1%) had less than 50% residual vital tumor (RVT) compared to 34 patients from Heidelberg (57.6%, p = 0.955). The model was able to predict tumor regression with an error of ±14.1% and an AUC of 0.648.

CONCLUSIONS: This study demonstrates that visual features extracted by deep learning from therapy-naïve biopsies of gastroesophageal adenocarcinomas correlate with positive lymph nodes and tumor regression. The results will be confirmed in prospective studies to achieve early allocation of patients to the most promising treatment.

PMID:39001507 | DOI:10.3390/cancers16132445

Categories: Literature Watch

Improved Pancreatic Cancer Detection and Localization on CT Scans: A Computer-Aided Detection Model Utilizing Secondary Features

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jun 29;16(13):2403. doi: 10.3390/cancers16132403.

ABSTRACT

The early detection of pancreatic ductal adenocarcinoma (PDAC) is essential for optimal treatment of pancreatic cancer patients. We propose a tumor detection framework to improve the detection of pancreatic head tumors on CT scans. In this retrospective research study, CT images of 99 patients with pancreatic head cancer and 98 control cases from the Catharina Hospital Eindhoven were collected. A multi-stage 3D U-Net-based approach was used for PDAC detection including clinically significant secondary features such as pancreatic duct and common bile duct dilation. The developed algorithm was evaluated using a local test set comprising 59 CT scans. The model was externally validated in 28 pancreatic cancer cases of a publicly available medical decathlon dataset. The tumor detection framework achieved a sensitivity of 0.97 and a specificity of 1.00, with an area under the receiver operating curve (AUROC) of 0.99, in detecting pancreatic head cancer in the local test set. In the external test set, we obtained similar results, with a sensitivity of 1.00. The model provided the tumor location with acceptable accuracy obtaining a DICE Similarity Coefficient (DSC) of 0.37. This study shows that a tumor detection framework utilizing CT scans and secondary signs of pancreatic cancer can detect pancreatic tumors with high accuracy.

PMID:39001465 | DOI:10.3390/cancers16132403

Categories: Literature Watch

Outcome Prediction Using Multi-Modal Information: Integrating Large Language Model-Extracted Clinical Information and Image Analysis

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jun 29;16(13):2402. doi: 10.3390/cancers16132402.

ABSTRACT

Survival prediction post-cystectomy is essential for the follow-up care of bladder cancer patients. This study aimed to evaluate artificial intelligence (AI)-large language models (LLMs) for extracting clinical information and improving image analysis, with an initial application involving predicting five-year survival rates of patients after radical cystectomy for bladder cancer. Data were retrospectively collected from medical records and CT urograms (CTUs) of bladder cancer patients between 2001 and 2020. Of 781 patients, 163 underwent chemotherapy, had pre- and post-chemotherapy CTUs, underwent radical cystectomy, and had an available post-surgery five-year survival follow-up. Five AI-LLMs (Dolly-v2, Vicuna-13b, Llama-2.0-13b, GPT-3.5, and GPT-4.0) were used to extract clinical descriptors from each patient's medical records. As a reference standard, clinical descriptors were also extracted manually. Radiomics and deep learning descriptors were extracted from CTU images. The developed multi-modal predictive model, CRD, was based on the clinical (C), radiomics (R), and deep learning (D) descriptors. The LLM retrieval accuracy was assessed. The performances of the survival predictive models were evaluated using AUC and Kaplan-Meier analysis. For the 163 patients (mean age 64 ± 9 years; M:F 131:32), the LLMs achieved extraction accuracies of 74%~87% (Dolly), 76%~83% (Vicuna), 82%~93% (Llama), 85%~91% (GPT-3.5), and 94%~97% (GPT-4.0). For a test dataset of 64 patients, the CRD model achieved AUCs of 0.89 ± 0.04 (manually extracted information), 0.87 ± 0.05 (Dolly), 0.83 ± 0.06~0.84 ± 0.05 (Vicuna), 0.81 ± 0.06~0.86 ± 0.05 (Llama), 0.85 ± 0.05~0.88 ± 0.05 (GPT-3.5), and 0.87 ± 0.05~0.88 ± 0.05 (GPT-4.0). This study demonstrates the use of LLM model-extracted clinical information, in conjunction with imaging analysis, to improve the prediction of clinical outcomes, with bladder cancer as an initial example.

PMID:39001463 | DOI:10.3390/cancers16132402

Categories: Literature Watch

Deep Learning Algorithms for Bladder Cancer Segmentation on Multi-Parametric MRI

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jun 26;16(13):2348. doi: 10.3390/cancers16132348.

ABSTRACT

BACKGROUND: Bladder cancer (BC) segmentation on MRI images is the first step to determining the presence of muscular invasion. This study aimed to assess the tumor segmentation performance of three deep learning (DL) models on multi-parametric MRI (mp-MRI) images.

METHODS: We studied 53 patients with bladder cancer. Bladder tumors were segmented on each slice of T2-weighted (T2WI), diffusion-weighted imaging/apparent diffusion coefficient (DWI/ADC), and T1-weighted contrast-enhanced (T1WI) images acquired at a 3Tesla MRI scanner. We trained Unet, MAnet, and PSPnet using three loss functions: cross-entropy (CE), dice similarity coefficient loss (DSC), and focal loss (FL). We evaluated the model performances using DSC, Hausdorff distance (HD), and expected calibration error (ECE).

RESULTS: The MAnet algorithm with the CE+DSC loss function gave the highest DSC values on the ADC, T2WI, and T1WI images. PSPnet with CE+DSC obtained the smallest HDs on the ADC, T2WI, and T1WI images. The segmentation accuracy overall was better on the ADC and T1WI than on the T2WI. The ECEs were the smallest for PSPnet with FL on the ADC images, while they were the smallest for MAnet with CE+DSC on the T2WI and T1WI.

CONCLUSIONS: Compared to Unet, MAnet and PSPnet with a hybrid CE+DSC loss function displayed better performances in BC segmentation depending on the choice of the evaluation metric.

PMID:39001410 | DOI:10.3390/cancers16132348

Categories: Literature Watch

Quantification and Profiling of Early and Late Differentiation Stage T Cells in Mantle Cell Lymphoma Reveals Immunotherapeutic Targets in Subsets of Patients

Sat, 2024-07-13 06:00

Cancers (Basel). 2024 Jun 21;16(13):2289. doi: 10.3390/cancers16132289.

ABSTRACT

With the aim to advance the understanding of immune regulation in MCL and to identify targetable T-cell subsets, we set out to combine image analysis and spatial omic technology focused on both early and late differentiation stages of T cells. MCL patient tissue (n = 102) was explored using image analysis and GeoMx spatial omics profiling of 69 proteins and 1812 mRNAs. Tumor cells, T helper (TH) cells and cytotoxic (TC) cells of early (CD57-) and late (CD57+) differentiation stage were analyzed. An image analysis workflow was developed based on fine-tuned Cellpose models for cell segmentation and classification. TC and CD57+ subsets of T cells were enriched in tumor-rich compared to tumor-sparse regions. Tumor-sparse regions had a higher expression of several key immune suppressive proteins, tentatively controlling T-cell expansion in regions close to the tumor. We revealed that T cells in late differentiation stages (CD57+) are enriched among MCL infiltrating T cells and are predictive of an increased expression of immune suppressive markers. CD47, IDO1 and CTLA-4 were identified as potential targets for patients with T-cell-rich MCL TIME, while GITR might be a feasible target for MCL patients with sparse T-cell infiltration. In subgroups of patients with a high degree of CD57+ TC-cell infiltration, several immune checkpoint inhibitors, including TIGIT, PD-L1 and LAG3 were increased, emphasizing the immune-suppressive features of this highly differentiated T-cell subset not previously described in MCL.

PMID:39001353 | DOI:10.3390/cancers16132289

Categories: Literature Watch

Evolving and Novel Applications of Artificial Intelligence in Thoracic Imaging

Sat, 2024-07-13 06:00

Diagnostics (Basel). 2024 Jul 8;14(13):1456. doi: 10.3390/diagnostics14131456.

ABSTRACT

The advent of artificial intelligence (AI) is revolutionizing medicine, particularly radiology. With the development of newer models, AI applications are demonstrating improved performance and versatile utility in the clinical setting. Thoracic imaging is an area of profound interest, given the prevalence of chest imaging and the significant health implications of thoracic diseases. This review aims to highlight the promising applications of AI within thoracic imaging. It examines the role of AI, including its contributions to improving diagnostic evaluation and interpretation, enhancing workflow, and aiding in invasive procedures. Next, it further highlights the current challenges and limitations faced by AI, such as the necessity of 'big data', ethical and legal considerations, and bias in representation. Lastly, it explores the potential directions for the application of AI in thoracic radiology.

PMID:39001346 | DOI:10.3390/diagnostics14131456

Categories: Literature Watch

Automated Laryngeal Invasion Detector of Boluses in Videofluoroscopic Swallowing Study Videos Using Action Recognition-Based Networks

Sat, 2024-07-13 06:00

Diagnostics (Basel). 2024 Jul 6;14(13):1444. doi: 10.3390/diagnostics14131444.

ABSTRACT

We aimed to develop an automated detector that determines laryngeal invasion during swallowing. Laryngeal invasion, which causes significant clinical problems, is defined as two or more points on the penetration-aspiration scale (PAS). We applied two three-dimensional (3D) stream networks for action recognition in videofluoroscopic swallowing study (VFSS) videos. To detect laryngeal invasion (PAS 2 or higher scores) in VFSS videos, we employed two 3D stream networks for action recognition. To establish the robustness of our model, we compared its performance with those of various current image classification-based architectures. The proposed model achieved an accuracy of 92.10%. Precision, recall, and F1 scores for detecting laryngeal invasion (≥PAS 2) in VFSS videos were 0.9470 each. The accuracy of our model in identifying laryngeal invasion surpassed that of other updated image classification models (60.58% for ResNet101, 60.19% for Swin-Transformer, 63.33% for EfficientNet-B2, and 31.17% for HRNet-W32). Our model is the first automated detector of laryngeal invasion in VFSS videos based on video action recognition networks. Considering its high and balanced performance, it may serve as an effective screening tool before clinicians review VFSS videos, ultimately reducing the burden on clinicians.

PMID:39001334 | DOI:10.3390/diagnostics14131444

Categories: Literature Watch

Pages