Deep learning
Deep learning algorithm applied to plain CT images to identify superior mesenteric artery abnormalities
Eur J Radiol. 2024 Feb 23;173:111388. doi: 10.1016/j.ejrad.2024.111388. Online ahead of print.
ABSTRACT
OBJECTIVES: Atypical presentations, lack of biomarkers, and low sensitivity of plain CT can delay the diagnosis of superior mesenteric artery (SMA) abnormalities, resulting in poor clinical outcomes. Our study aims to develop a deep learning (DL) model for detecting SMA abnormalities in plain CT and evaluate its performance in comparison with a clinical model and radiologist assessment.
MATERIALS AND METHODS: A total of 1048 patients comprised the internal (474 patients with SMA abnormalities, 474 controls) and external testing (50 patients with SMA abnormalities, 50 controls) cohorts. The internal cohort was divided into the training cohort (n = 776), validation cohort (n = 86), and internal testing cohort (n = 86). A total of 5 You Only Look Once version 8 (YOLOv8)-based DL submodels were developed, and the performance of the optimal submodel was compared with that of a clinical model and of experienced radiologists.
RESULTS: Of the submodels, YOLOv8x had the best performance. The area under the curve (AUC) of the YOLOv8x submodel was higher than that of the clinical model (internal test set: 0.990 vs 0.878, P =.002; external test set: 0.967 vs 0.912, P =.140) and that of all radiologists (P <.001). The YOLOv8x submodel, when compared with radiologist assessment, demonstrated higher sensitivity (internal test set: 100.0 % vs 70.7 %, P =.002; external test set: 96.0 % vs 68.8 %, P <.001) and specificity (internal test set: 90.7 % vs 66.0 %, P =.025; external test set: = 88.0 % vs 66.0 %, P <.001).
CONCLUSION: Using plain CT images, YOLOv8x was able to efficiently identify cases of SMA abnormalities. This could potentially improve early diagnosis accuracy and thus improve clinical outcomes.
PMID:38412582 | DOI:10.1016/j.ejrad.2024.111388
A deep learning based holistic diagnosis system for immunohistochemistry interpretation and molecular subtyping
Neoplasia. 2024 Feb 26;50:100976. doi: 10.1016/j.neo.2024.100976. Online ahead of print.
ABSTRACT
BACKGROUND: Breast cancer in different molecular subtypes, which is determined by the overexpression rates of human epidermal growth factor receptor 2 (HER2), estrogen receptor (ER), progesterone receptor (PR), and Ki67, exhibit distinct symptom characteristics and sensitivity to different treatment. The immunohistochemical method, one of the most common detecting tools for tumour markers, is heavily relied on artificial judgment and in clinical practice, with an inherent limitation in interpreting stability and operating efficiency. Here, a holistic intelligent breast tumour diagnosis system has been developed for tumour-markeromic analysis, combining the automatic interpretation and clinical suggestion.
METHODS: The holistic intelligent breast tumour diagnosis system included two main modules. The interpreting modules were constructed based on convolutional neural network, for comprehensively extracting and analyzing the multi-features of immunostaining. Referring to the clinical classification criteria, the interpreting results were encoded in a low-dimensional feature representation in the subtyping module, to efficiently output a holistic detecting result of the critical tumour-markeromic with diagnosis suggestions on molecular subtypes.
RESULTS: The overexpression rates of HER2, ER, PR, and Ki67, as well as an effective determination of molecular subtypes were successfully obtained by this diagnosis system, with an average sensitivity of 97.6 % and an average specificity of 96.1 %, among those, the sensitivity and specificity for interpreting HER2 were up to 99.8 % and 96.9 %.
CONCLUSION: The holistic intelligent breast tumour diagnosis system shows improved performance in the interpretation of immunohistochemical images over pathologist-level, which can be expected to overcome the limitations of conventional manual interpretation in efficiency, precision, and repeatability.
PMID:38412576 | DOI:10.1016/j.neo.2024.100976
Comparison of clinical geneticist and computer visual attention in assessing genetic conditions
PLoS Genet. 2024 Feb 27;20(2):e1011168. doi: 10.1371/journal.pgen.1011168. Online ahead of print.
ABSTRACT
The use of artificial intelligence (AI) for facial diagnostics is increasingly used in the genetics clinic to evaluate patients with potential genetic conditions. Current approaches focus on one type of AI called Deep Learning (DL). While DL- based facial diagnostic platforms have a high accuracy rate for many conditions, less is understood about how this technology assesses and classifies (categorizes) images, and how this compares to humans. To compare human and computer attention, we performed eye-tracking analyses of geneticist clinicians (n = 22) and non-clinicians (n = 22) who viewed images of people with 10 different genetic conditions, as well as images of unaffected individuals. We calculated the Intersection-over-Union (IoU) and Kullback-Leibler divergence (KL) to compare the visual attentions of the two participant groups, and then the clinician group against the saliency maps of our deep learning classifier. We found that human visual attention differs greatly from DL model's saliency results. Averaging over all the test images, IoU and KL metric for the successful (accurate) clinician visual attentions versus the saliency maps were 0.15 and 11.15, respectively. Individuals also tend to have a specific pattern of image inspection, and clinicians demonstrate different visual attention patterns than non-clinicians (IoU and KL of clinicians versus non-clinicians were 0.47 and 2.73, respectively). This study shows that humans (at different levels of expertise) and computer vision model examine images differently. Understanding these differences can improve the design and use of AI tools, and lead to more meaningful interactions between clinicians and AI technologies.
PMID:38412177 | DOI:10.1371/journal.pgen.1011168
Source-Free Domain Adaptation With Domain Generalized Pretraining for Face Anti-Spoofing
IEEE Trans Pattern Anal Mach Intell. 2024 Feb 27;PP. doi: 10.1109/TPAMI.2024.3370721. Online ahead of print.
ABSTRACT
Source-free domain adaptation (SFDA) shows the potential to improve the generalizability of deep learning-based face anti-spoofing (FAS) while preserving the privacy and security of sensitive human faces. However, existing SFDA methods are significantly degraded without accessing source data due to the inability to mitigate domain and identity bias in FAS. In this paper, we propose a novel Source-free Domain Adaptation framework for FAS (SDA-FAS) that systematically addresses the challenges of source model pre-training, source knowledge adaptation, and target data exploration under the source-free setting. Specifically, we develop a generalized method for source model pre-training that leverages a causality-inspired PatchMix data augmentation to diminish domain bias and designs the patch-wise contrastive loss to alleviate identity bias. For source knowledge adaptation, we propose a contrastive domain alignment module to align conditional distribution across domains with a theoretical equivalence to adaptation based on source data. Furthermore, target data exploration is achieved via self-supervised learning with patch shuffle augmentation to identify unseen attack types, which is ignored in existing SFDA methods. To our best knowledge, this paper provides the first full-stack privacy-preserving framework to address the generalization problem in FAS. Extensive experiments on nineteen cross-dataset scenarios show our framework considerably outperforms state-of-the-art methods.
PMID:38412088 | DOI:10.1109/TPAMI.2024.3370721
Scalable Moment Propagation and Analysis of Variational Distributions for Practical Bayesian Deep Learning
IEEE Trans Neural Netw Learn Syst. 2024 Feb 27;PP. doi: 10.1109/TNNLS.2024.3367363. Online ahead of print.
ABSTRACT
Bayesian deep learning is one of the key frameworks employed in handling predictive uncertainty. Variational inference (VI), an extensively used inference method, derives the predictive distributions by Monte Carlo (MC) sampling. The drawback of MC sampling is its extremely high computational cost compared to that of ordinary deep learning. In contrast, the moment propagation (MP)-based approach propagates the output moments of each layer to derive predictive distributions instead of MC sampling. Because of this computational property, it is expected to realize faster inference than MC-based approaches. However, the applicability of the MP-based method in deep models has not been explored sufficiently, even though some studies have demonstrated the effectiveness of MP only in small toy models. One of the reasons is that it is difficult to train deep models by MP because of the large variance in activations. To realize MP in deep models, some normalization layers are required but have not yet been studied. In addition, it is still difficult to design well-calibrated MP-based models, because the effectiveness of MP-based methods under various variational distributions has also not been investigated. In this study, we propose a fast and reliable MP-based Bayesian deep-learning method. First, to train deep-learning models using MP, we introduce a batch normalization layer extended to random variables to prevent increases in the variance of activations. Second, to identify the appropriate variational distribution in MP, we investigate the treatment of moments of several variational distributions and evaluate their uncertainty quality of predictions. Experiments with regression tasks demonstrate that the MP-based method provides qualitatively and quantitatively equivalent predictive performance to MC-based methods regardless of variational distributions. In the classification tasks, we show that we can train MP-based deep models by extended batch normalization. We also show that the MP-based approach realizes 2.0-2.8 times faster inference than the MC-based approach while maintaining the predictive performance. The results of this study can help realize a fast and well-calibrated uncertainty estimation method that can be deployed in a wider range of reliability-aware applications.
PMID:38412086 | DOI:10.1109/TNNLS.2024.3367363
Leveraging Brain Modularity Prior for Interpretable Representation Learning of fMRI
IEEE Trans Biomed Eng. 2024 Feb 27;PP. doi: 10.1109/TBME.2024.3370415. Online ahead of print.
ABSTRACT
Resting-state functional magnetic resonance imaging (rs-fMRI) can reflect spontaneous neural activities in the brain and is widely used for brain disorder analysis. Previous studies focus on extracting fMRI representations using machine/deep learning methods, but these features typically lack biological interpretability. The human brain exhibits a remarkable modular structure in spontaneous brain functional networks, with each module comprised of functionally interconnected brain regions-of-interest (ROIs). However, existing learning-based methods cannot adequately utilize such brain modularity prior. In this paper, we propose a brain modularity-constrained dynamic representation learning framework for interpretable fMRI analysis, consisting of dynamic graph construction, dynamic graph learning via a novel modularity-constrained graph neural network (MGNN), and prediction and biomarker detection. The designed MGNN is constrained by three core neurocognitive modules (i.e., salience network, central executive network, and default mode network), encouraging ROIs within the same module to share similar representations. To further enhance discriminative ability of learned features, we encourage the MGNN to preserve network topology of input graphs via a graph topology reconstruction constraint. Experimental results on 534 subjects with rs-fMRI scans from two datasets validate the effectiveness of the proposed method. The identified discriminative brain ROIs and functional connectivities can be regarded as potential fMRI biomarkers to aid in clinical diagnosis.
PMID:38412079 | DOI:10.1109/TBME.2024.3370415
Real-Time Non-Invasive Imaging and Detection of Spreading Depolarizations through EEG: An Ultra-Light Explainable Deep Learning Approach
IEEE J Biomed Health Inform. 2024 Feb 27;PP. doi: 10.1109/JBHI.2024.3370502. Online ahead of print.
ABSTRACT
A core aim of neurocritical care is to prevent secondary brain injury. Spreading depolarizations (SDs) have been identified as an important independent cause of secondary brain injury. SDs are usually detected using invasive electrocorticography recorded at high sampling frequency. Recent pilot studies suggest a possible utility of scalp electrodes generated electroencephalogram (EEG) for non-invasive SD detection. However, noise and attenuation of EEG signals makes this detection task extremely challenging. Previous methods focus on detecting temporal power change of EEG over a fixed high-density map of scalp electrodes, which is not always clinically feasible. Having a specialized spectrogram as an input to the automatic SD detection model, this study is the first to transform SD identification problem from a detection task on a 1-D time-series wave to a task on a sequential 2-D rendered imaging. This study presented a novel ultra-light-weight multi-modal deep-learning network to fuse EEG spectrogram imaging and temporal power vectors to enhance SD identification accuracy over each single electrode, allowing flexible EEG map and paving the way for SD detection on ultra-low-density EEG with variable electrode positioning. Our proposed model has an ultra-fast processing speed (<0.3 sec). Compared to the conventional methods (2 hours), this is a huge advancement towards early SD detection and to facilitate instant brain injury prognosis. Seeing SDs with a new dimension - frequency on spectrograms, we demonstrated that such additional dimension could improve SD detection accuracy, providing preliminary evidence to support the hypothesis that SDs may show implicit features over the frequency profile.
PMID:38412076 | DOI:10.1109/JBHI.2024.3370502
Revealing the Denoising Principle of Zero-Shot N2N-Based Algorithm from 1D Spectrum to 2D Image
Anal Chem. 2024 Feb 27. doi: 10.1021/acs.analchem.3c04608. Online ahead of print.
ABSTRACT
Denoising is a necessary step in image analysis to extract weak signals, especially those hardly identified by the naked eye. Unlike the data-driven deep-learning denoising algorithms relying on a clean image as the reference, Noise2Noise (N2N) was able to denoise the noise image, providing sufficiently noise images with the same subject but randomly distributed noise. Further, by introducing data augmentation to create a big data set and regularization to prevent model overfitting, zero-shot N2N-based denoising was proposed in which only a single noisy image was needed. Although various N2N-based denoising algorithms have been developed with high performance, their complicated black box operation prevented the lightweight. Therefore, to reveal the working function of the zero-shot N2N-based algorithm, we proposed a lightweight Peak2Peak algorithm (P2P) and qualitatively and quantitatively analyzed its denoising behavior on the 1D spectrum and 2D image. We found that the high-performance denoising originates from the trade-off balance between the loss function and regularization in the denoising module, where regularization is the switch of denoising. Meanwhile, the signal extraction is mainly from the self-supervised characteristic learning in the data augmentation module. Further, the lightweight P2P improved the denoising speed by at least ten times but with little performance loss, compared with that of the current N2N-based algorithms. In general, the visualization of P2P provides a reference for revealing the working function of zero-shot N2N-based algorithms, which would pave the way for the application of these algorithms toward real-time (in situ, in vivo, and operando) research improving both temporal and spatial resolutions. The P2P is open-source at https://github.com/3331822w/Peak2Peakand will be accessible online access at https://ramancloud.xmu.edu.cn/tutorial.
PMID:38412039 | DOI:10.1021/acs.analchem.3c04608
Automated Machine Learning versus Expert-Designed Models in Ocular Toxoplasmosis: Detection and Lesion Localization Using Fundus Images
Ocul Immunol Inflamm. 2024 Feb 27:1-7. doi: 10.1080/09273948.2024.2319281. Online ahead of print.
ABSTRACT
PURPOSE: Automated machine learning (AutoML) allows clinicians without coding experience to build their own deep learning (DL) models. This study assesses the performance of AutoML in detecting and localizing ocular toxoplasmosis (OT) lesions in fundus images and compares it to expert-designed models.
METHODS: Ophthalmology trainees without coding experience designed AutoML models using 304 labelled fundus images. We designed a binary model to differentiate OT from normal and an object detection model to visually identify OT lesions.
RESULTS: The AutoML model had an area under the precision-recall curve (AuPRC) of 0.945, sensitivity of 100%, specificity of 83% and accuracy of 93.5% (vs. 94%, 86% and 91% for the bespoke models). The AutoML object detection model had an AuPRC of 0.600 with a precision of 93.3% and recall of 56%. Using a diversified external validation dataset, our model correctly labeled 15 normal fundus images (100%) and 15 OT fundus images (100%), with a mean confidence score of 0.965 and 0.963, respectively.
CONCLUSION: AutoML models created by ophthalmologists without coding experience were comparable or better than expert-designed bespoke models trained on the same dataset. By creatively using AutoML to identify OT lesions on fundus images, our approach brings the whole spectrum of DL model design into the hands of clinicians.
PMID:38411944 | DOI:10.1080/09273948.2024.2319281
Deep learning model based on multi-lesion and time series CT images for predicting the benefits from anti-HER2 targeted therapy in stage IV gastric cancer
Insights Imaging. 2024 Feb 27;15(1):59. doi: 10.1186/s13244-024-01639-2.
ABSTRACT
OBJECTIVE: To develop and validate a deep learning model based on multi-lesion and time series CT images in predicting overall survival (OS) in patients with stage IV gastric cancer (GC) receiving anti-HER2 targeted therapy.
METHODS: A total of 207 patients were enrolled in this multicenter study, with 137 patients for retrospective training and internal validation, 33 patients for prospective validation, and 37 patients for external validation. All patients received anti-HER2 targeted therapy and underwent pre- and post-treatment CT scans (baseline and at least one follow-up). The proposed deep learning model evaluated the multiple lesions in time series CT images to predict risk probabilities. We further evaluated and validated the risk score of the nomogram combining a two-follow-up lesion-based deep learning model (LDLM-2F), tumor markers, and clinical information for predicting the benefits from treatment (Nomo-LDLM-2F).
RESULTS: In the internal validation and prospective cohorts, the one-year AUCs for Nomo-LDLM-2F using the time series medical images and tumor markers were 0.894 (0.728-1.000) and 0.809 (0.561-1.000), respectively. In the external validation cohort, the one-year AUC of Nomo-LDLM-2F without tumor markers was 0.771 (0.510-1.000). Patients with a low Nomo-LDLM-2F score derived survival benefits from anti-HER2 targeted therapy significantly compared to those with a high Nomo-LDLM-2F score (all p < 0.05).
CONCLUSION: The Nomo-LDLM-2F score derived from multi-lesion and time series CT images holds promise for the effective readout of OS probability in patients with HER2-positive stage IV GC receiving anti-HER2 therapy.
CRITICAL RELEVANCE STATEMENT: The deep learning model using baseline and early follow-up CT images aims to predict OS in patients with stage IV gastric cancer receiving anti-HER2 targeted therapy. This model highlights the spatiotemporal heterogeneity of stage IV GC, assisting clinicians in the early evaluation of the efficacy of anti-HER2 therapy.
KEY POINTS: • Multi-lesion and time series model revealed the spatiotemporal heterogeneity in anti-HER2 therapy. • The Nomo-LDLM-2F score was a valuable prognostic marker for anti-HER2 therapy. • CT-based deep learning model incorporating time-series tumor markers improved performance.
PMID:38411839 | DOI:10.1186/s13244-024-01639-2
Prediction of lymph node metastasis in operable cervical cancer using clinical parameters and deep learning with MRI data: a multicentre study
Insights Imaging. 2024 Feb 27;15(1):56. doi: 10.1186/s13244-024-01618-7.
ABSTRACT
OBJECTIVES: To develop and validate a magnetic resonance imaging-based (MRI) deep multiple instance learning (D-MIL) model and combine it with clinical parameters for preoperative prediction of lymph node metastasis (LNM) in operable cervical cancer.
METHODS: A total of 392 patients with cervical cancer were retrospectively enrolled. Clinical parameters were analysed by logistical regression to construct a clinical model (M1). A ResNet50 structure is applied to extract features at the instance level without using manual annotations about the tumour region and then construct a D-MIL model (M2). A hybrid model (M3) was constructed by M1 and M2 scores. The diagnostic performance of each model was evaluated by the area under the receiver operating characteristic curve (AUC) and compared using the Delong method. Disease-free survival (DFS) was evaluated by the Kaplan‒Meier method.
RESULTS: SCC-Ag, maximum lymph node short diameter (LNmax), and tumour volume were found to be independent predictors of M1 model. For the diagnosis of LNM, the AUC of the training/internal/external cohort of M1 was 0.736/0.690/0.732, the AUC of the training/internal/external cohort of M2 was 0.757/0.714/0.765, and the AUC of the training/internal/external cohort of M3 was 0.838/0.764/0.835. M3 showed better performance than M1 and M2. Through the survival analysis, patients with higher hybrid model scores had a shorter time to reach DFS.
CONCLUSION: The proposed hybrid model could be used as a personalised non-invasive tool, which is helpful for predicting LNM in operable cervical cancer. The score of the hybrid model could also reflect the DFS of operable cervical cancer.
CRITICAL RELEVANCE STATEMENT: Lymph node metastasis is an important factor affecting the prognosis of cervical cancer. Preoperative prediction of lymph node status is helpful to make treatment decisions, improve prognosis, and prolong survival time.
KEY POINTS: • The MRI-based deep-learning model can predict the LNM in operable cervical cancer. • The hybrid model has the highest diagnostic efficiency for the LNM prediction. • The score of the hybrid model can reflect the DFS of operable cervical cancer.
PMID:38411729 | DOI:10.1186/s13244-024-01618-7
A novel deep learning-based perspective for tooth numbering and caries detection
Clin Oral Investig. 2024 Feb 27;28(3):178. doi: 10.1007/s00784-024-05566-w.
ABSTRACT
OBJECTIVES: The aim of this study was automatically detecting and numbering teeth in digital bitewing radiographs obtained from patients, and evaluating the diagnostic efficiency of decayed teeth in real time, using deep learning algorithms.
METHODS: The dataset consisted of 1170 anonymized digital bitewing radiographs randomly obtained from faculty archives. After image evaluation and labeling process, the dataset was split into training and test datasets. This study proposed an end-to-end pipeline architecture consisting of three stages for matching tooth numbers and caries lesions to enhance treatment outcomes and prevent potential issues. Initially, a pre-trained convolutional neural network (CNN) utilized to determine the side of the bitewing images. Then, an improved CNN model YOLOv7 was proposed for tooth numbering and caries detection. In the final stage, our developed algorithm assessed which teeth have caries by comparing the numbered teeth with the detected caries, using the intersection over union value for the matching process.
RESULTS: According to test results, the recall, precision, and F1-score values were 0.994, 0.987 and 0.99 for teeth detection, 0.974, 0.985 and 0.979 for teeth numbering, and 0.833, 0.866 and 0.822 for caries detection, respectively. For teeth numbering and caries detection matching performance; the accuracy, recall, specificity, precision and F1-Score values were 0.934, 0.834, 0.961, 0.851 and 0.842, respectively.
CONCLUSIONS: The proposed model exhibited good achievement, highlighting the potential use of CNNs for tooth detection, numbering, and caries detection, concurrently.
CLINICAL SIGNIFICANCE: CNNs can provide valuable support to clinicians by automating the detection and numbering of teeth, as well as the detection of caries on bitewing radiographs. By enhancing overall performance, these algorithms have the capacity to efficiently save time and play a significant role in the assessment process.
PMID:38411726 | DOI:10.1007/s00784-024-05566-w
Diagnostic performance and image quality of an image-based denoising algorithm applied to radiation dose-reduced CT in diagnosing acute appendicitis
Abdom Radiol (NY). 2024 Feb 27. doi: 10.1007/s00261-024-04246-3. Online ahead of print.
ABSTRACT
PURPOSE: To evaluate diagnostic performance and image quality of ultralow-dose CT (ULDCT) in diagnosing acute appendicitis with an image-based deep-learning denoising algorithm (IDLDA).
METHODS: This retrospective multicenter study included 180 patients (mean ± standard deviation, 29 ± 9 years; 91 female) who underwent contrast-enhanced 2-mSv CT for suspected appendicitis from February 2014 to August 2016. We simulated ULDCT from 2-mSv CT, reducing the dose by at least 50%. Then we applied an IDLDA on ULDCT to produce denoised ULDCT (D-ULDCT). Six radiologists with different experience levels (three board-certified radiologists and three residents) independently reviewed the ULDCT and D-ULDCT. They rated the likelihood of appendicitis and subjective image qualities (subjective image noise, diagnostic acceptability, and artificial sensation). One radiologist measured image noise, signal-to-noise ratio (SNR), and contrast-to-noise ratio (CNR). We used the receiver operating characteristic (ROC) analyses, Wilcoxon's signed-rank tests, and paired t-tests.
RESULTS: The area under the ROC curves (AUC) for diagnosing appendicitis ranged 0.90-0.97 for ULDCT and 0.94-0.97 for D-ULDCT. The AUCs of two residents were significantly higher on D-ULDCT (AUC difference = 0.06 [95% confidence interval, 0.01-0.11; p = .022] and 0.05 [0.00-0.10; p = .046], respectively). D-ULDCT provided better subjective image noise and diagnostic acceptability to all six readers. However, the response of board-certified radiologists and residents differed in artificial sensation (all p ≤ .003). D-ULDCT showed significantly lower image noise, higher SNR, and higher CNR (all p < .001).
CONCLUSION: An IDLDA can provide better ULDCT image quality and enhance diagnostic performance for less-experienced radiologists.
PMID:38411690 | DOI:10.1007/s00261-024-04246-3
Patterns of subregional cerebellar atrophy across epilepsy syndromes: An ENIGMA-Epilepsy study
Epilepsia. 2024 Feb 27. doi: 10.1111/epi.17881. Online ahead of print.
ABSTRACT
OBJECTIVE: The intricate neuroanatomical structure of the cerebellum is of longstanding interest in epilepsy, but has been poorly characterized within the current corticocentric models of this disease. We quantified cross-sectional regional cerebellar lobule volumes using structural magnetic resonance imaging in 1602 adults with epilepsy and 1022 healthy controls across 22 sites from the global ENIGMA-Epilepsy working group.
METHODS: A state-of-the-art deep learning-based approach was employed that parcellates the cerebellum into 28 neuroanatomical subregions. Linear mixed models compared total and regional cerebellar volume in (1) all epilepsies, (2) temporal lobe epilepsy with hippocampal sclerosis (TLE-HS), (3) nonlesional temporal lobe epilepsy, (4) genetic generalized epilepsy, and (5) extratemporal focal epilepsy (ETLE). Relationships were examined for cerebellar volume versus age at seizure onset, duration of epilepsy, phenytoin treatment, and cerebral cortical thickness.
RESULTS: Across all epilepsies, reduced total cerebellar volume was observed (d = .42). Maximum volume loss was observed in the corpus medullare (dmax = .49) and posterior lobe gray matter regions, including bilateral lobules VIIB (dmax = .47), crus I/II (dmax = .39), VIIIA (dmax = .45), and VIIIB (dmax = .40). Earlier age at seizure onset ( η ρ max 2 $$ \eta {\mathit{\mathsf{\rho}}}_{\mathsf{max}}^{\mathsf{2}} $$ = .05) and longer epilepsy duration ( η ρ max 2 $$ \eta {\mathit{\mathsf{\rho}}}_{\mathsf{max}}^{\mathsf{2}} $$ = .06) correlated with reduced volume in these regions. Findings were most pronounced in TLE-HS and ETLE, with distinct neuroanatomical profiles observed in the posterior lobe. Phenytoin treatment was associated with reduced posterior lobe volume. Cerebellum volume correlated with cerebral cortical thinning more strongly in the epilepsy cohort than in controls.
SIGNIFICANCE: We provide robust evidence of deep cerebellar and posterior lobe subregional gray matter volume loss in patients with chronic epilepsy. Volume loss was maximal for posterior subregions implicated in nonmotor functions, relative to motor regions of both the anterior and posterior lobe. Associations between cerebral and cerebellar changes, and variability of neuroanatomical profiles across epilepsy syndromes argue for more precise incorporation of cerebellar subregional damage into neurobiological models of epilepsy.
PMID:38411286 | DOI:10.1111/epi.17881
Deep learning-based skin care product recommendation: A focus on cosmetic ingredient analysis and facial skin conditions
J Cosmet Dermatol. 2024 Feb 27. doi: 10.1111/jocd.16218. Online ahead of print.
ABSTRACT
BACKGROUND: Recommendations for cosmetics are gaining popularity, but they are not being made with consideration of the analysis of cosmetic ingredients, which customers consider important when selecting cosmetics.
AIMS: This article aims to propose a method for estimating the efficacy of cosmetics based on their ingredients and introduces a system that recommends personalized products for consumers, combined with AI skin analysis.
METHODS: We constructed a deep neural network architecture to analyze sequentially arranged cosmetic ingredients in the product and incorporated skin analysis models to get the precise skin status of users from frontal face images. Our recommendation system makes decisions based on the results optimized for the individual.
RESULTS: Our cosmetic recommendation system has shown its effectiveness through reliable evaluation metrics, and numerous examples have demonstrated its ability to make reasonable recommendations for various skin problems.
CONCLUSION: The result shows that deep learning methods can be used to predict the effects of products based on their cosmetic ingredients and are available for use in personalized cosmetic recommendations.
PMID:38411029 | DOI:10.1111/jocd.16218
Deep learning applications for kidney histology analysis
Curr Opin Nephrol Hypertens. 2024 Feb 20. doi: 10.1097/MNH.0000000000000973. Online ahead of print.
ABSTRACT
PURPOSE OF REVIEW: Nephropathology is increasingly incorporating computational methods to enhance research and diagnostic accuracy. The widespread adoption of digital pathology, coupled with advancements in deep learning, will likely transform our pathology practices. Here, we discuss basic concepts of deep learning, recent applications in nephropathology, current challenges in implementation and future perspectives.
RECENT FINDINGS: Deep learning models have been developed and tested in various areas of nephropathology, for example, predicting kidney disease progression or diagnosing diseases based on imaging and clinical data. Despite their promising potential, challenges remain that hinder a wider adoption, for example, the lack of prospective evidence and testing in real-world scenarios.
SUMMARY: Deep learning offers great opportunities to improve quantitative and qualitative kidney histology analysis for research and clinical nephropathology diagnostics. Although exciting approaches already exist, the potential of deep learning in nephropathology is only at its beginning and we can expect much more to come.
PMID:38411024 | DOI:10.1097/MNH.0000000000000973
3D Convolutional Deep Learning for Nonlinear Estimation of Body Composition from Whole-Body Morphology
Res Sq [Preprint]. 2024 Feb 13:rs.3.rs-3935042. doi: 10.21203/rs.3.rs-3935042/v1.
ABSTRACT
Total and regional body composition are strongly correlated with metabolic syndrome and have been estimated non-invasively from 3D optical scans using linear parameterizations of body shape and linear regression models. Prior works produced accurate and precise predictions on many, but not all, body composition targets relative to the reference dual X-Ray absorptiometry (DXA) measurement. Here, we report the effects of replacing linear models with nonlinear parameterization and regression models on the precision and accuracy of body composition estimation in a novel application of deep 3D convolutional graph networks to human body composition modeling. We assembled an ensemble dataset of 4286 topologically standardized 3D optical scans from four different human body shape databases, DFAUST, CAESAR, Shape Up! Adults, and Shape Up! Kids and trained a parameterized shape model using a graph convolutional 3D autoencoder (3DAE) in lieu of linear PCA. We trained a nonlinear Gaussian process regression (GPR) on the 3DAE parameter space to predict body composition via correlations to paired DXA reference measurements from the Shape Up! scan subset. We tested our model on a set of 424 randomly withheld test meshes and compared the effects of nonlinear computation against prior linear models. Nonlinear GPR produced up to 20% reduction in prediction error and up to 30% increase in precision over linear regression for both sexes in 10 tested body composition variables. Deep shape features produced 6-8% reduction in prediction error over linear PCA features for males only and a 4-14% reduction in precision error for both sexes. Our best performing nonlinear model predicting body composition from deep features outperformed prior work using linear methods on all tested body composition prediction metrics in both precision and accuracy. All coefficients of determination (R 2 ) for all predicted variables were above 0.86. We show that GPR is a more precise and accurate method for modeling body composition mappings from body shape features than linear regression. Deep 3D features learned by a graph convolutional autoencoder only improved male body composition accuracy but improved precision in both sexes. Our work achieved lower estimation RMSEs than all previous work on 10 metrics of body composition.
PMID:38410459 | PMC:PMC10896405 | DOI:10.21203/rs.3.rs-3935042/v1
Use of Artificial Intelligence in the Diagnosis of Colorectal Cancer
Cureus. 2024 Jan 26;16(1):e53024. doi: 10.7759/cureus.53024. eCollection 2024 Jan.
ABSTRACT
Colorectal cancer (CRC) is one of the most common forms of cancer. Therefore, diagnosing the condition early and accurately is critical for improved patient outcomes and effective treatment. Recently, artificial intelligence (AI) algorithms such as support vector machine (SVM) and convolutional neural network (CNN) have demonstrated promise in medical image analysis. This paper, conducted from a systematic review perspective, aimed to determine the effectiveness of AI integration in CRC diagnosis, emphasizing accuracy, sensitivity, and specificity. From a methodological perspective, articles that were included were those that had been conducted in the past decade. Also, the articles needed to have been documented in English, with databases such as Embase, PubMed, and Google Scholar used to obtain relevant research studies. Similarly, keywords were used to arrive at relevant articles. These keywords included AI, CRC, specificity, sensitivity, accuracy, efficacy, effectiveness, disease diagnosis, screening, machine learning, area under the curve (AUC), and deep learning. From the results, most scholarly studies contend that AI is superior in medical image analysis, the development of subtle patterns, and decision support. However, while deploying these algorithms, a key theme is that the collaboration between medical experts and AI systems needs to be seamless. In addition, the AI algorithms ought to be refined continuously in the current world of big data and ensure that they undergo rigorous validation to provide more informed decision-making for or against adopting those AI tools in clinical settings. In conclusion, therefore, balancing between human expertise and technological innovation is likely to pave the way for the realization of AI's full potential concerning its promising role in improving CRC diagnosis, upon which there might be significant patient outcome improvements, disease detection, and the achievement of a more effective healthcare system.
PMID:38410294 | PMC:PMC10895204 | DOI:10.7759/cureus.53024
Deep learning denoising reconstruction for improved image quality in fetal cardiac cine MRI
Front Cardiovasc Med. 2024 Feb 12;11:1323443. doi: 10.3389/fcvm.2024.1323443. eCollection 2024.
ABSTRACT
PURPOSE: This study aims to evaluate deep learning (DL) denoising reconstructions for image quality improvement of Doppler ultrasound (DUS)-gated fetal cardiac MRI in congenital heart disease (CHD).
METHODS: Twenty-five fetuses with CHD (mean gestational age: 35 ± 1 weeks) underwent fetal cardiac MRI at 3T. Cine imaging was acquired using a balanced steady-state free precession (bSSFP) sequence with Doppler ultrasound gating. Images were reconstructed using both compressed sensing (bSSFP CS) and a pre-trained convolutional neural network trained for DL denoising (bSSFP DL). Images were compared qualitatively based on a 5-point Likert scale (from 1 = non-diagnostic to 5 = excellent) and quantitatively by calculating the apparent signal-to-noise ratio (aSNR) and contrast-to-noise ratio (aCNR). Diagnostic confidence was assessed for the atria, ventricles, foramen ovale, valves, great vessels, aortic arch, and pulmonary veins.
RESULTS: Fetal cardiac cine MRI was successful in 23 fetuses (92%), with two studies excluded due to extensive fetal motion. The image quality of bSSFP DL cine reconstructions was rated superior to standard bSSFP CS cine images in terms of contrast [3 (interquartile range: 2-4) vs. 5 (4-5), P < 0.001] and endocardial edge definition [3 (2-4) vs. 4 (4-5), P < 0.001], while the extent of artifacts was found to be comparable [4 (3-4.75) vs. 4 (3-4), P = 0.40]. bSSFP DL images had higher aSNR and aCNR compared with the bSSFP CS images (aSNR: 13.4 ± 6.9 vs. 8.3 ± 3.6, P < 0.001; aCNR: 26.6 ± 15.8 vs. 14.4 ± 6.8, P < 0.001). Diagnostic confidence of the bSSFP DL images was superior for the evaluation of cardiovascular structures (e.g., atria and ventricles: P = 0.003).
CONCLUSION: DL image denoising provides superior quality for DUS-gated fetal cardiac cine imaging of CHD compared to standard CS image reconstruction.
PMID:38410246 | PMC:PMC10894983 | DOI:10.3389/fcvm.2024.1323443
Research progress in predicting visceral pleural invasion of lung cancer: a narrative review
Transl Cancer Res. 2024 Jan 31;13(1):462-470. doi: 10.21037/tcr-23-1318. Epub 2023 Dec 26.
ABSTRACT
BACKGROUND AND OBJECTIVE: In lung cancer, visceral pleural invasion (VPI) affects the selection of surgical methods, the scope of lymph node dissection and the need for adjuvant chemotherapy. Preoperative or intraoperative prediction and diagnosis of VPI of lung cancer is helpful for choosing the best treatment plan and improving the prognosis of patients. This review aims to summarize the research progress of the clinical significance of VPI assessment, the intraoperative diagnosis technology of VPI, and various imaging methods for preoperative prediction of VPI. The diagnostic efficacy, advantages and disadvantages of various methods were summarized. The challenges and prospects for future research will also be discussed.
METHODS: A comprehensive, non-systematic review of the latest literature was carried out in order to investigate the progress of predicting VPI. PubMed database was being examined and the last run was on 4 August 2022.
KEY CONTENT AND FINDINGS: The pathological diagnosis and clinical significance of VPI of lung cancer were discussed in this review. The research progress of prediction and diagnosis of VPI in recent years was summarized. The results showed that preoperative imaging examination and intraoperative freezing pathology were of great value.
CONCLUSIONS: VPI is one of the adverse prognostic factors in patients with lung cancer. Accurate prediction of VPI status before surgery can provide guidance and help for the selection of clinical operation and postoperative treatment. There are some advantages and limitations in predicting VPI based on traditional computed tomography (CT) signs, 18F-fluorodeoxyglucose (18F-FDG) positron emission tomography (PET)/CT and magnetic resonance imaging (MRI) techniques. As an emerging technology, radiomics and deep learning show great potential and represent the future research direction.
PMID:38410233 | PMC:PMC10894335 | DOI:10.21037/tcr-23-1318