Deep learning
Investigating the effects of artificial intelligence on the personalization of breast cancer management: a systematic study
BMC Cancer. 2024 Jul 18;24(1):852. doi: 10.1186/s12885-024-12575-1.
ABSTRACT
BACKGROUND: Providing appropriate specialized treatment to the right patient at the right time is considered necessary in cancer management. Targeted therapy tailored to the genetic changes of each breast cancer patient is a desirable feature of precision oncology, which can not only reduce disease progression but also potentially increase patient survival. The use of artificial intelligence alongside precision oncology can help physicians by identifying and selecting more effective treatment factors for patients.
METHOD: A systematic review was conducted using the PubMed, Embase, Scopus, and Web of Science databases in September 2023. We performed the search strategy with keywords, namely: Breast Cancer, Artificial intelligence, and precision Oncology along with their synonyms in the article titles. Descriptive, qualitative, review, and non-English studies were excluded. The quality assessment of the articles and evaluation of bias were determined based on the SJR journal and JBI indices, as well as the PRISMA2020 guideline.
RESULTS: Forty-six studies were selected that focused on personalized breast cancer management using artificial intelligence models. Seventeen studies using various deep learning methods achieved a satisfactory outcome in predicting treatment response and prognosis, contributing to personalized breast cancer management. Two studies utilizing neural networks and clustering provided acceptable indicators for predicting patient survival and categorizing breast tumors. One study employed transfer learning to predict treatment response. Twenty-six studies utilizing machine-learning methods demonstrated that these techniques can improve breast cancer classification, screening, diagnosis, and prognosis. The most frequent modeling techniques used were NB, SVM, RF, XGBoost, and Reinforcement Learning. The average area under the curve (AUC) for the models was 0.91. Moreover, the average values for accuracy, sensitivity, specificity, and precision were reported to be in the range of 90-96% for the models.
CONCLUSION: Artificial intelligence has proven to be effective in assisting physicians and researchers in managing breast cancer treatment by uncovering hidden patterns in complex omics and genetic data. Intelligent processing of omics data through protein and gene pattern classification and the utilization of deep neural patterns has the potential to significantly transform the field of complex disease management.
PMID:39026174 | DOI:10.1186/s12885-024-12575-1
Synthetic temporal bone CT generation from UTE-MRI using a cycleGAN-based deep learning model: advancing beyond CT-MR imaging fusion
Eur Radiol. 2024 Jul 18. doi: 10.1007/s00330-024-10967-2. Online ahead of print.
ABSTRACT
OBJECTIVES: The aim of this study is to develop a deep-learning model to create synthetic temporal bone computed tomography (CT) images from ultrashort echo-time magnetic resonance imaging (MRI) scans, thereby addressing the intrinsic limitations of MRI in localizing anatomic landmarks in temporal bone CT.
MATERIALS AND METHODS: This retrospective study included patients who underwent temporal MRI and temporal bone CT within one month between April 2020 and March 2023. These patients were randomly divided into training and validation datasets. A CycleGAN model for generating synthetic temporal bone CT images was developed using temporal bone CT and pointwise encoding-time reduction with radial acquisition (PETRA). To assess the model's performance, the pixel count in mastoid air cells was measured. Two neuroradiologists evaluated the successful generation rates of 11 anatomical landmarks.
RESULTS: A total of 102 patients were included in this study (training dataset, n = 54, mean age 58 ± 14, 34 females (63%); validation dataset, n = 48, mean age 61 ± 13, 29 females (60%)). In the pixel count of mastoid air cells, no difference was observed between synthetic and real images (679 ± 342 vs 738 ± 342, p = 0.13). For the six major anatomical sites, the positive generation rates were 97-100%, whereas those of the five major anatomical structures ranged from 24% to 83%.
CONCLUSION: We developed a model to generate synthetic temporal bone CT images using PETRA MRI. This model can provide information regarding the major anatomic sites of the temporal bone using MRI.
CLINICAL RELEVANCE STATEMENT: The proposed algorithm addresses the primary limitations of MRI in localizing anatomic sites within the temporal bone.
KEY POINTS: CT is preferred for imaging the temporal bone, but has limitations in differentiating pathology there. The model achieved a high success rate in generating synthetic images of six anatomic sites. This can overcome the limitations of MRI in visualizing key anatomic sites in the temporal skull.
PMID:39026063 | DOI:10.1007/s00330-024-10967-2
Prediction of dose distributions for non-small cell lung cancer patients using MHA-ResUNet
Med Phys. 2024 Jul 18. doi: 10.1002/mp.17308. Online ahead of print.
ABSTRACT
BACKGROUND: The current level of automation in the production of radiotherapy plans for lung cancer patients is relatively low. With the development of artificial intelligence, it has become a reality to use neural networks to predict dose distributions and provide assistance for radiation therapy planning. However, due to the significant individual variability in the distribution of non-small cell lung cancer (NSCLC) planning target volume (PTV) and the complex spatial relationships between the PTV and organs at risk (OARs), there is still a lack of a high-precision dose prediction network tailored to the characteristics of NSCLC.
PURPOSE: To assist in the development of volumetric modulated arc therapy (VMAT) plans for non-small cell lung cancer patients, a deep neural network is proposed to predict high-precision dose distribution.
METHODS: This study has developed a network called MHA-ResUNet, which combines a large-kernel dilated convolution module and multi-head attention (MHA) modules. The network was trained based on 80 VMAT plans of NSCLC patients. CT images, PTV, and OARs were fed into the independent input channel. The dose distribution was taken as the output to train the model. The performance of this network was compared with that of several commonly used networks, and the networks' performance was evaluated based on the voxel-level mean absolute error (MAE) within the PTV and OARs, as well as the error in clinical dose-volume metrics.
RESULTS: The MAE between the predicted dose distribution and the manually planned dose distribution within the PTV is 1.43 Gy, and the D95 error is less than 1 Gy. Compared with the other three commonly used networks, the dose error of the MHA-ResUNet is the smallest in PTV and OARs.
CONCLUSIONS: The proposed MHA-ResUNet network improves the receptive field and filters the shallow features to learn the relative spatial relation between the PTV and the OARs, enabling accurate prediction of dose distributions in NSCLC patients undergoing VMAT radiotherapy.
PMID:39024495 | DOI:10.1002/mp.17308
Enhanced multistage deep learning for diagnosing anterior disc displacement in temporomandibular joint using magnetic resonance imaging
Dentomaxillofac Radiol. 2024 Jul 18:twae033. doi: 10.1093/dmfr/twae033. Online ahead of print.
ABSTRACT
OBJECTIVES: This study aimed to propose a new method for the automatic diagnosis of anterior disc displacement of the temporomandibular joint (TMJ) using magnetic resonance imaging (MRI) and deep learning. By employing a multistage approach, the factors affecting the final result can be easily identified and improved.
METHODS: This study introduces a multistage automatic diagnostic technique using deep learning. This process involves segmenting the target from MR images, extracting distance parameters, and classifying the diagnosis into three classes. MRI exams of 368 TMJs from 204 patients were evaluated for anterior disc displacement. In the first stage, five algorithms were used for the semantic segmentation of the disc and condyle. In the second stage, 54 distance parameters were extracted from the segments. In the third stage, a rule-based decision model was developed to link the parameters with the expert diagnosis results.
RESULTS: In the first stage, DeepLabV3+ showed the best result (95% Hausdorff distance, Dice coefficient, and sensitivity of 6.47 ± 7.22, 0.84 ± 0.07, and 0.84 ± 0.09, respectively). This study used the original MRI exams as input without preprocessing and showed high segmentation performance compared with that of previous studies. In the third stage, the combination of SegNet and a random forest model yielded an accuracy of 0.89 ± 0.06.
CONCLUSIONS: An algorithm was developed to automatically diagnose TMJ-anterior disc displacement using MRI. Through a multistage approach, this algorithm facilitated the improvement of results and demonstrated high accuracy from more complex inputs. Furthermore, existing radiological knowledge was applied and validated.
PMID:39024472 | DOI:10.1093/dmfr/twae033
Binding and sensing diverse small molecules using shape-complementary pseudocycles
Science. 2024 Jul 19;385(6706):276-282. doi: 10.1126/science.adn3780. Epub 2024 Jul 18.
ABSTRACT
We describe an approach for designing high-affinity small molecule-binding proteins poised for downstream sensing. We use deep learning-generated pseudocycles with repeating structural units surrounding central binding pockets with widely varying shapes that depend on the geometry and number of the repeat units. We dock small molecules of interest into the most shape complementary of these pseudocycles, design the interaction surfaces for high binding affinity, and experimentally screen to identify designs with the highest affinity. We obtain binders to four diverse molecules, including the polar and flexible methotrexate and thyroxine. Taking advantage of the modular repeat structure and central binding pockets, we construct chemically induced dimerization systems and low-noise nanopore sensors by splitting designs into domains that reassemble upon ligand addition.
PMID:39024436 | DOI:10.1126/science.adn3780
Achieving Occam's razor: Deep learning for optimal model reduction
PLoS Comput Biol. 2024 Jul 18;20(7):e1012283. doi: 10.1371/journal.pcbi.1012283. Online ahead of print.
ABSTRACT
All fields of science depend on mathematical models. Occam's razor refers to the principle that good models should exclude parameters beyond those minimally required to describe the systems they represent. This is because redundancy can lead to incorrect estimates of model parameters from data, and thus inaccurate or ambiguous conclusions. Here, we show how deep learning can be powerfully leveraged to apply Occam's razor to model parameters. Our method, FixFit, uses a feedforward deep neural network with a bottleneck layer to characterize and predict the behavior of a given model from its input parameters. FixFit has three major benefits. First, it provides a metric to quantify the original model's degree of complexity. Second, it allows for the unique fitting of data. Third, it provides an unbiased way to discriminate between experimental hypotheses that add value versus those that do not. In three use cases, we demonstrate the broad applicability of this method across scientific domains. To validate the method using a known system, we apply FixFit to recover known composite parameters for the Kepler orbit model and a dynamic model of blood glucose regulation. In the latter, we demonstrate the ability to fit the latent parameters to real data. To illustrate how the method can be applied to less well-established fields, we use it to identify parameters for a multi-scale brain model and reduce the search space for viable candidate mechanisms.
PMID:39024398 | DOI:10.1371/journal.pcbi.1012283
Advancements in urban scene segmentation using deep learning and generative adversarial networks for accurate satellite image analysis
PLoS One. 2024 Jul 18;19(7):e0307187. doi: 10.1371/journal.pone.0307187. eCollection 2024.
ABSTRACT
In the urban scene segmentation, the "image-to-image translation issue" refers to the fundamental task of transforming input images into meaningful segmentation maps, which essentially involves translating the visual information present in the input image into semantic labels for different classes. When this translation process is inaccurate or incomplete, it can lead to failed segmentation results where the model struggles to correctly classify pixels into the appropriate semantic categories. The study proposed a conditional Generative Adversarial Network (cGAN), for creating high-resolution urban maps from satellite images. The method combines semantic and spatial data using cGAN framework to produce realistic urban scenes while maintaining crucial details. To assess the performance of the proposed method, extensive experiments are performed on benchmark datasets, the ISPRS Potsdam and Vaihingen datasets. Intersection over Union (IoU) and Pixel Accuracy are two quantitative metrics used to evaluate the segmentation accuracy of the produced maps. The proposed method outperforms traditional methods with an IoU of 87% and a Pixel Accuracy of 93%. The experimental findings show that the suggested cGAN-based method performs better than traditional techniques, attaining better segmentation accuracy and generating better urban maps with finely detailed information. The suggested approach provides a framework for resolving the image-to-image translation difficulties in urban scene segmentation, demonstrating the potential of cGANs for producing excellent urban maps from satellite data.
PMID:39024353 | DOI:10.1371/journal.pone.0307187
PW-SMD: A Plane-Wave Implicit Solvation Model Based on Electron Density for Surface Chemistry and Crystalline Systems in Aqueous Solution
J Chem Theory Comput. 2024 Jul 18. doi: 10.1021/acs.jctc.4c00594. Online ahead of print.
ABSTRACT
Electron density-based implicit solvation models are a class of techniques for quantifying solvation effects and calculating free energies of solvation without an explicit representation of solvent molecules. Integral to the accuracy of solvation modeling is the proper definition of the solvation shell separating the solute molecule from the solvent environment, allowing for a physical partitioning of the free energies of solvation. Unlike state-of-the-art implicit solvation models for molecular quantum chemistry calculations, e.g., the solvation model based on solute electron density (SMD), solvation models for systems under periodic boundary conditions with plane-wave (PW) basis sets have been limited in their accuracy. Furthermore, a unified implicit solvation model with both homogeneous solution-phase and heterogeneous interfacial structures treated on equal footing is needed. In order to address this challenge, we developed a high-accuracy solvation model for periodic PW calculations that is applicable to molecular, ionic, interfacial, and bulk-phase chemistry. Our model, PW-SMD, is an extension of the SMD molecular solvation model to periodic systems in water. The free energy of solvation is partitioned into the electrostatic and cavity-dispersion-solvent structure (CDS) contributions. The electrostatic contributions of the solvation shell surrounding solute structures are parametrized based on their geometric and physical properties. In addition, the nonelectrostatic contribution to the solvation energy is accounted for by extending the CDS formalism of SMD to incorporate periodic boundary conditions. We validate the accuracy and robustness of our solvation model by comparing predicted solvation free energies against experimental data for molecular and ionic systems, carved-cluster composite energetic models of solvated reaction energies and barriers on surface systems, and deep-learning-accelerated ab initio molecular dynamics (AIMD). Our developed periodic implicit solvation model shows significantly improved accuracy compared to previous work (namely, solvation models in aqueous solution) and can be applied to simulate solvent effects in a wide range of surface and crystalline materials.
PMID:39024317 | DOI:10.1021/acs.jctc.4c00594
Biobjective gradient descent for feature selection on high dimension, low sample size data
PLoS One. 2024 Jul 18;19(7):e0305654. doi: 10.1371/journal.pone.0305654. eCollection 2024.
ABSTRACT
Even though deep learning shows impressive results in several applications, its use on problems with High Dimensions and Low Sample Size, such as diagnosing rare diseases, leads to overfitting. One solution often proposed is feature selection. In deep learning, along with feature selection, network sparsification is also used to improve the results when dealing with high dimensions low sample size data. However, most of the time, they are tackled as separate problems. This paper proposes a new approach that integrates feature selection, based on sparsification, into the training process of a deep neural network. This approach uses a constrained biobjective gradient descent method. It provides a set of Pareto optimal neural networks that make a trade-off between network sparsity and model accuracy. Results on both artificial and real datasets show that using a constrained biobjective gradient descent increases the network sparsity without degrading the classification performances. With the proposed approach, on an artificial dataset, the feature selection score reached 0.97 with a sparsity score of 0.92 with an accuracy of 0.9. For the same accuracy, none of the other methods reached a feature score above 0.20 and sparsity score of 0.35. Finally, statistical tests validate the results obtained on all datasets.
PMID:39024199 | DOI:10.1371/journal.pone.0305654
Machine/deep learning-assisted hemoglobin level prediction using palpebral conjunctival images
Br J Haematol. 2024 Jul 18. doi: 10.1111/bjh.19621. Online ahead of print.
ABSTRACT
Palpebral conjunctival hue alteration is used in non-invasive screening for anaemia, whereas it is a qualitative measure. This study constructed machine/deep learning models for predicting haemoglobin values using 150 palpebral conjunctival images taken by a smartphone. The median haemoglobin value was 13.1 g/dL, including 10 patients with <11 g/dL. A segmentation model using U-net was successfully constructed. The segmented images were subjected to non-convolutional neural network (CNN)-based and CNN-based regression models for predicting haemoglobin values. The correlation coefficients between the actual and predicted haemoglobin values were 0.38 and 0.44 in the non-CNN-based and CNN-based models, respectively. The sensitivity and specificity for anaemia detection were 13% and 98% for the non-CNN-based model and 20% and 99% for the CNN-based model. The performance of the CNN-based model did not improve with a mask layer guiding the model's attention towards the conjunctival regions, however, slightly improved with correction by the aspect ratio and exposure time of input images. The gradient-weighted class activation mapping heatmap indicated that the lower half area of the conjunctiva was crucial for haemoglobin value prediction. In conclusion, the CNN-based model had better results than the non-CNN-based model. The prediction accuracy would improve by using more input data with anaemia.
PMID:39024119 | DOI:10.1111/bjh.19621
Neurosymbolic AI for Reasoning Over Knowledge Graphs: A Survey
IEEE Trans Neural Netw Learn Syst. 2024 Jul 18;PP. doi: 10.1109/TNNLS.2024.3420218. Online ahead of print.
ABSTRACT
Neurosymbolic artificial intelligence (AI) is an increasingly active area of research that combines symbolic reasoning methods with deep learning to leverage their complementary benefits. As knowledge graphs (KGs) are becoming a popular way to represent heterogeneous and multirelational data, methods for reasoning on graph structures have attempted to follow this neurosymbolic paradigm. Traditionally, such approaches have utilized either rule-based inference or generated representative numerical embeddings from which patterns could be extracted. However, several recent studies have attempted to bridge this dichotomy to generate models that facilitate interpretability, maintain competitive performance, and integrate expert knowledge. Therefore, we survey methods that perform neurosymbolic reasoning tasks on KGs and propose a novel taxonomy by which we can classify them. Specifically, we propose three major categories: 1) logically informed embedding approaches; 2) embedding approaches with logical constraints; and 3) rule-learning approaches. Alongside the taxonomy, we provide a tabular overview of the approaches and links to their source code, if available, for more direct comparison. Finally, we discuss the unique characteristics and limitations of these methods and then propose several prospective directions toward which this field of research could evolve.
PMID:39024082 | DOI:10.1109/TNNLS.2024.3420218
Automatic Deep Learning Detection of Overhanging Restorations in Bitewing Radiographs
Dentomaxillofac Radiol. 2024 Jul 18:twae036. doi: 10.1093/dmfr/twae036. Online ahead of print.
ABSTRACT
OBJECTIVE: This study aimed to assess the effectiveness of deep convolutional neural network (CNN) algorithms in detecting and segmentation of overhanging dental restorations in bitewing radiographs.
METHOD AND MATERIALS: A total of 1160 anonymized bitewing radiographs were used to progress the Artificial Intelligence (AI) system) for the detection and segmentation of overhanging restorations. The data was then divided into three groups: 80% for training (930 images, 2399 labels), 10% for validation (115 images, 273 labels), and 10% for testing (115 images, 306 labels). A CNN model known as you only look once (YOLOv5) was trained to detect overhanging restorations in bitewing radiographs. After utilizing the remaining 115 radiographs to evaluate the efficacy of the proposed CNN model, the accuracy, sensitivity, precision, F1 score, and area under the receiver operating characteristic curve (AUC) were computed.
RESULTS: The model demonstrated a precision of 90.9%, a sensitivity of 85.3%, and an F1 score of 88.0%. Furthermore, the model achieved an AUC of 0.859 on the Receiver Operating Characteristic (ROC) curve. The mean average precision (mAP) at an intersection over a union (IoU) threshold of 0.5 was notably high at 0.87.
CONCLUSION: The findings suggest that deep CNN algorithms are highly effective in the detection and diagnosis of overhanging dental restorations in bitewing radiographs. The high levels of precision, sensitivity, and F1 score, along with the significant AUC and mAP values, underscore the potential of these advanced deep learning techniques in revolutionizing dental diagnostic procedures.
PMID:39024043 | DOI:10.1093/dmfr/twae036
A Multimorbidity Analysis of Hospitalized Patients With COVID-19 in Northwest Italy: Longitudinal Study Using Evolutionary Machine Learning and Health Administrative Data
JMIR Public Health Surveill. 2024 Jul 18;10:e52353. doi: 10.2196/52353.
ABSTRACT
BACKGROUND: Multimorbidity is a significant public health concern, characterized by the coexistence and interaction of multiple preexisting medical conditions. This complex condition has been associated with an increased risk of COVID-19. Individuals with multimorbidity who contract COVID-19 often face a significant reduction in life expectancy. The postpandemic period has also highlighted an increase in frailty, emphasizing the importance of integrating existing multimorbidity details into epidemiological risk assessments. Managing clinical data that include medical histories presents significant challenges, particularly due to the sparsity of data arising from the rarity of multimorbidity conditions. Also, the complex enumeration of combinatorial multimorbidity features introduces challenges associated with combinatorial explosions.
OBJECTIVE: This study aims to assess the severity of COVID-19 in individuals with multiple medical conditions, considering their demographic characteristics such as age and sex. We propose an evolutionary machine learning model designed to handle sparsity, analyzing preexisting multimorbidity profiles of patients hospitalized with COVID-19 based on their medical history. Our objective is to identify the optimal set of multimorbidity feature combinations strongly associated with COVID-19 severity. We also apply the Apriori algorithm to these evolutionarily derived predictive feature combinations to identify those with high support.
METHODS: We used data from 3 administrative sources in Piedmont, Italy, involving 12,793 individuals aged 45-74 years who tested positive for COVID-19 between February and May 2020. From their 5-year pre-COVID-19 medical histories, we extracted multimorbidity features, including drug prescriptions, disease diagnoses, sex, and age. Focusing on COVID-19 hospitalization, we segmented the data into 4 cohorts based on age and sex. Addressing data imbalance through random resampling, we compared various machine learning algorithms to identify the optimal classification model for our evolutionary approach. Using 5-fold cross-validation, we evaluated each model's performance. Our evolutionary algorithm, utilizing a deep learning classifier, generated prediction-based fitness scores to pinpoint multimorbidity combinations associated with COVID-19 hospitalization risk. Eventually, the Apriori algorithm was applied to identify frequent combinations with high support.
RESULTS: We identified multimorbidity predictors associated with COVID-19 hospitalization, indicating more severe COVID-19 outcomes. Frequently occurring morbidity features in the final evolved combinations were age>53, R03BA (glucocorticoid inhalants), and N03AX (other antiepileptics) in cohort 1; A10BA (biguanide or metformin) and N02BE (anilides) in cohort 2; N02AX (other opioids) and M04AA (preparations inhibiting uric acid production) in cohort 3; and G04CA (Alpha-adrenoreceptor antagonists) in cohort 4.
CONCLUSIONS: When combined with other multimorbidity features, even less prevalent medical conditions show associations with the outcome. This study provides insights beyond COVID-19, demonstrating how repurposed administrative data can be adapted and contribute to enhanced risk assessment for vulnerable populations.
PMID:39024001 | DOI:10.2196/52353
Deep learning reconstruction for coronary CT angiography in patients with origin anomaly, stent or bypass graft
Radiol Med. 2024 Jul 18. doi: 10.1007/s11547-024-01846-3. Online ahead of print.
ABSTRACT
PURPOSE: To develop and validate a deep learning (DL)-model for automatic reconstruction for coronary CT angiography (CCTA) in patients with origin anomaly, stent or bypass graft.
MATERIAL AND METHODS: In this retrospective study, a DL model for automatic CCTA reconstruction was developed with training and validation sets from 6063 and 1962 patients. The algorithm was evaluated on an independent external test set of 812 patients (357 with origin anomaly or revascularization, 455 without). The image quality of DL reconstruction and manual reconstruction (using dedicated cardiac reconstruction software provided by CT vendors) was compared using a 5-point scale. The successful reconstruction rates and post-processing time for two methods were recorded.
RESULTS: In the external test set, 812 patients (mean age, 64.0 ± 11.6, 100 with origin anomalies, 152 with stents, 105 with bypass grafts) were evaluated. The successful rates for automatic reconstruction were 100% (455/455), 97% (97/100), 100% (152/152), and 76.2% (80/105) in patients with native vessel, origin anomaly, stent, and bypass graft, respectively. The image quality scores were significantly higher for DL reconstruction than those for manual approach in all subgroups (4 vs. 3 for native vessel, 4 vs. 4 for origin anomaly, 4 vs. 3 for stent and 4 vs. 3 for bypass graft, all p < 0.001). The overall post-processing time was remarkably reduced for DL reconstruction compared to manual method (11 s vs. 465 s, p < 0.001).
CONCLUSIONS: The developed DL model enabled accurate automatic CCTA reconstruction of bypass graft, stent and origin anomaly. It significantly reduced post-processing time and improved clinical workflow.
PMID:39023665 | DOI:10.1007/s11547-024-01846-3
Evaluation of a deep image-to-image network (DI2IN) auto-segmentation algorithm across a network of cancer centers
J Cancer Res Ther. 2024 Apr 1;20(3):1020-1025. doi: 10.4103/jcrt.jcrt_769_23. Epub 2024 Jan 22.
ABSTRACT
PURPOSE/OBJECTIVE S: Due to manual OAR contouring challenges, various automatic contouring solutions have been introduced. Historically, common clinical auto-segmentation algorithms used were atlas-based, which required maintaining a library of self-made contours. Searching the collection was computationally intensive and could take several minutes to complete. Deep learning approaches have shown significant benefits compared to atlas-based methods in improving segmentation accuracy and efficiency in auto-segmentation algorithms. This work represents the first multi-institutional study to describe and evaluate an AI algorithm for the auto-segmentation of organs at risk (OARs) based on a deep image-to-image network (DI2IN).
MATERIALS/METHODS: The AI-Rad Companion Organs RT (AIRC) algorithm (Siemens Healthineers, Erlangen, Germany) uses a two-step approach for segmentation. In the first step, the target organ region in the optimal input image is extracted using a trained deep reinforcement learning network (DRL), which is then used as input to create the contours in the second step based on DI2IN. The study was initially designed as a prospective single-center evaluation. The automated contours generated by AIRC were evaluated by three experienced board-certified radiation oncologists using a four-point scale where 4 is clinically usable and 1 requires re-contouring. After seeing favorable results in a single-center pilot study, we decided to expand the study to six additional institutions, encompassing eight additional evaluators for a total of 11 physician evaluators across seven institutions.
RESULTS: One hundred and fifty-six patients and 1366 contours were prospectively evaluated. The five most commonly contoured organs were the lung (136 contours, average rating = 4.0), spinal cord (106 contours, average rating = 3.1), eye globe (80 contours, average rating = 3.9), lens (77 contours, average rating = 3.9), and optic nerve (75 contours, average rating = 4.0). The average rating per evaluator per contour was 3.6. On average, 124 contours were evaluated by each evaluator. 65% of the contours were rated as 4, and 31% were rated as 3. Only 4% of contours were rated as 1 or 2. Thirty-three organs were evaluated in the study, with 19 structures having a 3.5 or above average rating (ribs, abdominopelvic cavity, skeleton, larynx, lung, aorta, brachial plexus, lens, eye globe, glottis, heart, parotid glands, bladder, kidneys, supraglottic larynx, submandibular glands, esophagus, optic nerve, oral cavity) and the remaining organs having a rating of 3.0 or greater (female breast, proximal femur, seminal vesicles, rectum, sternum, brainstem, prostate, brain, lips, mandible, liver, optic chiasm, spinal cord, spleen). No organ had an average rating below 3.
CONCLUSION: AIRC performed well with greater than 95% of contours accepted by treating physicians with no or minor edits. It supported a fully automated workflow with the potential for time savings and increased standardization with the use of AI-powered algorithms for high-quality OAR contouring.
PMID:39023610 | DOI:10.4103/jcrt.jcrt_769_23
AI-Based Strain Estimation in Echocardiography Using Open and Collaborative Data: The More Experts the Better?
JACC Cardiovasc Imaging. 2024 Jul 3:S1936-878X(24)00232-8. doi: 10.1016/j.jcmg.2024.05.020. Online ahead of print.
NO ABSTRACT
PMID:39023498 | DOI:10.1016/j.jcmg.2024.05.020
Multi-Plexus Nonperfusion Area Segmentation in Widefield OCT Angiography Using a Deep Convolutional Neural Network
Transl Vis Sci Technol. 2024 Jul 1;13(7):15. doi: 10.1167/tvst.13.7.15.
ABSTRACT
PURPOSE: To train and validate a convolutional neural network to segment nonperfusion areas (NPAs) in multiple retinal vascular plexuses on widefield optical coherence tomography angiography (OCTA).
METHODS: This cross-sectional study included 202 participants with a full range of diabetic retinopathy (DR) severities (diabetes mellitus without retinopathy, mild to moderate non-proliferative DR, severe non-proliferative DR, and proliferative DR) and 39 healthy participants. Consecutive 6 × 6-mm OCTA scans at the central macula, optic disc, and temporal region in one eye from 202 participants in a clinical DR study were acquired with a 70-kHz OCT commercial system (RTVue-XR). Widefield OCTA en face images were generated by montaging the scans from these three regions. A projection-resolved OCTA algorithm was applied to remove projection artifacts at the voxel scale. A deep convolutional neural network with a parallel U-Net module was designed to detect NPAs and distinguish signal reduction artifacts from flow deficits in the superficial vascular complex (SVC), intermediate capillary plexus (ICP), and deep capillary plexus (DCP). Expert graders manually labeled NPAs and signal reduction artifacts for the ground truth. Sixfold cross-validation was used to evaluate the proposed algorithm on the entire dataset.
RESULTS: The proposed algorithm showed high agreement with the manually delineated ground truth for NPA detection in three retinal vascular plexuses on widefield OCTA (mean ± SD F-score: SVC, 0.84 ± 0.05; ICP, 0.87 ± 0.04; DCP, 0.83 ± 0.07). The extrafoveal avascular area in the DCP showed the best sensitivity for differentiating eyes with diabetes but no retinopathy (77%) from healthy controls and for differentiating DR by severity: DR versus no DR, 77%; referable DR (rDR) versus non-referable DR (nrDR), 79%; vision-threatening DR (vtDR) versus non-vision-threatening DR (nvtDR), 60%. The DCP also showed the best area under the receiver operating characteristic curve for distinguishing diabetes from healthy controls (96%), DR versus no DR (95%), and rDR versus nrDR (96%). The three-plexus-combined OCTA achieved the best result in differentiating vtDR and nvtDR (81.0%).
CONCLUSIONS: A deep learning network can accurately segment NPAs in individual retinal vascular plexuses and improve DR diagnostic accuracy.
TRANSLATIONAL RELEVANCE: Using a deep learning method to segment nonperfusion areas in widefield OCTA can potentially improve the diagnostic accuracy of diabetic retinopathy by OCT/OCTA systems.
PMID:39023443 | DOI:10.1167/tvst.13.7.15
Pre-processing visual scenes for retinal prosthesis systems: A comprehensive review
Artif Organs. 2024 Jul 18. doi: 10.1111/aor.14824. Online ahead of print.
ABSTRACT
BACKGROUND: Retinal prostheses offer hope for individuals with degenerative retinal diseases by stimulating the remaining retinal cells to partially restore their vision. This review delves into the current advancements in retinal prosthesis technology, with a special emphasis on the pivotal role that image processing and machine learning techniques play in this evolution.
METHODS: We provide a comprehensive analysis of the existing implantable devices and optogenetic strategies, delineating their advantages, limitations, and challenges in addressing complex visual tasks. The review extends to various image processing algorithms and deep learning architectures that have been implemented to enhance the functionality of retinal prosthetic devices. We also illustrate the testing results by demonstrating the clinical trials or using Simulated Prosthetic Vision (SPV) through phosphene simulations, which is a critical aspect of simulating visual perception for retinal prosthesis users.
RESULTS: Our review highlights the significant progress in retinal prosthesis technology, particularly its capacity to augment visual perception among the visually impaired. It discusses the integration between image processing and deep learning, illustrating their impact on individual interactions and navigations within the environment through applying clinical trials and also illustrating the limitations of some techniques to be used with current devices, as some approaches only use simulation even on sighted-normal individuals or rely on qualitative analysis, where some consider realistic perception models and others do not.
CONCLUSION: This interdisciplinary field holds promise for the future of retinal prostheses, with the potential to significantly enhance the quality of life for individuals with retinal prostheses. Future research directions should pivot towards optimizing phosphene simulations for SPV approaches, considering the distorted and confusing nature of phosphene perception, thereby enriching the visual perception provided by these prosthetic devices. This endeavor will not only improve navigational independence but also facilitate a more immersive interaction with the environment.
PMID:39023279 | DOI:10.1111/aor.14824
Innovative approaches for coronary heart disease management: integrating biomedical sensors, deep learning, and stellate ganglion modulation
Comput Methods Biomech Biomed Engin. 2024 Jul 18:1-18. doi: 10.1080/10255842.2024.2378099. Online ahead of print.
ABSTRACT
Coronary heart disease (CHD) is a significant global health concern, necessitating continuous advancements in treatment modalities to improve patient outcomes. Traditional Chinese medicine (TCM) offers alternative therapeutic approaches, but integration with modern biomedical technologies remains relatively unexplored. This study aimed to assess the efficacy of a combined treatment approach for CHD, integrating traditional Chinese medicinal interventions with modern biomedical sensors and stellate ganglion modulation. The objective was to evaluate the impact of this combined treatment on symptom relief, clinical outcomes, hemorheological indicators, and inflammatory biomarkers. A randomized controlled trial was conducted on 117 CHD patients with phlegm-turbidity congestion and excessiveness type. Patients were divided into a combined treatment group (CTG) and a traditional Chinese medicinal group (CMG). The CTG group received a combination of herbal decoctions, thread-embedding therapy, and stellate ganglion modulation, while the CMG group only received traditional herbal decoctions. The CTG demonstrated superior outcomes compared to the CMG across multiple parameters. Significant reductions in TCM symptom scores, improved clinical effects, reduced angina manifestation, favorable changes in hemorheological indicators, and decreased serum inflammatory biomarkers were observed in the CTG post-intervention. The combination of traditional Chinese medicinal interventions with modern biomedical sensors and stellate ganglion modulation has shown promising results in improving symptoms, clinical outcomes, and inflammatory markers in CHD patients. This holistic approach enhances treatment efficacy and patient outcomes. Further research and advancements in sensor technology are needed to optimize this approach.
PMID:39023137 | DOI:10.1080/10255842.2024.2378099
Causality-inspired crop pest recognition based on Decoupled Feature Learning
Pest Manag Sci. 2024 Jul 18. doi: 10.1002/ps.8314. Online ahead of print.
ABSTRACT
BACKGROUND: Ensuring the efficient recognition and management of crop pests is crucial for maintaining the balance in global agricultural ecosystems and ecological harmony. Deep learning-based methods have shown promise in crop pest recognition. However, prevailing methods often fail to address a critical issue: biased pest training dataset distribution stemming from the tendency to collect images primarily in certain environmental contexts, such as paddy fields. This oversight hampers recognition accuracy when encountering pest images dissimilar to training samples, highlighting the need for a novel approach to overcome this limitation.
RESULTS: We introduce the Decoupled Feature Learning (DFL) framework, leveraging causal inference techniques to handle training dataset bias. DFL manipulates the training data based on classification confidence to construct different training domains and employs center triplet loss for learning class-core features. The proposed DFL framework significantly boosts existing baseline models, attaining unprecedented recognition accuracies of 95.33%, 92.59%, and 74.86% on the Li, DFSPD, and IP102 datasets, respectively.
CONCLUSION: Extensive testing on three pest datasets using standard baseline models demonstrates the superiority of DFL in pest recognition. The visualization results show that DFL encourages the baseline models to capture the class-core features. The proposed DFL marks a pivotal step in mitigating the issue of data distribution bias, enhancing the reliability of deep learning in agriculture. © 2024 Society of Chemical Industry.
PMID:39022822 | DOI:10.1002/ps.8314