Deep learning
Diffusion network with spatial channel attention infusion and frequency spatial attention for brain tumor segmentation
Med Phys. 2024 Oct 30. doi: 10.1002/mp.17482. Online ahead of print.
ABSTRACT
BACKGROUND: Accurate segmentation of gliomas is crucial for diagnosis, treatment planning, and evaluating therapeutic efficacy. Physicians typically analyze and delineate tumor regions in brain magnetic resonance imaging (MRI) images based on personal experience, which is often time-consuming and subject to individual interpretation. Despite advancements in deep learning technology for image segmentation, current techniques still face challenges in clearly defining tumor boundary contours and enhancing segmentation accuracy.
PURPOSE: To address these issues, this paper proposes a conditional diffusion network (SF-Diff) with a spatial channel attention infusion (SCAI) module and a frequency spatial attention (FSA) mechanism to achieve accurate segmentation of the whole tumor (WT) region in brain tumors.
METHODS: SF-Diff initially extracts multiscale information from multimodal MRI images and subsequently employs a diffusion model to restore boundaries and details, thereby enabling accurate brain tumor segmentation (BraTS). Specifically, a SCAI module is developed to capture multiscale information within and between encoder layers. A dual-channel upsampling block (DUB) is designed to assist in detail recovery during upsampling. A FSA mechanism is introduced to better match the conditional features with the diffusion probability distribution information. Furthermore, a cross-model loss function has been implemented to supervise the feature extraction of the conditional model and the noise distribution of the diffusion model.
RESULTS: The dataset used in this paper is publicly available and includes 369 patient cases from the Multimodal BraTS Challenge 2020 (BraTS2020). The conducted experiments on BraTS2020 demonstrate that SF-Diff performs better than other state-of-the-art models. The method achieved a Dice score of 91.87%, a Hausdorff 95 of 5.47 mm, an IoU of 84.96%, a sensitivity of 92.29%, and a specificity of 99.95% on BraTS2020.
CONCLUSIONS: The proposed SF-Diff performs well in identifying the WT region of the brain tumors compared to other state-of-the-art models, especially in terms of boundary contours and non-contiguous lesion regions, which is clinically significant. In the future, we will further develop this method for brain tumor three-class segmentation task.
PMID:39476317 | DOI:10.1002/mp.17482
Assessing small molecule conformational sampling methods in molecular docking
J Comput Chem. 2024 Oct 30. doi: 10.1002/jcc.27516. Online ahead of print.
ABSTRACT
Small molecule conformational sampling plays a pivotal role in molecular docking. Recent advancements have led to the emergence of various conformational sampling methods, each employing distinct algorithms. This study investigates the impact of different small molecule conformational sampling methods in molecular docking using UCSF DOCK 3.7. Specifically, six traditional sampling methods (Omega, BCL::Conf, CCDC Conformer Generator, ConfGenX, Conformator, RDKit ETKDGv3) and a deep learning-based model (Torsional Diffusion) for generating conformational ensembles are evaluated. These ensembles are subsequently docked against the Platinum Diverse Dataset, the PoseBusters dataset and the DUDE-Z dataset to assess binding pose reproducibility and screening power. Notably, different sampling methods exhibit varying performance due to their unique preferences, such as dihedral angle sampling ranges on rotatable bonds. Combining complementary methods may lead to further improvements in docking performance.
PMID:39476310 | DOI:10.1002/jcc.27516
Automated Posterior Scleral Topography Assessment for Enhanced Staphyloma Visualization and Quantification With Improved Maculopathy Correlation
Transl Vis Sci Technol. 2024 Oct 1;13(10):41. doi: 10.1167/tvst.13.10.41.
ABSTRACT
PURPOSE: To quantitatively characterize the posterior morphology of high myopia eyes with posterior staphyloma.
METHODS: Surface points of the eyeball were automatically extracted from magnetic resonance imaging scans using deep learning. Subsequently, the topography of posterior staphylomas was constructed to facilitate accurate visualization and quantification of their location and severity. In the three-dimensional Cartesian coordinate system established with surface points, measurements of distances (D) from each point to the hypothetical pre-elongation eye center within the eyeball and local curvatures (C) at each point on the posterior sclera were computed. Using this data, specific parameters were formulated. The concordance of these parameters with traditional staphyloma classification methods and their association with myopic traction maculopathy (MTM) grades based on the ATN classifications were investigated.
RESULTS: The study included 102 eyes from 52 participants. The measured parameters, particularly the variance of distance (Dvar) and the maximum value of the curvature and distance product (C · Dmax), demonstrated efficacy in differentiating various types of posterior staphyloma and exhibited strong correlations with the grades of MTM.
CONCLUSIONS: The automated generation of the posterior scleral topography facilitated visualization and quantification of staphyloma location and severity. Simple geometric parameters can quantify staphyloma shape and correlate well with retinal complications. Future works on expanding these measures to more imaging modalities could improve their clinical use and deepen insights into the link between posterior staphyloma and related retinal diseases.
TRANSLATIONAL RELEVANCE: This work has the potential to be translated into clinical practice, allowing for the accurate assessment of staphyloma severity and ultimately improving disease management.
PMID:39476086 | DOI:10.1167/tvst.13.10.41
MosquitoSong+: A noise-robust deep learning model for mosquito classification from wingbeat sounds
PLoS One. 2024 Oct 30;19(10):e0310121. doi: 10.1371/journal.pone.0310121. eCollection 2024.
ABSTRACT
In order to assess risk of mosquito-vector borne disease and to effectively target and monitor vector control efforts, accurate information about mosquito vector population densities is needed. The traditional and still most common approach to this involves the use of traps along with manual counting and classification of mosquito species, but the costly and labor-intensive nature of this approach limits its widespread use. Numerous previous studies have sought to address this problem by developing machine learning models to automatically identify species and sex of mosquitoes based on their wingbeat sounds. Yet little work has addressed the issue of robust classification in the presence of environmental background noise, which is essential to making the approach practical. In this paper, we propose a new deep learning model, MosquitoSong+, to identify the species and sex of mosquitoes from raw wingbeat sounds so that it is robust to the environmental noise and the relative volume of the mosquito's flight tone. The proposed model extends the existing 1D-CNN model by adjusting its architecture and introducing two data augmentation techniques during model training: noise augmentation and wingbeat volume variation. Experiments show that the new model has very good generalizability, with species classification accuracy above 80% on several wingbeat datasets with various background noise. It also has an accuracy of 93.3% for species and sex classification on wingbeat sounds overlaid with various background noises. These results suggest that the proposed approach may be a practical means to develop classification models that can perform well in the field.
PMID:39475971 | DOI:10.1371/journal.pone.0310121
Prediction of fellow eye neovascularization in type 3 macular neovascularization (Retinal angiomatous proliferation) using deep learning
PLoS One. 2024 Oct 30;19(10):e0310097. doi: 10.1371/journal.pone.0310097. eCollection 2024.
ABSTRACT
PURPOSE: To establish a deep learning artificial intelligence model to predict the risk of long-term fellow eye neovascularization in unilateral type 3 macular neovascularization (MNV).
METHODS: This retrospective study included 217 patients (199 in the training/validation of the AI model and 18 in the testing set) with a diagnosis of unilateral type 3 MNV. The purpose of the AI model was to predict fellow eye neovascularization within 24 months after the initial diagnosis. The data used to train the AI model included a baseline fundus image and horizontal/vertical cross-hair scan optical coherence tomography images in the fellow eye. The neural network of this study for AI-learning was based on the visual geometry group with modification. The precision, recall, accuracy, and the area under the curve values of receiver operating characteristics (AUCROC) were calculated for the AI model. The accuracy of an experienced (examiner 1) and less experienced (examiner 2) human examiner was also evaluated.
RESULTS: The incidence of fellow eye neovascularization over 24 months was 28.6% in the training/validation set and 38.9% in the testing set (P = 0.361). In the AI model, precision was 0.562, recall was 0.714, accuracy was 0.667, and the AUCROC was 0.675. The sensitivity, specificity, and accuracy were 0.429, 0.727, and 0.611, respectively, for examiner 1, and 0.143, 0.636, and 0.444, respectively, for examiner 2.
CONCLUSIONS: This is the first AI study focusing on the clinical course of type 3 MNV. While our AI model exhibited accuracy comparable to that of human examiners, overall accuracy was not high. This may partly be a result of the relatively small number of patients used for AI training, suggesting the need for future multi-center studies to improve the accuracy of the model.
PMID:39475903 | DOI:10.1371/journal.pone.0310097
DKVMN&MRI: A new deep knowledge tracing model based on DKVMN incorporating multi-relational information
PLoS One. 2024 Oct 30;19(10):e0312022. doi: 10.1371/journal.pone.0312022. eCollection 2024.
ABSTRACT
Knowledge tracing is a technology that models students' changing knowledge state over learning time based on their historical answer records, thus predicting their learning ability. It is the core module that supports the intelligent education system. To address the problems of sparse input data, lack of interpretability and weak capacity to capture the relationship between exercises in the existing models, this paper build a deep knowledge tracing model DKVMN&MRI based on the Dynamic Key-Value Memory Network (DKVMN) that incorporates multiple relationship information including exercise-knowledge point relations, exercise-exercise relations, and learning-forgetting relations. In the model, firstly, the Q-matrix is utilized to map the link between knowledge points and exercises to the input layer; secondly, improved DKVMN and LSTM are used to model the learning process of learners, then the Ebbinghaus forgetting curve function is introduced to simulate the process of memory forgetting in learners, and finally, the prediction strategies of Item Response Theory (IRT) and attention mechanism are used to combine the similarity relationship between learners' knowledge state and exercises to calculate the probability that learners would correctly respond during the subsequent time step. Through extensive experiments on three real-world datasets, we demonstrate that DKVMN&MRI has significant improvements in both AUC and ACC metrics contrast with the latest models. Furthermore, the study provides explanations at both the exercise level and learner knowledge state level, demonstrating the interpretability and efficacy of the proposed model.
PMID:39475856 | DOI:10.1371/journal.pone.0312022
Targeting COVID-19 and Human Resources for Health News Information Extraction: Algorithm Development and Validation
JMIR AI. 2024 Oct 30;3:e55059. doi: 10.2196/55059.
ABSTRACT
BACKGROUND: Global pandemics like COVID-19 put a high amount of strain on health care systems and health workers worldwide. These crises generate a vast amount of news information published online across the globe. This extensive corpus of articles has the potential to provide valuable insights into the nature of ongoing events and guide interventions and policies. However, the sheer volume of information is beyond the capacity of human experts to process and analyze effectively.
OBJECTIVE: The aim of this study was to explore how natural language processing (NLP) can be leveraged to build a system that allows for quick analysis of a high volume of news articles. Along with this, the objective was to create a workflow comprising human-computer symbiosis to derive valuable insights to support health workforce strategic policy dialogue, advocacy, and decision-making.
METHODS: We conducted a review of open-source news coverage from January 2020 to June 2022 on COVID-19 and its impacts on the health workforce from the World Health Organization (WHO) Epidemic Intelligence from Open Sources (EIOS) by synergizing NLP models, including classification and extractive summarization, and human-generated analyses. Our DeepCovid system was trained on 2.8 million news articles in English from more than 3000 internet sources across hundreds of jurisdictions.
RESULTS: Rules-based classification with hand-designed rules narrowed the data set to 8508 articles with high relevancy confirmed in the human-led evaluation. DeepCovid's automated information targeting component reached a very strong binary classification performance of 98.98 for the area under the receiver operating characteristic curve (ROC-AUC) and 47.21 for the area under the precision recall curve (PR-AUC). Its information extraction component attained good performance in automatic extractive summarization with a mean Recall-Oriented Understudy for Gisting Evaluation (ROUGE) score of 47.76. DeepCovid's final summaries were used by human experts to write reports on the COVID-19 pandemic.
CONCLUSIONS: It is feasible to synergize high-performing NLP models and human-generated analyses to benefit open-source health workforce intelligence. The DeepCovid approach can contribute to an agile and timely global view, providing complementary information to scientific literature.
PMID:39475833 | DOI:10.2196/55059
Leveraging Artificial Intelligence and Data Science for Integration of Social Determinants of Health in Emergency Medicine: Scoping Review
JMIR Med Inform. 2024 Oct 30;12:e57124. doi: 10.2196/57124.
ABSTRACT
BACKGROUND: Social determinants of health (SDOH) are critical drivers of health disparities and patient outcomes. However, accessing and collecting patient-level SDOH data can be operationally challenging in the emergency department (ED) clinical setting, requiring innovative approaches.
OBJECTIVE: This scoping review examines the potential of AI and data science for modeling, extraction, and incorporation of SDOH data specifically within EDs, further identifying areas for advancement and investigation.
METHODS: We conducted a standardized search for studies published between 2015 and 2022, across Medline (Ovid), Embase (Ovid), CINAHL, Web of Science, and ERIC databases. We focused on identifying studies using AI or data science related to SDOH within emergency care contexts or conditions. Two specialized reviewers in emergency medicine (EM) and clinical informatics independently assessed each article, resolving discrepancies through iterative reviews and discussion. We then extracted data covering study details, methodologies, patient demographics, care settings, and principal outcomes.
RESULTS: Of the 1047 studies screened, 26 met the inclusion criteria. Notably, 9 out of 26 (35%) studies were solely concentrated on ED patients. Conditions studied spanned broad EM complaints and included sepsis, acute myocardial infarction, and asthma. The majority of studies (n=16) explored multiple SDOH domains, with homelessness/housing insecurity and neighborhood/built environment predominating. Machine learning (ML) techniques were used in 23 of 26 studies, with natural language processing (NLP) being the most commonly used approach (n=11). Rule-based NLP (n=5), deep learning (n=2), and pattern matching (n=4) were the most commonly used NLP techniques. NLP models in the reviewed studies displayed significant predictive performance with outcomes, with F1-scores ranging between 0.40 and 0.75 and specificities nearing 95.9%.
CONCLUSIONS: Although in its infancy, the convergence of AI and data science techniques, especially ML and NLP, with SDOH in EM offers transformative possibilities for better usage and integration of social data into clinical care and research. With a significant focus on the ED and notable NLP model performance, there is an imperative to standardize SDOH data collection, refine algorithms for diverse patient groups, and champion interdisciplinary synergies. These efforts aim to harness SDOH data optimally, enhancing patient care and mitigating health disparities. Our research underscores the vital need for continued investigation in this domain.
PMID:39475815 | DOI:10.2196/57124
Assessing the Generalizability of Deep Learning Models Trained on Standardized and Nonstandardized Images and Their Performance Against Teledermatologists: Retrospective Comparative Study
JMIR Dermatol. 2022 Sep 12;5(3):e35150. doi: 10.2196/35150.
ABSTRACT
BACKGROUND: Convolutional neural networks (CNNs) are a type of artificial intelligence that shows promise as a diagnostic aid for skin cancer. However, the majority are trained using retrospective image data sets with varying image capture standardization.
OBJECTIVE: The aim of our study was to use CNN models with the same architecture-trained on image sets acquired with either the same image capture device and technique (standardized) or with varied devices and capture techniques (nonstandardized)-and test variability in performance when classifying skin cancer images in different populations.
METHODS: In all, 3 CNNs with the same architecture were trained. CNN nonstandardized (CNN-NS) was trained on 25,331 images taken from the International Skin Imaging Collaboration (ISIC) using different image capture devices. CNN standardized (CNN-S) was trained on 177,475 MoleMap images taken with the same capture device, and CNN standardized number 2 (CNN-S2) was trained on a subset of 25,331 standardized MoleMap images (matched for number and classes of training images to CNN-NS). These 3 models were then tested on 3 external test sets: 569 Danish images, the publicly available ISIC 2020 data set consisting of 33,126 images, and The University of Queensland (UQ) data set of 422 images. Primary outcome measures were sensitivity, specificity, and area under the receiver operating characteristic curve (AUROC). Teledermatology assessments available for the Danish data set were used to determine model performance compared to teledermatologists.
RESULTS: When tested on the 569 Danish images, CNN-S achieved an AUROC of 0.861 (95% CI 0.830-0.889) and CNN-S2 achieved an AUROC of 0.831 (95% CI 0.798-0.861; standardized models), with both outperforming CNN-NS (nonstandardized model; P=.001 and P=.009, respectively), which achieved an AUROC of 0.759 (95% CI 0.722-0.794). When tested on 2 additional data sets (ISIC 2020 and UQ), CNN-S (P<.001 and P<.001, respectively) and CNN-S2 (P=.08 and P=.35, respectively) still outperformed CNN-NS. When the CNNs were matched to the mean sensitivity and specificity of the teledermatologists on the Danish data set, the models' resultant sensitivities and specificities were surpassed by the teledermatologists. However, when compared to CNN-S, the differences were not statistically significant (sensitivity: P=.10; specificity: P=.053). Performance across all CNN models as well as teledermatologists was influenced by image quality.
CONCLUSIONS: CNNs trained on standardized images had improved performance and, therefore, greater generalizability in skin cancer classification when applied to unseen data sets. This finding is an important consideration for future algorithm development, regulation, and approval.
PMID:39475778 | DOI:10.2196/35150
Addressing visual impairments: Essential software requirements for image caption solutions
Assist Technol. 2024 Oct 30:1-16. doi: 10.1080/10400435.2024.2413650. Online ahead of print.
ABSTRACT
Visually impaired individuals actively utilize devices like computers, tablets, and smartphones, due to advancements in screen reader technologies. Integrating freely available deep learning models, image captioning can further enhance these readers, providing an affordable assistive tech solution. This research outlines the critical software requirements necessary for image captioning tools to effectively serve this demographic. Two qualitative investigations were conducted to determine these requirements. An online survey was first conducted to identify the main preferences of visually impaired users in relation to audio descriptive software, with findings visualized using word clouds. A subsequent study evaluated the proficiency of existing deep learning captioning models in addressing these stipulated requirements. Emphasizing the need for comprehensive image data, the results highlighted three primary areas: 1) characteristics of individuals, 2) color specifics of objects, and 3) the overall context of images. The research indicates that current captioning tools are not entirely effective for the visually impaired. Based on the delineated requirements and suggested future research paths, there is potential for the development of improved image captioning systems, advancing digital accessibility for the visually impaired.
PMID:39475401 | DOI:10.1080/10400435.2024.2413650
Automatic tumor segmentation and lymph node metastasis prediction in papillary thyroid carcinoma using ultrasound keyframes
Med Phys. 2024 Oct 30. doi: 10.1002/mp.17498. Online ahead of print.
ABSTRACT
BACKGROUND: Accurate preoperative prediction of cervical lymph node metastasis (LNM) for papillary thyroid carcinoma (PTC) patients is essential for disease staging and individualized treatment planning, which can improve prognosis and facilitate better management.
PURPOSE: To establish a fully automated deep learning-enabled model (FADLM) for automated tumor segmentation and cervical LNM prediction in PTC using ultrasound (US) video keyframes.
METHODS: The bicentral study retrospective enrolled 518 PTC patients, who were then randomly divided into the training (Hospital 1, n = 340), internal test (Hospital 1, n = 83), and external test cohorts (Hospital 2, n = 95). The FADLM integrated mask region-based convolutional neural network (Mask R-CNN) for automatic thyroid primary tumor segmentation and ResNet34 with Bayes strategy for cervical LNM diagnosis. A radiomics model (RM) using the same automated segmentation method, a traditional radiomics model (TRM) using manual segmentation, and a clinical-semantic model (CSM) were developed for comparison. The dice similarity coefficient (DSC) was used to evaluate segmentation performance. The prediction performance of the models was validated in terms of discrimination and clinical utility with the area under the receiver operator characteristic curve (AUC), heatmap analysis, and decision curve analysis (DCA). The comparison of the predictive performance among different models was conducted by DeLong test. The performances of two radiologists compared with FADLM and the diagnostic augmentation with FADLM's assistance were analyzed in terms of accuracy, sensitivity and specificity using McNemar's x2 test. The p-value less than 0.05 was defined as a statistically significant difference. The Benjamini-Hochberg procedure was applied for multiple comparisons to deal with Type I error.
RESULTS: The FADLM yielded promising segmentation results in training (DSC: 0.88 ± 0.23), internal test (DSC: 0.88 ± 0.23), and external test cohorts (DSC: 0.85 ± 0.24). The AUCs of FADLM for cervical LNM prediction were 0.78 (95% CI: 0.73, 0.83), 0.83 (95% CI: 0.74, 0.92), and 0.83 (95% CI: 0.75, 0.92), respectively. It all significantly outperformed the RM (AUCs: 0.78 vs. 0.72; 0.83 vs. 0.65; 0.83 vs. 0.68, all adjusted p-values < 0.05) and CSM (AUCs: 0.78 vs. 0.71; 0.83 vs. 0.62; 0.83 vs. 0.68, all adjusted p-values < 0.05) across the three cohorts. The RM offered similar performance to that of TRM (AUCs: 0.61 vs. 0.63, adjusted p-value = 0.60) while significantly reducing the segmentation time (3.3 ± 3.8 vs. 14.1 ± 4.2 s, p-value < 0.001). Under the assistance of FADLM, the accuracies of junior and senior radiologists were improved by 18% and 15% (all adjusted p-values < 0.05) and the sensitivities by 25% and 21% (all adjusted p-values < 0.05) in the external test cohort.
CONCLUSION: The FADLM with elaborately designed automated strategy using US video keyframes holds good potential to provide an efficient and consistent prediction of cervical LNM in PTC. The FADLM displays superior performance to RM, CSM, and radiologists with promising efficacy.
PMID:39475358 | DOI:10.1002/mp.17498
Investigating the role of auditory cues in modulating motor timing: insights from EEG and deep learning
Cereb Cortex. 2024 Oct 3;34(10):bhae427. doi: 10.1093/cercor/bhae427.
ABSTRACT
Research on action-based timing has shed light on the temporal dynamics of sensorimotor coordination. This study investigates the neural mechanisms underlying action-based timing, particularly during finger-tapping tasks involving synchronized and syncopated patterns. Twelve healthy participants completed a continuation task, alternating between tapping in time with an auditory metronome (pacing) and continuing without it (continuation). Electroencephalography data were collected to explore how neural activity changes across these coordination modes and phases. We applied deep learning methods to classify single-trial electroencephalography data and predict behavioral timing conditions. Results showed significant classification accuracy for distinguishing between pacing and continuation phases, particularly during the presence of auditory cues, emphasizing the role of auditory input in motor timing. However, when auditory components were removed from the electroencephalography data, the differentiation between phases became inconclusive. Mean accuracy asynchrony, a measure of timing error, emerged as a superior predictor of performance variability compared to inter-response interval. These findings highlight the importance of auditory cues in modulating motor timing behaviors and present the challenges of isolating motor activation in the absence of auditory stimuli. Our study offers new insights into the neural dynamics of motor timing and demonstrates the utility of deep learning in analyzing single-trial electroencephalography data.
PMID:39475113 | DOI:10.1093/cercor/bhae427
Prediction of Cervical Cancer Lymph Node Metastasis via a Multimodal Transfer Learning Approach
Br J Hosp Med (Lond). 2024 Oct 30;85(10):1-14. doi: 10.12968/hmed.2024.0428. Epub 2024 Oct 29.
ABSTRACT
Aims/Background In the treatment of patients with cervical cancer, lymph node metastasis (LNM) is an important indicator for stratified treatment and prognosis of cervical cancer. This study aimed to develop and validate a multimodal model based on contrast-enhanced multiphase computed tomography (CT) images and clinical variables to accurately predict LNM in patients with cervical cancer. Methods This study included 233 multiphase contrast-enhanced CT images of patients with pathologically confirmed cervical malignancies treated at the Affiliated Dongyang Hospital of Wenzhou Medical University. A three-dimensional MedicalNet pre-trained model was used to extract features. Minimum redundancy-maximum correlation, and least absolute shrinkage and selection operator regression were used to screen the features that were ultimately combined with clinical candidate predictors to build the prediction model. The area under the curve (AUC) was used to assess the predictive efficacy of the model. Results The results indicate that the deep transfer learning model exhibited high diagnostic performance within the internal validation set, with an AUC of 0.82, accuracy of 0.88, sensitivity of 0.83, and specificity of 0.89. Conclusion We constructed a comprehensive, multiparameter model based on the concept of deep transfer learning, by pre-training the model with contrast-enhanced multiphase CT images and an array of clinical variables, for predicting LNM in patients with cervical cancer, which could aid the clinical stratification of these patients via a noninvasive manner.
PMID:39475034 | DOI:10.12968/hmed.2024.0428
Artificial Intelligence Assisted Surgical Scene Recognition: A Comparative Study Amongst Healthcare Professionals
Ann Surg. 2024 Oct 30. doi: 10.1097/SLA.0000000000006577. Online ahead of print.
ABSTRACT
OBJECTIVE: This study aimed to compare the ability of a deep-learning platform (the MACSSwin-T model) with healthcare professionals in detecting cerebral aneurysms from operative videos. Secondly, we aimed to compare the neurosurgical team's ability to detect cerebral aneurysms with and without AI-assistance.
BACKGROUND: Modern microscopic surgery enables the capture of operative video data on an unforeseen scale. Advances in computer vision, a branch of artificial intelligence (AI), have enabled automated analysis of operative video. These advances are likely to benefit clinicians, healthcare systems, and patients alike, yet such benefits are yet to be realised.
METHODS: In a cross-sectional comparative study, neurosurgeons, anaesthetists, and operating room (OR) nurses, all at varying stages of training and experience, reviewed still frames of aneurysm clipping operations and labelled frames as "aneurysm not in frame" or "aneurysm in frame". Frames then underwent analysis by the AI platform. A second round of data collection was performed whereby the neurosurgical team had AI-assistance. Accuracy of aneurysm detection was calculated for human only, AI only, and AI-assisted human groups.
RESULTS: 5,154 individual frame reviews were collated from 338 healthcare professionals. Healthcare professionals correctly labelled 70% of frames without AI assistance, compared to 78% with AI-assistance (OR 1.77, P<0.001). Neurosurgical Attendings showed the greatest improvement, from 77% to 92% correct predictions with AI-assistance (OR 4.24, P=0.003).
CONCLUSION: AI-assisted human performance surpassed both human and AI alone. Notably, across healthcare professionals surveyed, frame accuracy improved across all subspecialties and experience levels, particularly among the most experienced healthcare professionals. These results challenge the prevailing notion that AI primarily benefits junior clinicians, highlighting its crucial role throughout the surgical hierarchy as an essential component of modern surgical practice.
PMID:39474680 | DOI:10.1097/SLA.0000000000006577
Rapid identification of chemical profiles in vitro and in vivo of Huan Shao Dan and potential anti-aging metabolites by high-resolution mass spectrometry, sequential metabolism, and deep learning model
Front Pharmacol. 2024 Oct 15;15:1432592. doi: 10.3389/fphar.2024.1432592. eCollection 2024.
ABSTRACT
BACKGROUND: Aging is marked by the gradual deterioration of cells, tissues, and organs and is a major risk factor for many chronic diseases. Considering the complex mechanisms of aging, traditional Chinese medicine (TCM) could offer distinct advantages. However, due to the complexity and variability of metabolites in TCM, the comprehensive screening of metabolites associated with pharmacology remains a significant issue.
METHODS: A reliable and integrated identification method based on UPLC-Q Exactive-Orbitrap HRMS was established to identify the chemical profiles of Huan Shao Dan (HSD). Then, based on the theory of sequential metabolism, the metabolic sites of HSD in vivo were further investigated. Finally, a deep learning model and a bioactivity assessment assay were applied to screen potential anti-aging metabolites.
RESULTS: This study identified 366 metabolites in HSD. Based on the results of sequential metabolism, 135 metabolites were then absorbed into plasma. A total of 178 peaks were identified from the sample after incubation with artificial gastric juice. In addition, 102 and 91 peaks were identified from the fecal and urine samples, respectively. Finally, based on the results of the deep learning model and bioactivity assay, ginsenoside Rg1, Rg2, and Rc, pseudoginsenoside F11, and jionoside B1 were selected as potential anti-aging metabolites.
CONCLUSION: This study provides a valuable reference for future research on the material basis of HSD by describing the chemical profiles both in vivo and in vitro. Moreover, the proposed screening approach may serve as a rapid tool for identifying potential anti-aging metabolites in TCM.
PMID:39474607 | PMC:PMC11518704 | DOI:10.3389/fphar.2024.1432592
Revolutionizing Radiology With Artificial Intelligence
Cureus. 2024 Oct 29;16(10):e72646. doi: 10.7759/cureus.72646. eCollection 2024 Oct.
ABSTRACT
Artificial intelligence (AI) is rapidly transforming the field of radiology, offering significant advancements in diagnostic accuracy, workflow efficiency, and patient care. This article explores AI's impact on various subfields of radiology, emphasizing its potential to improve clinical practices and enhance patient outcomes. AI-driven technologies such as machine learning, deep learning, and natural language processing (NLP) are playing a pivotal role in automating routine tasks, aiding in early disease detection, and supporting clinical decision-making, allowing radiologists to focus on more complex diagnostic challenges. Key applications of AI in radiology include improving image analysis through computer-aided diagnosis (CAD) systems, which enhance the detection of abnormalities in imaging, such as tumors. AI tools have demonstrated high accuracy in analyzing medical images, integrating data from multiple imaging modalities such as CT, MRI, and PET to provide comprehensive diagnostic insights. These advancements facilitate personalized treatment planning and complement radiologists' workflows. However, for AI to be fully integrated into radiology workflows, several challenges must be addressed, including ensuring transparency in how AI algorithms work, protecting patient data, and avoiding biases that could affect diverse populations. Developing explainable AI systems that can clearly show how decisions are made is crucial, as is ensuring AI tools can seamlessly fit into existing radiology systems. Collaboration between radiologists, AI developers, and policymakers, alongside strong ethical guidelines and regulatory oversight, will be key to ensuring AI is implemented safely and effectively in clinical practice. Overall, AI holds tremendous promise in revolutionizing radiology. Through its ability to automate complex tasks, enhance diagnostic capabilities, and streamline workflows, AI has the potential to significantly improve the quality and efficiency of radiology practices. Continued research, development, and collaboration will be crucial in unlocking AI's full potential and addressing the challenges that accompany its adoption.
PMID:39474591 | PMC:PMC11521355 | DOI:10.7759/cureus.72646
Enhancing Open-World Bacterial Raman Spectra Identification by Feature Regularization for Improved Resilience against Unknown Classes
Chem Biomed Imaging. 2024 May 6;2(6):442-452. doi: 10.1021/cbmi.4c00007. eCollection 2024 Jun 24.
ABSTRACT
The combination of deep learning techniques and Raman spectroscopy shows great potential offering precise and prompt identification of pathogenic bacteria in clinical settings. However, the traditional closed-set classification approaches assume that all test samples belong to one of the known pathogens, and their applicability is limited since the clinical environment is inherently unpredictable and dynamic, unknown, or emerging pathogens may not be included in the available catalogs. We demonstrate that the current state-of-the-art neural networks identifying pathogens through Raman spectra are vulnerable to unknown inputs, resulting in an uncontrollable false positive rate. To address this issue, first we developed an ensemble of ResNet architectures combined with the attention mechanism that achieves a 30-isolate accuracy of 87.8 ± 0.1%. Second, through the integration of feature regularization by the Objectosphere loss function, our model both achieves high accuracy in identifying known pathogens from the catalog and effectively separates unknown samples drastically reducing the false positive rate. Finally, the proposed feature regularization method during training significantly enhances the performance of out-of-distribution detectors during the inference phase improving the reliability of the detection of unknown classes. Our algorithm for Raman spectroscopy empowers the identification of previously unknown, uncataloged, and emerging pathogens ensuring adaptability to future pathogens that may surface. Moreover, it can be extended to enhance open-set medical image classification, bolstering its reliability in dynamic operational settings.
PMID:39474520 | PMC:PMC11503672 | DOI:10.1021/cbmi.4c00007
Impact of phased COVID-19 vaccine rollout on anxiety and depression among US adult population, January 2019-February 2023: a population-based interrupted time series analysis
Lancet Reg Health Am. 2024 Aug 9;37:100852. doi: 10.1016/j.lana.2024.100852. eCollection 2024 Sep.
ABSTRACT
BACKGROUND: Existing research lacks information on the potential impacts of multi-phased coronavirus disease 2019 (COVID-19) vaccine rollouts on population mental health. This study aims to evaluate the impact of various COVID-19 vaccine rollout phases on trends and prevalence of anxiety and depression among US adults at a population level.
METHODS: We performed a US population-based multi-intervention interrupted time series analysis through Deep Learning and autoregressive integrated moving average (ARIMA) approaches, analyzing 4 waves of US CDC's Behavioral Risk Factor Surveillance System (BRFSS) data (January 2019-February 2023) to assess changes in the weekly prevalence of anxiety and depression following interruptions, including all major COVID-19 vaccine rollout phases from 2020 to early 2023 while considering pandemic-related events.
FINDINGS: Among 1,615,643 US adults (1,011,300 [76.4%] aged 18-64 years, 867,826 [51.2%] female, 126,594 [16.9%] Hispanic, 120,380 [11.9%] non-Hispanic Black, 1,191,668 [61.7%] non-Hispanic White, and 113,461 [9.5%] other non-Hispanic people of color), we found that three COVID-19 vaccine rollout phases (ie, prioritization for educational/childcare workers, boosters for all US adults, authorization for young children) were associated with a 0.93 percentage-point (95% CI -1.81 to -0.04, p = 0.041), 1.28 percentage-point (95% CI -2.32 to -0.24, p = 0.017), and 0.89 percentage-point (95% CI -1.56 to -0.22, p = 0.010) reduction, respectively, in anxiety and depression prevalence among the general US adult population despite an upward trend in the prevalence of anxiety and depression from 2019 to early 2023. Among different population groups, Phase 1 was associated with increases in anxiety and depression prevalence among Black/African Americans (2.26 percentage-point, 95% CI 0.24-4.28, p = 0.029), other non-Hispanic people of color (2.68 percentage-point, 95% CI 0.36-5.00, p = 0.024), and lower-income individuals (3.95 percentage-point, 95% CI 2.20-5.71, p < 0.0001).
INTERPRETATION: Our findings suggest disparate effects of phased COVID-19 vaccine rollout on mental health across US populations, underlining the need for careful planning in future strategies for phased disease prevention and interventions.
FUNDING: None.
PMID:39474466 | PMC:PMC11519686 | DOI:10.1016/j.lana.2024.100852
Feasibility validation of automatic diagnosis of mitral valve prolapse from multi-view echocardiographic sequences based on deep neural network
Eur Heart J Imaging Methods Pract. 2024 Oct 28;2(4):qyae086. doi: 10.1093/ehjimp/qyae086. eCollection 2024 Oct.
ABSTRACT
AIMS: To address the limitations of traditional diagnostic methods for mitral valve prolapse (MVP), specifically fibroelastic deficiency (FED) and Barlow's disease (BD), by introducing an automated diagnostic approach utilizing multi-view echocardiographic sequences and deep learning.
METHODS AND RESULTS: An echocardiographic data set, collected from Zhongshan Hospital, Fudan University, containing apical 2 chambers (A2C), apical 3 chambers (A3C), and apical 4 chambers (A4C) views, was employed to train the deep learning models. We separately trained view-specific and view-agnostic deep neural network models, which were denoted as MVP-VS and MVP view-agonistic (VA), for MVP diagnosis. Diagnostic accuracy, precision, sensitivity, F1-score, and specificity were evaluated for both BD and FED phenotypes. MVP-VS demonstrated an overall diagnostic accuracy of 0.94 for MVP. In the context of BD diagnosis, precision, sensitivity, F1-score, and specificity were 0.83, 1.00, 0.90, and 0.92, respectively. For FED diagnosis, the metrics were 1.00, 0.83, 0.91, and 1.00. MVP-VA exhibited an overall accuracy of 0.95, with BD-specific metrics of 0.85, 1.00, 0.92, and 0.94 and FED-specific metrics of 1.00, 0.83, 0.91, and 1.00. In particular, the MVP-VA model using mixed views for training demonstrated efficient diagnostic performance, eliminating the need for repeated development of MVP-VS models and improving the efficiency of the clinical pipeline by using arbitrary views in the deep learning model.
CONCLUSION: This study pioneers the integration of artificial intelligence into MVP diagnosis and demonstrates the effectiveness of deep neural networks in overcoming the challenges of traditional diagnostic methods. The efficiency and accuracy of the proposed automated approach suggest its potential for clinical applications in the diagnosis of valvular heart disease.
PMID:39474265 | PMC:PMC11519029 | DOI:10.1093/ehjimp/qyae086
Minimum imaging dose for deep learning-based pelvic synthetic computed tomography generation from cone beam images
Phys Imaging Radiat Oncol. 2024 Mar 22;30:100569. doi: 10.1016/j.phro.2024.100569. eCollection 2024 Apr.
ABSTRACT
BACKGROUND AND PURPOSE: Daily cone-beam computed tomography (CBCT) in image-guided radiotherapy administers radiation exposure and subjects patients to secondary cancer risk. Reducing imaging dose remains challenging as image quality deteriorates. We investigated three imaging dose levels by reducing projections and correcting images using two deep learning algorithms, aiming at identifying the lowest achievable imaging dose.
MATERIALS AND METHODS: CBCTs were reconstructed with 100%, 25%, 15% and 10% projections. Models were trained (30), validated (3) and tested (8) with prostate cancer patient datasets. We optimized and compared the performance of 1) a cycle generative adversarial network (cycleGAN) with residual connection and 2) a contrastive unpaired translation network (CUT) to generate synthetic computed tomography (sCT) from reduced imaging dose CBCTs. Volumetric modulated arc therapy plans were optimized on a reference intensity-corrected full dose CBCTcor and recalculated on sCTs. Hounsfield unit (HU) and positioning accuracy were evaluated. Bladder and rectum were manually delineated to determine anatomical fidelity.
RESULTS: All sCTs achieved average mean absolute mean absolute error/structural similarity index measure/peak signal-to-noise ratio of ⩽ 59HU/ ⩾ 0.94/ ⩾ 33 dB. All dose-volume histogram parameter differences were within 2 Gy or 2 % . Positioning differences were ⩽ 0.30 mm or 0.30°. cycleGAN with Dice similarity coefficients (DSC) for bladder/rectum of ⩾ 0.85/ ⩾ 0.81 performed better than CUT ( ⩾ 0.83/ ⩾ 0.76). A significantly lower DSC accuracy was observed for 15 % and 10 % sCTs. cycleGAN performed better than CUT for contouring, however both yielded comparable outcomes in other evaluations.
CONCLUSION: sCTs based on different CBCT doses using cycleGAN and CUT were investigated. Based on segmentation accuracy, 25 % is the minimum imaging dose.
PMID:39474260 | PMC:PMC11519690 | DOI:10.1016/j.phro.2024.100569