Deep learning

Hierarchical Graph Attention Network with Positive and Negative Attentions for Improved Interpretability: ISA-PN

Mon, 2024-12-09 06:00

J Chem Inf Model. 2024 Dec 9. doi: 10.1021/acs.jcim.4c01035. Online ahead of print.

ABSTRACT

With the advancement of deep learning (DL) methods in chemistry and materials science, the interpretability of DL models has become a critical issue in elucidating quantitative (molecular) structure-property relationships. Although attention mechanisms have been generally employed to explain the importance of molecular substructures that contribute to molecular properties, their interpretability remains limited. In this work, we introduce a versatile segmentation method and develop an interpretable subgraph attention (ISA) network with positive and negative streams (ISA-PN) to enhance the understanding of molecular structure-property relationships. The predictive performance of the ISA models was validated using data sets for aqueous solubility, lipophilicity, and melting temperature, with a particular focus on evaluating interpretability for the aqueous solubility data set. The ISA-PN model enables the quantification of the contributions of molecular substructures through positive and negative attention scores. Comparative analyses of the ISA, ISA-PN, and GC-Net (group contribution network) models demonstrate that the ISA-PN model significantly improves interpretability while maintaining similar accuracy levels. This study highlights the efficacy of the ISA-PN model in providing meaningful insights into the contributions of molecular substructures to molecular properties, thereby enhancing the interpretability of DL models in chemical applications.

PMID:39654089 | DOI:10.1021/acs.jcim.4c01035

Categories: Literature Watch

Prediction of Nursing Need Proxies Using Vital Signs and Biomarkers Data: Application of Deep Learning Models

Mon, 2024-12-09 06:00

J Clin Nurs. 2024 Dec 9. doi: 10.1111/jocn.17612. Online ahead of print.

ABSTRACT

AIM: To develop deep learning models to predict nursing need proxies among hospitalised patients and compare their predictive efficacy to that of a traditional regression model.

DESIGN: This methodological study employed a cross-sectional secondary data analysis.

METHODS: This study used de-identified electronic health records data from 20,855 adult patients aged 20 years or older, admitted to the general wards at a tertiary hospital. The models utilised patient information covering the preceding 2 days, comprising vital signs, biomarkers and demographic data. To create nursing need proxies, we identified the six highest-workload nursing tasks. We structured the collected data sequentially to facilitate processing via recurrent neural network (RNN) and long short-term memory (LSTM) algorithms. The STROBE checklist for cross-sectional studies was used for reporting.

RESULTS: Both the RNN and LSTM predicted nursing need proxies more effectively than the traditional regression model. However, upon testing the models using a sample case dataset, we observed a notable reduction in prediction accuracy during periods marked by rapid change.

CONCLUSIONS: The RNN and LSTM, which enhanced predictive performance for nursing needs, were developed using iterative learning processes. The RNN and LSTM demonstrated predictive capabilities superior to the traditional multiple regression model for nursing need proxies.

IMPLICATIONS FOR THE PROFESSION: Applying these predictive models in clinical settings where medical care complexity and diversity are increasing could substantially mitigate the uncertainties inherent in decision-making processes.

PATIENT OR PUBLIC CONTRIBUTION: We used de-identified electronic health record data of 20,855 adult patients about vital signs, biomarkers and nursing activities.

REPORTING METHOD: The authors state that they have adhered to relevant EQUATOR guidelines: STROBE statement for cross-sectional studies.

IMPACT: Despite widespread adoption of deep learning algorithms in various industries, their application in nursing administration for workload distribution and staffing adequacy remains limited. This study amalgamated deep learning technology to develop a predictive model to proactively forecast nursing need proxies. Our study demonstrates that both the RNN and LSTM models outperform a traditional regression model in predicting nursing need proxies. The proactive application of deep learning methods for nursing need prediction could help facilitate timely detection of changes in patient nursing demands, enabling the effective and safe nursing services.

PMID:39654010 | DOI:10.1111/jocn.17612

Categories: Literature Watch

Fusion Learning from Non-contrast CT Scans for the Detection of Hemorrhagic Transformation in Stroke Patients

Mon, 2024-12-09 06:00

J Imaging Inform Med. 2024 Dec 9. doi: 10.1007/s10278-024-01350-0. Online ahead of print.

ABSTRACT

Hemorrhagic transformation (HT) is a potentially catastrophic complication after acute ischemic stroke. Prevention of HT risk is crucial because it worsens prognosis and increases mortality. This study aimed at developing and validating a computer-aided diagnosis system using pretreatment non-contrast computed tomography (CT) scans for HT prediction in stroke patients undergoing revascularization. This retrospective study included all acute ischemic stroke patients with non-contrast CT before reperfusion therapy who also underwent follow-up MRI from January 2018 to December 2022. Among the 188 evaluated patients, any degree of HT at follow-up imaging was observed in 103 patients. HT diagnosis via MRI was defined as the reference standard for neuroradiologists. Using a database of 2076 serial non-contrast CT images of the brain, pretrained deep learning architectures such as convolutional neural networks and vision transformers (ViTs) were used for feature extraction. The performance of the predictive HT risk model was evaluated via tenfold cross-validation in machine learning classifiers. The accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC) were evaluated. Using an individual deep learning architecture, DenseNet201 features achieved the highest accuracy of 87% and an AUC of 0.8863 in the classifier of the subspace ensemble k-nearest neighbor. By combining the DenseNet201 and ViT features, the accuracy and AUC can be improved to 88% and 0.8987, respectively, which are significantly better than those of using ViT alone. Detecting HT in stroke patients is a meaningful but challenging issue. On the basis of the model approach, HT diagnosis would be more automatic, efficient, and consistent, which would be helpful in clinic use.

PMID:39653876 | DOI:10.1007/s10278-024-01350-0

Categories: Literature Watch

An explainable deep learning platform for molecular discovery

Mon, 2024-12-09 06:00

Nat Protoc. 2024 Dec 9. doi: 10.1038/s41596-024-01084-x. Online ahead of print.

ABSTRACT

Deep learning approaches have been increasingly applied to the discovery of novel chemical compounds. These predictive approaches can accurately model compounds and increase true discovery rates, but they are typically black box in nature and do not generate specific chemical insights. Explainable deep learning aims to 'open up' the black box by providing generalizable and human-understandable reasoning for model predictions. These explanations can augment molecular discovery by identifying structural classes of compounds with desired activity in lieu of lone compounds. Additionally, these explanations can guide hypothesis generation and make searching large chemical spaces more efficient. Here we present an explainable deep learning platform that enables vast chemical spaces to be mined and the chemical substructures underlying predicted activity to be identified. The platform relies on Chemprop, a software package implementing graph neural networks as a deep learning model architecture. In contrast to similar approaches, graph neural networks have been shown to be state of the art for molecular property prediction. Focusing on discovering structural classes of antibiotics, this protocol provides guidelines for experimental data generation, model implementation and model explainability and evaluation. This protocol does not require coding proficiency or specialized hardware, and it can be executed in as little as 1-2 weeks, starting from data generation and ending in the testing of model predictions. The platform can be broadly applied to discover structural classes of other small molecules, including anticancer, antiviral and senolytic drugs, as well as to discover structural classes of inorganic molecules with desired physical and chemical properties.

PMID:39653800 | DOI:10.1038/s41596-024-01084-x

Categories: Literature Watch

ATP_mCNN: Predicting ATP binding sites through pretrained language models and multi-window neural networks

Mon, 2024-12-09 06:00

Comput Biol Med. 2024 Dec 8;185:109541. doi: 10.1016/j.compbiomed.2024.109541. Online ahead of print.

ABSTRACT

Adenosine triphosphate plays a vital role in providing energy and enabling key cellular processes through interactions with binding proteins. The increasing amount of protein sequence data necessitates computational methods for identifying binding sites. However, experimental identification of adenosine triphosphate-binding residues remains challenging. To address the challenge, we developed a multi-window convolutional neural network architecture taking pre-trained protein language model embeddings as input features. In particular, multiple parallel convolutional layers scan for motifs localized to different window sizes. Max pooling extracts salient features concatenated across windows into a final multi-scale representation for residue-level classification. On benchmark datasets, our model achieves an area under the ROC curve of 0.95, significantly improving on prior sequence-based models and outperforming convolutional neural network baselines. This demonstrates the utility of pre-trained language models and multi-window convolutional neural networks for advanced sequence-based prediction of adenosine triphosphate-binding residues. Our approach provides a promising new direction for elucidating binding mechanisms and interactions from primary structure.

PMID:39653625 | DOI:10.1016/j.compbiomed.2024.109541

Categories: Literature Watch

CGPDTA: An Explainable Transfer Learning-Based Predictor With Molecule Substructure Graph for Drug-Target Binding Affinity

Mon, 2024-12-09 06:00

J Comput Chem. 2025 Jan 5;46(1):e27538. doi: 10.1002/jcc.27538.

ABSTRACT

Identifying interactions between drugs and targets is crucial for drug discovery and development. Nevertheless, the determination of drug-target binding affinities (DTAs) through traditional experimental methods is a time-consuming process. Conventional approaches to predicting drug-target interactions (DTIs) frequently prove inadequate due to an insufficient representation of drugs and targets, resulting in ineffective feature capture and questionable interpretability of results. To address these challenges, we introduce CGPDTA, a novel deep learning framework empowered by transfer learning, designed explicitly for the accurate prediction of DTAs. CGPDTA leverages the complementarity of drug-drug and protein-protein interaction knowledge through advanced drug and protein language models. It further enhances predictive capability and interpretability by incorporating molecular substructure graphs and protein pocket sequences to represent local features of drugs and targets effectively. Our findings demonstrate that CGPDTA not only outperforms existing methods in accuracy but also provides meaningful insights into the predictive process, marking a significant advancement in the field of drug discovery.

PMID:39653581 | DOI:10.1002/jcc.27538

Categories: Literature Watch

Current State of Community-Driven Radiological AI Deployment in Medical Imaging

Mon, 2024-12-09 06:00

JMIR AI. 2024 Dec 9;3:e55833. doi: 10.2196/55833.

ABSTRACT

Artificial intelligence (AI) has become commonplace in solving routine everyday tasks. Because of the exponential growth in medical imaging data volume and complexity, the workload on radiologists is steadily increasing. AI has been shown to improve efficiency in medical image generation, processing, and interpretation, and various such AI models have been developed across research laboratories worldwide. However, very few of these, if any, find their way into routine clinical use, a discrepancy that reflects the divide between AI research and successful AI translation. The goal of this paper is to give an overview of the intersection of AI and medical imaging landscapes. We also want to inform the readers about the importance of using standards in their radiology workflow and the challenges associated with deploying AI models in the clinical workflow. The main focus of this paper is to examine the existing condition of radiology workflow and identify the challenges hindering the implementation of AI in hospital settings. This report reflects extensive weekly discussions and practical problem-solving expertise accumulated over multiple years by industry experts, imaging informatics professionals, research scientists, and clinicians. To gain a deeper understanding of the requirements for deploying AI models, we introduce a taxonomy of AI use cases, supplemented by real-world instances of AI model integration within hospitals. We will also explain how the need for AI integration in radiology can be addressed using the Medical Open Network for AI (MONAI). MONAI is an open-source consortium for providing reproducible deep learning solutions and integration tools for radiology practice in hospitals.

PMID:39653370 | DOI:10.2196/55833

Categories: Literature Watch

Artificial Intelligence-based Detection of Tent-Like Signs in Intracardiac Echocardiography to Assist Transseptal Puncture

Mon, 2024-12-09 06:00

J Am Soc Echocardiogr. 2024 Dec 7:S0894-7317(24)00626-6. doi: 10.1016/j.echo.2024.11.010. Online ahead of print.

NO ABSTRACT

PMID:39653193 | DOI:10.1016/j.echo.2024.11.010

Categories: Literature Watch

Deep learning using histological images for gene mutation prediction in lung cancer: a multicentre retrospective study

Mon, 2024-12-09 06:00

Lancet Oncol. 2024 Dec 6:S1470-2045(24)00599-0. doi: 10.1016/S1470-2045(24)00599-0. Online ahead of print.

ABSTRACT

BACKGROUND: Accurate detection of driver gene mutations is crucial for treatment planning and predicting prognosis for patients with lung cancer. Conventional genomic testing requires high-quality tissue samples and is time-consuming and resource-consuming, and as a result, is not available for most patients, especially those in low-resource settings. We aimed to develop an annotation-free Deep learning-enabled artificial intelligence method to predict GEne Mutations (DeepGEM) from routinely acquired histological slides.

METHODS: In this multicentre retrospective study, we collected data for patients with lung cancer who had a biopsy and multigene next-generation sequencing done at 16 hospitals in China (with no restrictions on age, sex, or histology type), to form a large multicentre dataset comprising paired pathological image and multiple gene mutation information. We also included patients from The Cancer Genome Atlas (TCGA) publicly available dataset. Our developed model is an instance-level and bag-level co-supervised multiple instance learning method with label disambiguation design. We trained and initially tested the DeepGEM model on the internal dataset (patients from the First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China), and further evaluated it on the external dataset (patients from the remaining 15 centres) and the public TCGA dataset. Additionally, a dataset of patients from the same medical centre as the internal dataset, but without overlap, was used to evaluate the model's generalisation ability to biopsy samples from lymph node metastases. The primary objective was the performance of the DeepGEM model in predicting gene mutations (area under the curve [AUC] and accuracy) in the four prespecified groups (ie, the hold-out internal test set, multicentre external test set, TCGA set, and lymph node metastases set).

FINDINGS: Assessable pathological images and multigene testing information were available for 3697 patients who had biopsy and multigene next-generation sequencing done between Jan 1, 2018, and March 31, 2022, at the 16 centres. We excluded 60 patients with low-quality images. We included 3767 images from 3637 consecutive patients (1978 [54·4%] men, 1514 [41·6%] women, 145 [4·0%] unknown; median age 60 years [IQR 52-67]), with 1716 patients in the internal dataset, 1718 patients in the external dataset, and 203 patients in the lymph node metastases dataset. The DeepGEM model showed robust performance in the internal dataset: for excisional biopsy samples, AUC values for gene mutation prediction ranged from 0·90 (95% CI 0·77-1·00) to 0·97 (0·93-1·00) and accuracy values ranged from 0·91 (0·85-0·98) to 0·97 (0·93-1·00); for aspiration biopsy samples, AUC values ranged from 0·85 (0·80-0·91) to 0·95 (0·86-1·00) and accuracy values ranged from 0·79 (0·74-0·85) to 0·99 (0·98-1·00). In the multicentre external dataset, for excisional biopsy samples, AUC values ranged from 0·80 (95% CI 0·75-0·85) to 0·91 (0·88-1·00) and accuracy values ranged from 0·79 (0·76-0·82) to 0·95 (0·93-0·96); for aspiration biopsy samples, AUC values ranged from 0·76 (0·70-0·83) to 0·87 (0·80-0·94) and accuracy values ranged from 0·76 (0·74-0·79) to 0·97 (0·96-0·98). The model also showed strong performance on the TCGA dataset (473 patients; 535 slides; AUC values ranged from 0·82 [95% CI 0·71-0·93] to 0·96 [0·91-1·00], accuracy values ranged from 0·79 [0·70-0·88] to 0·95 [0·90-1·00]). The DeepGEM model, trained on primary region biopsy samples, could be generalised to biopsy samples from lymph node metastases, with AUC values of 0·91 (95% CI 0·88-0·94) for EGFR and 0·88 (0·82-0·93) for KRAS and accuracy values of 0·85 (0·80-0·88) for EGFR and 0·95 (0·92-0·96) for KRAS and showed potential for prognostic prediction of targeted therapy. The model generated spatial gene mutation maps, indicating gene mutation spatial distribution.

INTERPRETATION: We developed an AI-based method that can provide an accurate, timely, and economical prediction of gene mutation and mutation spatial distribution. The method showed substantial potential as an assistive tool for guiding the clinical treatment of patients with lung cancer.

FUNDING: National Natural Science Foundation of China, the Science and Technology Planning Project of Guangzhou, and the National Key Research and Development Program of China.

TRANSLATION: For the Chinese translation of the abstract see Supplementary Materials section.

PMID:39653054 | DOI:10.1016/S1470-2045(24)00599-0

Categories: Literature Watch

A pathway from surface to deep online language learning approach: The crucial role of online self-regulation

Mon, 2024-12-09 06:00

Acta Psychol (Amst). 2024 Dec 8;251:104644. doi: 10.1016/j.actpsy.2024.104644. Online ahead of print.

ABSTRACT

Students' approaches to online technologies (SAOLT) have garnered significant support among scholars from a variety of fields, particularly in the field of science, technology, engineering, and mathematics education (STEM). The humanistic field needs to explore learners' approaches to target learning contexts, especially in the fields of computer-assisted language learning (CALL) and psycholinguistics. To fill this gap, the researchers explored 686 Iranian high school EFL learners' approaches to online language learning (OLL) with regard to their pedagogical, technical, and peer support, as well as the mediation role of online self-regulation. The partial least squares modeling approach's (PLS-SEM) reflective analysis validated the factorial structure of the conceptual study framework in online language learning and secondary education. The formative model found that instructional and peer support positively impacted language learners' deep and surface approaches to online language learning, thereby maximizing and minimizing their online presence and meaningful language learning. Furthermore, the mediation analysis revealed the significant moderating influence of learners' online self-regulation, which shifted the correlations from relating both learners' perceived support to their surface and deep learning approaches to only relating to the deep learning approach. Consequently, this study has pedagogical implications as it introduces a new conceptual framework to the CALL and psycholinguistic domains, specifically incorporating a new psychological factor related to language learners' complex dynamic systems, known as Language Learners' Approaches to Online Language Learning (OLLA). To do so, the researchers suggested that instructors should design more practical activities for theoretical subjects such as English language learning in online learning and shift online teaching from teacher-centered to peer-centered so as to foster their learners' deep learning approach to online language learning.

PMID:39652985 | DOI:10.1016/j.actpsy.2024.104644

Categories: Literature Watch

Volumetric Breast Density Estimation From Three-Dimensional Reconstructed Digital Breast Tomosynthesis Images Using Deep Learning

Mon, 2024-12-09 06:00

JCO Clin Cancer Inform. 2024 Dec;8:e2400103. doi: 10.1200/CCI.24.00103. Epub 2024 Dec 9.

ABSTRACT

PURPOSE: Breast density is a widely established independent breast cancer risk factor. With the increasing utilization of digital breast tomosynthesis (DBT) in breast cancer screening, there is an opportunity to estimate volumetric breast density (VBD) routinely. However, current available methods extrapolate VBD from two-dimensional (2D) images acquired using DBT and/or depend on the existence of raw DBT data, which is rarely archived by clinical centers because of storage constraints.

METHODS: We retrospectively analyzed 1,080 nonactionable three-dimensional (3D) reconstructed DBT screening examinations acquired between 2011 and 2016. Reference tissue segmentations were generated using previously validated software that uses 3D reconstructed slices and raw 2D DBT data. We developed a deep learning (DL) model that segments dense and fatty breast tissue from background. We then applied this model to estimate %VBD and absolute dense volume (ADV) in cm3 in a separate case-control sample (180 cases and 654 controls). We created two conditional logistic regression models, relating each model-derived density measurement to likelihood of contralateral breast cancer diagnosis, adjusted for age, BMI, family history, and menopausal status.

RESULTS: The DL model achieved unweighted and weighted Dice scores of 0.88 (standard deviation [SD] = 0.08) and 0.76 (SD = 0.15), respectively, on the held-out test set, demonstrating good agreement between the model and 3D reference segmentations. There was a significant association between the odds of breast cancer diagnosis and model-derived VBD (odds ratio [OR], 1.41 [95 % CI, 1.13 to 1.77]; P = .002), with an AUC of 0.65 (95% CI, 0.60 to 0.69). ADV was also significantly associated with breast cancer diagnosis (OR, 1.45 [95% CI, 1.22 to 1.73]; P < .001) with an AUC of 0.67 (95% CI, 0.62 to 0.71).

CONCLUSION: DL-derived density measures derived from 3D reconstructed DBT images are associated with breast cancer diagnosis.

PMID:39652797 | DOI:10.1200/CCI.24.00103

Categories: Literature Watch

Recognition of Conus species using a combined approach of supervised learning and deep learning-based feature extraction

Mon, 2024-12-09 06:00

PLoS One. 2024 Dec 9;19(12):e0313329. doi: 10.1371/journal.pone.0313329. eCollection 2024.

ABSTRACT

Cone snails are venomous marine gastropods comprising more than 950 species widely distributed across different habitats. Their conical shells are remarkably similar to those of other invertebrates in terms of color, pattern, and size. For these reasons, assigning taxonomic signatures to cone snail shells is a challenging task. In this report, we propose an ensemble learning strategy based on the combination of Random Forest (RF) and XGBoost (XGB) methods. We used 47,600 cone shell images of uniform size (224 x 224 pixels), which were split into an 80:20 train-test ratio. Prior to performing subsequent operations, these images were subjected to pre-processing and transformation. After applying a deep learning approach (Visual Geometry Group with a 16-layer deep model architecture) for feature extraction, model specificity was further assessed by including multiple related and unrelated seashell images. Both classifiers demonstrated comparable recognition ability on random test samples. The evaluation results suggested that RF outperformed XGB due to its high accuracy in recognizing Conus species, with an average precision of 95.78%. The area under the receiver operating characteristic curve was 0.99, indicating the model's optimal performance. The learning and validation curves also demonstrated a robust fit, with the training score reaching 1 and the validation score gradually increasing to 95 as more data was provided. These values indicate a well-trained model that generalizes effectively to validation data without significant overfitting. The gradual improvement in the validation score curve is crucial for ensuring model reliability and minimizing the risk of overfitting. Our findings revealed an interactive visualization. The performance of our proposed model suggests its potential for use with datasets of other mollusks, and optimal results may be achieved for their categorization and taxonomical characterization.

PMID:39652613 | DOI:10.1371/journal.pone.0313329

Categories: Literature Watch

Integrating Faith and Learning Using a Biblical Concept-Based Curriculum

Mon, 2024-12-09 06:00

J Christ Nurs. 2025 Jan-Mar 01;42(1):22-27. doi: 10.1097/CNJ.0000000000001226. Epub 2024 Dec 9.

ABSTRACT

The integration of faith and learning (IFL) in nursing has deep historical roots. However, barriers have developed to successful IFL in Christian higher education. A biblically based concept curriculum (BBCC) is proposed that emphasizes deep learning, critical thinking, and student-centered learning. Examples are provided of BBCC curricular integration, pedagogical methods, and assessments related to the IFL. Results of student evaluations before and after implementation of the BBCC identified an increase in students' perceptions of IFL in nursing courses.

PMID:39652482 | DOI:10.1097/CNJ.0000000000001226

Categories: Literature Watch

Automated segmentation of dental restorations using deep learning: exploring data augmentation techniques

Mon, 2024-12-09 06:00

Oral Radiol. 2024 Dec 9. doi: 10.1007/s11282-024-00794-y. Online ahead of print.

ABSTRACT

OBJECTIVES: Deep learning has revolutionized image analysis for dentistry. Automated segmentation of dental radiographs is of great importance towards digital dentistry. The performance of deep learning models heavily relies on the quality and diversity of the training data. Data augmentation is a widely used technique implemented in machine learning and deep learning to artificially increase the size and diversity of a training dataset by applying various transformations to the original data.

METHODS: This work aims to automatically segment implants, prostheses, and fillings in panoramic images using 9 different deep learning segmentation models. Later, it explores the effect of data augmentation methods on segmentation performance of the models. Eight different data augmentation techniques are examined. Performance is evaluated by well-accepted metrics such as intersection over union (IoU) and Dice coefficient.

RESULTS: While averaging the segmentation results for the three classes, IoU varies between 0.62 and 0.82 while Dice score is between 0.75 and 0.9 among deep learning models used. Augmentation techniques provided performance improvements of up to 3.37%, 5.75% and 8.75% for implant, prosthesis and filling classes, respectively.

CONCLUSIONS: Findings reveal that choosing optimal augmentation strategies depends on both model architecture and dental structure type.

PMID:39652261 | DOI:10.1007/s11282-024-00794-y

Categories: Literature Watch

Classification of speech arrests and speech impairments during awake craniotomy: a multi-databases analysis

Mon, 2024-12-09 06:00

Int J Comput Assist Radiol Surg. 2024 Dec 9. doi: 10.1007/s11548-024-03301-0. Online ahead of print.

ABSTRACT

PURPOSE: Awake craniotomy presents a unique opportunity to map and preserve critical brain functions, particularly speech, during tumor resection. The ability to accurately assess linguistic functions in real-time not only enhances surgical precision, but also contributes significantly to improving postoperative outcomes. However, today, its evaluation is subjective as it relies on a clinician's observations only. This paper explores the use of a deep learning based model for the objective assessment of speech arrest and speech impairments during awake craniotomy.

METHODS: We extracted 1883 3-second audio clips containing the patient's response following direct electrical stimulation from 23 awake craniotomies recorded from two operating rooms of the Tokyo Women's Medical University Hospital (Japan) and two awake craniotomies recorded from the University Hospital of Brest (France). A Wav2Vec2-based model has been trained and used to detect speech arrests and speech impairments. Experiments were performed with different datasets settings and preprocessing techniques and the performances of the model were evaluated using the F1-score.

RESULTS: The F1-score was 84.12% when the model was trained and tested on Japanese data only. In a cross-language situation, the F1-score was 74.68% when the model was trained on Japanese data and tested on French data.

CONCLUSIONS: The results are encouraging even in a cross-language situation but further evaluation is required. The integration of preprocessing techniques, in particular noise reduction, improved the results significantly.

PMID:39652158 | DOI:10.1007/s11548-024-03301-0

Categories: Literature Watch

A dual-decoder banded convolutional attention network for bone segmentation in ultrasound images

Mon, 2024-12-09 06:00

Med Phys. 2024 Dec 9. doi: 10.1002/mp.17545. Online ahead of print.

ABSTRACT

BACKGROUND: Ultrasound (US) has great potential for application in computer-assisted orthopedic surgery (CAOS) due to its non-radiative, cost-effective, and portable traits. However, bone segmentation from low-quality US images has been challenging. Traditional segmentation methods cannot achieve satisfactory results due to their high customization and dependence on bone morphology. Existing deep learning-based methods make it difficult to ensure efficient and accurate segmentation due to the ignorance of prior knowledge of bone features during feature learning.

PURPOSE: This paper aims to systematically investigate feature extraction and segmentation methodologies of bone US images and then proposes an innovative convolutional neural network to address the need for precise and efficient bone structure extraction in CAOS.

METHODS: This paper has proposed a dual-decoder banded convolutional attention network (BCA-Net), which takes the raw US image as input and simplified U-Net as the baseline network. Multiscale banded convolution kernels are employed internally in the BCA-Net model, leveraging the prior knowledge that bone surfaces in US images are exhibited as bright bands of a few millimeters in width. Additionally, a shared encoder to extract input features and two independent decoders to generate outputs for the bone surface and bone shadow mask are integrated into the BCA-Net model, leveraging the prior knowledge that US bone surfaces manifest low-intensity hollow shadows below. Then, a new task consistency loss is introduced that can utilize inter-task dependency fully and enhance the performance of our model. In the network construction process, a dataset containing 1623 sets of US images was adopted, and a five-fold cross-validation strategy was divided into the training and validation sets for the model's training and validation. Many vital metrics were introduced to comprehensively evaluate the model performance, including overlap, edge distance, area under curve, and model efficiency. Finally, the model performance was subjected to a confidence interval, Tukey's honest significant difference, and Cohen's d statistics at a significance level (5%) to ensure the accuracy and reliability of the obtained findings.

RESULTS: The experimental results show that the BCA-Net model performs well in the bone surface segmentation task. Its average Dice coefficient reaches 87.51%, 4.04% higher than U-Net's, proving its superior bone surface segmentation accuracy. Meanwhile, the average distance error is 0.2 mm, 0.33 mm lower than U-Net's, highlighting its accuracy in detail capture and boundary recognition. Using a confidence distance threshold of 1.02 mm, the Dice coefficient of the BCA-Net model exceeds 98%, an improvement of 1.87% over U-Net's, which is highly consistent with manual labeling. The BCA-Net model achieves a statistical significance of p-values < 0.05 in the above accuracy comparisons. In addition, the BCA-Net model has a small parameter count (5.58 M) and high computational efficiency (35.85 frames per second), further validating its excellent potential in bone surface segmentation tasks.

CONCLUSIONS: The proposed method achieves excellent performance with high accuracy and efficiency, aligning well with clinical requirements and holding excellent potential for advancing the utilization of US images in CAOS.

PMID:39651711 | DOI:10.1002/mp.17545

Categories: Literature Watch

A comparison of super-resolution microscopy techniques for imaging tightly packed microcolonies of an obligate intracellular bacterium

Mon, 2024-12-09 06:00

J Microsc. 2024 Dec 9. doi: 10.1111/jmi.13376. Online ahead of print.

ABSTRACT

Conventional optical microscopy imaging of obligate intracellular bacteria is hampered by the small size of bacterial cells, tight clustering exhibited by some bacterial species and challenges relating to labelling such as background from host cells, a lack of validated reagents, and a lack of tools for genetic manipulation. In this study, we imaged intracellular bacteria from the species Orientia tsutsugamushi (Ot) using five different fluorescence microscopy techniques: standard confocal, Airyscan confocal, instant Structured Illumination Microscopy (iSIM), three-dimensional Structured Illumination Microscopy (3D-SIM) and Stimulated Emission Depletion Microscopy (STED). We compared the ability of each to resolve bacterial cells in intracellular clumps in the lateral (xy) axis, using full width half-maximum (FWHM) measurements of a labelled outer membrane protein (ScaA) and the ability to detect small, outer membrane vesicles external to the cells. Comparing the techniques readily available to us (above), 3D-SIM microscopy, in combination with the shortest-wavelength dyes, was found overall to give the best lateral resolution. We next compared the ability of each technique to sufficiently resolve bacteria in the axial (z) direction and found 3D-STED to be the most successful method for this. We then combined this 3D-STED approach with a custom 3D cell segmentation and analysis pipeline using the open-source, deep learning software, Cellpose to segment the cells and subsequently the commercial software Imaris to analyse their 3D shape and size. Using this combination, we demonstrated differences in bacterial shape, but not their size, when grown in different mammalian cell lines. Overall, we compare the advantages and disadvantages of different super-resolution microscopy techniques for imaging this cytoplasmic obligate intracellular bacterium based on the specific research question being addressed.

PMID:39651611 | DOI:10.1111/jmi.13376

Categories: Literature Watch

Prognostic Modeling for Liver Cirrhosis Mortality Prediction and Real-Time Health Monitoring from Electronic Health Data

Mon, 2024-12-09 06:00

Big Data. 2024 Dec 9. doi: 10.1089/big.2024.0071. Online ahead of print.

ABSTRACT

Liver cirrhosis stands as a prominent contributor to mortality, impacting millions across the United States. Enabling health care providers to predict early mortality among patients with cirrhosis holds the potential to enhance treatment efficacy significantly. Our hypothesis centers on the correlation between mortality and laboratory test results along with relevant diagnoses in this patient cohort. Additionally, we posit that a deep learning model could surpass the predictive capabilities of the existing Model for End-Stage Liver Disease score. This research seeks to advance prognostic accuracy and refine approaches to address the critical challenges posed by cirrhosis-related mortality. This study evaluates the performance of an artificial neural network model for liver disease classification using various training dataset sizes. Through meticulous experimentation, three distinct training proportions were analyzed: 70%, 80%, and 90%. The model's efficacy was assessed using precision, recall, F1-score, accuracy, and support metrics, alongside receiver operating characteristic (ROC) and precision-recall (PR) curves. The ROC curves were quantified using the area under the curve (AUC) metric. Results indicated that the model's performance improved with an increased size of the training dataset. Specifically, the 80% training data model achieved the highest AUC, suggesting superior classification ability over the models trained with 70% and 90% data. PR analysis revealed a steep trade-off between precision and recall across all datasets, with 80% training data again demonstrating a slightly better balance. This is indicative of the challenges faced in achieving high precision with a concurrently high recall, a common issue in imbalanced datasets such as those found in medical diagnostics.

PMID:39651607 | DOI:10.1089/big.2024.0071

Categories: Literature Watch

Prediction of Visual Acuity After Cataract Surgery by Deep Learning Methods Using Clinical Information and Color Fundus Photography

Mon, 2024-12-09 06:00

Curr Eye Res. 2024 Dec 9:1-6. doi: 10.1080/02713683.2024.2430212. Online ahead of print.

ABSTRACT

PURPOSE: To examine the performance of deep-learning models that predicts the visual acuity after cataract surgery using preoperative clinical information and color fundus photography (CFP).

METHODS: We retrospectively collected the age, sex, and logMAR preoperative best corrected visual acuity (preoperative-BCVA) and CFP from patients who underwent cataract surgeries from 2020 to 2021 at National Taiwan University Hospital. Feature extraction of CFP was performed using a pre-existing image classification model, Xception. The CFP-extracted features and pre-operative clinical information were then fed to a downstream neural network for final prediction. We assessed the model performance by calculating the mean absolute error (MAE) between the predicted and the true logMAR of postoperative BCVA. A nested 10-fold cross-validation was performed for model validation.

RESULTS: A total of 673 fundus images from 446 patients were collected. The mean preoperative BCVA and postoperative BCVA was 0.60 ± 0.39 and 0.14 ± 0.18, respectively. The model using age and sex as predictors achieved an MAE of 0.121 ± 0.016 in postoperative BCVA prediction. The inclusion of CFP as additional predictor in the model (predictors: age, sex and CFP) did not further improve the predictive performance (MAE = 0.120 ± 0.023, p = 0.375), while adding the preoperative BCVA as an additional predictor resulted in a 4.13% improvement (predictors: age, sex and preoperative BCVA, MAE = 0.116 ± 0.016, p = 0.013).

CONCLUSIONS: Our multimodal models including both CFP and clinical information achieved excellent accuracy in predicting BCVA after cataract surgery, while the learning models input with only clinical information performed similarly. Future studies are needed to clarify the effects of multimodal input on this task.

PMID:39651583 | DOI:10.1080/02713683.2024.2430212

Categories: Literature Watch

Performance of large language models in the National Dental Licensing Examination in China: a comparative analysis of ChatGPT, GPT-4, and New Bing

Mon, 2024-12-09 06:00

Int J Comput Dent. 2024 Dec 9;27(4):401-411. doi: 10.3290/j.ijcd.b5870240.

ABSTRACT

AIM: The objective of the present study was to investigate the clinical understanding and reasoning abilities of large language models (LLMs); namely, ChatGPT, GPT-4, and New Bing, by evaluating their performance in the NDLE (National Dental Licensing Examination) in China.

MATERIALS AND METHODS: Questions from the NDLE from 2020 to 2022 were selected based on subject weightings. Standardized prompts were utilized to regulate the output of LLMs for acquiring more precise answers. The performance of each model across each subject category and for the subjects overall was analyzed employing the McNemar's test.

RESULTS: The percentage scores obtained by ChatGPT, GPT-4, and New Bing were 42.6% (138/324), 63.0% (204/324), and 72.5% (235/324), respectively. Significant variance was seen between the performance of New Bing compared with ChatGPT and GPT-4. GPT-4 and New Bing outperformed ChatGPT across all subjects, with New Bing surpassing GPT-4 in most subjects.

CONCLUSION: GPT-4 and New Bing exhibited promising capabilities in the NDLE. However, their performance in specific subjects such as prosthodontics and oral and maxillofacial surgery requires improvement. This performance gap can be attributed to limited dental training data and the inherent complexity of these subjects.

PMID:39651568 | DOI:10.3290/j.ijcd.b5870240

Categories: Literature Watch

Pages