Deep learning

A transformer-based deep learning model for identifying the occurrence of acute hematogenous osteomyelitis and predicting blood culture results

Thu, 2024-11-21 06:00

Front Microbiol. 2024 Nov 6;15:1495709. doi: 10.3389/fmicb.2024.1495709. eCollection 2024.

ABSTRACT

BACKGROUND: Acute hematogenous osteomyelitis is the most common form of osteomyelitis in children. In recent years, the incidence of osteomyelitis has been steadily increasing. For pediatric patients, clearly describing their symptoms can be quite challenging, which often necessitates the use of complex diagnostic methods, such as radiology. For those who have been diagnosed, the ability to culture the pathogenic bacteria significantly affects their treatment plan.

METHOD: A total of 634 patients under the age of 18 were included, and the correlation between laboratory indicators and osteomyelitis, as well as several diagnoses often confused with osteomyelitis, was analyzed. Based on this, a Transformer-based deep learning model was developed to identify osteomyelitis patients. Subsequently, the correlation between laboratory indicators and the length of hospital stay for osteomyelitis patients was examined. Finally, the correlation between the successful cultivation of pathogenic bacteria and laboratory indicators in osteomyelitis patients was analyzed, and a deep learning model was established for prediction.

RESULT: The laboratory indicators of patients are correlated with the presence of acute hematogenous osteomyelitis, and the deep learning model developed based on this correlation can effectively identify patients with acute hematogenous osteomyelitis. The laboratory indicators of patients with acute hematogenous osteomyelitis can partially reflect their length of hospital stay. Although most laboratory indicators lack a direct correlation with the ability to culture pathogenic bacteria in patients with acute hematogenous osteomyelitis, our model can still predict whether the bacteria can be successfully cultured.

CONCLUSION: Laboratory indicators, as easily accessible medical information, can identify osteomyelitis in pediatric patients. They can also predict whether pathogenic bacteria can be successfully cultured, regardless of whether the patient has received antibiotics beforehand. This not only simplifies the diagnostic process for pediatricians but also provides a basis for deciding whether to use empirical antibiotic therapy or discontinue treatment for blood cultures.

PMID:39568996 | PMC:PMC11578118 | DOI:10.3389/fmicb.2024.1495709

Categories: Literature Watch

ProBID-Net: a deep learning model for protein-protein binding interface design

Thu, 2024-11-21 06:00

Chem Sci. 2024 Oct 30. doi: 10.1039/d4sc02233e. Online ahead of print.

ABSTRACT

Protein-protein interactions are pivotal in numerous biological processes. The computational design of these interactions facilitates the creation of novel binding proteins, crucial for advancing biopharmaceutical products. With the evolution of artificial intelligence (AI), protein design tools have swiftly transitioned from scoring-function-based to AI-based models. However, many AI models for protein design are constrained by assuming complete unfamiliarity with the amino acid sequence of the input protein, a feature most suited for de novo design but posing challenges in designing protein-protein interactions when the receptor sequence is known. To bridge this gap in computational protein design, we introduce ProBID-Net. Trained using natural protein-protein complex structures and protein domain-domain interface structures, ProBID-Net can discern features from known target protein structures to design specific binding proteins based on their binding sites. In independent tests, ProBID-Net achieved interface sequence recovery rates of 52.7%, 43.9%, and 37.6%, surpassing or being on par with ProteinMPNN in binding protein design. Validated using AlphaFold-Multimer, the sequences designed by ProBID-Net demonstrated a close correspondence between the design target and the predicted structure. Moreover, the model's output can predict changes in binding affinity upon mutations in protein complexes, even in scenarios where no data on such mutations were provided during training (zero-shot prediction). In summary, the ProBID-Net model is poised to significantly advance the design of protein-protein interactions.

PMID:39568891 | PMC:PMC11575592 | DOI:10.1039/d4sc02233e

Categories: Literature Watch

Medical meteorological forecast for ischemic stroke: random forest regression vs long short-term memory model

Wed, 2024-11-20 06:00

Int J Biometeorol. 2024 Nov 21. doi: 10.1007/s00484-024-02818-y. Online ahead of print.

ABSTRACT

Ischemic stroke (IS) is one of the top risk factors for death and disability. Meteorological conditions have an effect on IS attack. In this study, we try to develop models of medical meteorological forecast for IS attack based on machine learning and deep learning algorithms. The medical meteorological forecast would be beneficial to public health in IS events prevention and treatment. We collected data on IS attacks and climatology in each day from 18th September 2016 to 31th December 2020 in Haikou. Data on IS attacks were from the number of hospital admissions due to IS attack among general population. The random forest (RF) regression and long short-term memory (LSTM) algorithms were respectively used to develop the predictive model based on meteorological data. Performance of the model was assessed by mean squared error (MSE) and root mean squared error (RMSE). A total of 42849 IS attacks was included in this study. IS attacks were significantly decreased in winter. The pattern of climatological data was observed the regularity in seasons. For the performance of RF regression model, the MSE is 243, and the RMSE is 15.6. For LSTM model, the MSE is 36, and the RMSE is 6. In conclusion, LSTM model is more accurate than RF regression model to predict IS attacks in general population based on meteorological data. LSTM model showed acceptable accuracy for the prediction and could be used as medical meteorological forecast to predict IS attack among population according to local climate.

PMID:39567379 | DOI:10.1007/s00484-024-02818-y

Categories: Literature Watch

Impact of Deep Learning-Based Computer-Aided Detection and Electronic Notification System for Pneumothorax on Time to Treatment: Clinical Implementation

Wed, 2024-11-20 06:00

J Am Coll Radiol. 2024 Nov 18:S1546-1440(24)00919-0. doi: 10.1016/j.jacr.2024.11.009. Online ahead of print.

ABSTRACT

OBJECTIVE: To assess whether the implementation of deep learning (DL) computer-aided detection (CAD) that screens for suspected pneumothorax (PTX) on chest radiography (CXR) combined with an electronic notification system (ENS) that simultaneously alerts both the radiologist and the referring clinician would affect time to treatment (TTT) in a real-world clinical practice.

METHODS: In May 2022, a commercial deep learning-based CAD and ENS (DL-CAD-ENS) was introduced for all CXRs at an 818-bed general hospital, with 33 attending doctors and their residents using ENS, while 155 others used only CAD. We used difference-in-differences estimates to compare TTT between the CAD and ENS group and the CAD-only group for the period from January 2018 to April 2022 and from May 2022 to April 2023.

RESULTS: A total of 603,028 CXRs from 140,841 unique patients were included, with a PTX prevalence of 2.0%. There was a significant reduction in TTT for supplemental oxygen therapy for the CAD and ENS group compared with the CAD-only group in the post-implementation period (-143.8 min; 95% confidence interval [CI], -277.8 to -9.9; P = .035). However, there was no significant difference in TTT for other treatments, including aspiration or tube-thoracostomy (14.4 min; 95% CI, -35.0 to 63.9) and consultation with the thoracic and cardiovascular surgery department (86.3 min; 95% CI, -175.1 to 347.6).

CONCLUSION: The introduction of a DL-CAD-ENS reduced the time to initiate oxygen supplementation for patients with PTX.

PMID:39566875 | DOI:10.1016/j.jacr.2024.11.009

Categories: Literature Watch

Blip-up blip-down circular EPI (BUDA-cEPI) for distortion-free dMRI with rapid unrolled deep learning reconstruction

Wed, 2024-11-20 06:00

Magn Reson Imaging. 2024 Nov 18:110277. doi: 10.1016/j.mri.2024.110277. Online ahead of print.

ABSTRACT

PURPOSE: BUDA-cEPI has been shown to achieve high-quality, high-resolution diffusion magnetic resonance imaging (dMRI) with fast acquisition time, particularly when used in conjunction with S-LORAKS reconstruction. However, this comes at a cost of more complex reconstruction that is computationally prohibitive. In this work we develop rapid reconstruction pipeline for BUDA-cEPI to pave the way for its deployment in routine clinical and neuroscientific applications. The proposed reconstruction includes the development of ML-based unrolled reconstruction as well as rapid ML-based B0 and eddy current estimations that are needed. The architecture of the unroll network was designed so that it can mimic S-LORAKS regularization well, with the addition of virtual coil channels.

METHODS: BUDA-cEPI RUN-UP - a model-based framework that incorporates off-resonance and eddy current effects was unrolled through an artificial neural network with only six gradient updates. The unrolled network alternates between data consistency (i.e., forward BUDA-cEPI and its adjoint) and regularization steps where U-Net plays a role as the regularizer. To handle the partial Fourier effect, the virtual coil concept was also introduced into the reconstruction to effectively take advantage of the smooth phase prior and trained to predict the ground-truth images obtained by BUDA-cEPI with S-LORAKS.

RESULTS: The introduction of the Virtual Coil concept into the unrolled network was shown to be key to achieving high-quality reconstruction for BUDA-cEPI. With the inclusion of an additional non-diffusion image (b-value = 0 s/mm2), a slight improvement was observed, with the normalized root mean square error further reduced by approximately 5 %. The reconstruction times for S-LORAKS and the proposed unrolled networks were approximately 225 and 3 s per slice, respectively.

CONCLUSION: BUDA-cEPI RUN-UP was shown to reduce the reconstruction time by ~88× when compared to the state-of-the-art technique, while preserving imaging details as demonstrated through DTI application.

PMID:39566835 | DOI:10.1016/j.mri.2024.110277

Categories: Literature Watch

Applications of artificial intelligence to inherited retinal diseases: A systematic review

Wed, 2024-11-20 06:00

Surv Ophthalmol. 2024 Nov 18:S0039-6257(24)00139-5. doi: 10.1016/j.survophthal.2024.11.007. Online ahead of print.

ABSTRACT

Artificial intelligence (AI)-based methods have been extensively used for the detection and management of various common retinal conditions, but their targeted development for inherited retinal diseases (IRD) is still nascent. In the context of limited availability of retinal subspecialists, genetic testing and genetic counseling, there is a high need for accurate and accessible diagnostic methods. The currently available AI studies, aiming for detection, classification, and prediction of IRD, remain mainly retrospective and include relatively limited numbers of patients due to their scarcity. We summarize the latest findings and clinical implications of machine-learning algorithms in IRD, highlighting the achievements and challenges of AI to assist ophthalmologists in their clinical practice.

PMID:39566565 | DOI:10.1016/j.survophthal.2024.11.007

Categories: Literature Watch

Cultivating diagnostic clarity: The importance of reporting artificial intelligence confidence levels in radiologic diagnoses

Wed, 2024-11-20 06:00

Clin Imaging. 2024 Nov 13;117:110356. doi: 10.1016/j.clinimag.2024.110356. Online ahead of print.

ABSTRACT

Accurate image interpretation is essential in the field of radiology to the healthcare team in order to provide optimal patient care. This article discusses the use of artificial intelligence (AI) confidence levels to enhance the accuracy and dependability of its radiological diagnoses. The current advances in AI technologies have changed how radiologists and clinicians make the diagnoses of pathological conditions such as aneurysms, hemorrhages, pneumothorax, pneumoperitoneum, and particularly fractures. To enhance the utility of these AI models, radiologists need a more comprehensive understanding of the model's levels of confidence and certainty behind the results they produce. This allows radiologists to make more informed decisions that have the potential to drastically change a patient's clinical management. Several AI models, especially those utilizing deep learning models (DL) with convolutional neural networks (CNNs), have demonstrated significant potential in identifying subtle findings in medical imaging that are often missed by radiologists. It is necessary to create standardized levels of confidence metrics in order for AI systems to be relevant and reliable in the clinical setting. Incorporating AI into clinical practice does have certain obstacles like the need for clinical validation, concerns regarding the interpretability of AI system results, and addressing confusion and misunderstandings within the medical community. This study emphasizes the importance of AI systems to clearly convey their level of confidence in radiological diagnosis. This paper highlights the importance of conducting research to establish AI confidence level metrics that are limited to a specific anatomical region or lesion type. KEY POINT OF THE VIEW: Accurate fracture diagnosis relies on radiologic certainty, where Artificial intelligence (AI), especially convolutional neural networks (CNNs) and deep learning (DL), shows promise in enhancing X-ray interpretation amidst a shortage of radiologists. Overcoming integration challenges through improved AI interpretability and education is crucial for widespread acceptance and better patient outcomes.

PMID:39566394 | DOI:10.1016/j.clinimag.2024.110356

Categories: Literature Watch

Long duration forecasting and its performance capability for seasonal variation modelling of residual chlorine concentrations: A comparative evaluation of two small-scale water distribution systems in Japan

Wed, 2024-11-20 06:00

Water Res. 2024 Nov 8;268(Pt B):122766. doi: 10.1016/j.watres.2024.122766. Online ahead of print.

ABSTRACT

Chlorine dosing control is a critical task for health security, complying with drinking water quality standards while achieving customer satisfaction. In Japan, this became more difficult due to decreasing number of technical personnel as a result of declining population and huge retirement of veterans. As experienced-based dosing control still exists and is considered inefficient due to risks of inaccurate dosing and need of experienced manpower, streamlining of operations is needed. This study aims to address these concerns through the development and evaluation of a deep learning model for residual chlorine concentrations capable of forecasting long durations. In this paper, the model is investigated in seasonal timeframe analyzing trends and variations. Two water distributions systems of varying network-simple and complex, are also compared to further analyze model performance and versatility. The model utilizes the past 24 h of flow rate, chlorine level at treatment plant, water temperature, and residual chlorine at private homes as input to predict the 12 h ahead of residual chlorine at private homes evaluated at hourly training lengths of 0.5, 1, 1.5, 2 years. Results revealed that the model achieved high accuracy in predicting hourly residual chlorine with general increasing model error from winter, spring, autumn, and summer due to the progressing instabilities in chlorine concentrations from low to high temperatures. Moreover, smaller system tends to be more unstable incurring lower model performance relative to larger system. In terms of optimal training length, ≥1Y training models are found to have lesser chance of data drift occurrence prospectively reducing model retraining frequency in the future.

PMID:39566282 | DOI:10.1016/j.watres.2024.122766

Categories: Literature Watch

HKA-Net: clinically-adapted deep learning for automated measurement of hip-knee-ankle angle on lower limb radiography for knee osteoarthritis assessment

Wed, 2024-11-20 06:00

J Orthop Surg Res. 2024 Nov 20;19(1):777. doi: 10.1186/s13018-024-05265-y.

ABSTRACT

BACKGROUND: Accurate measurement of the hip-knee-ankle (HKA) angle is essential for informed clinical decision-making in the management of knee osteoarthritis (OA). Knee OA is commonly associated with varus deformity, where the alignment of the knee shifts medially, leading to increased stress and deterioration of the medial compartment. The HKA angle, which quantifies this alignment, is a critical indicator of the severity of varus deformity and helps guide treatment strategies, including corrective surgeries. Current manual methods are labor-intensive, time-consuming, and prone to inter-observer variability. Developing an automated model for HKA angle measurement is challenging due to the elaborate process of generating handcrafted anatomical landmarks, which is more labor-intensive than the actual measurement. This study aims to develop a ResNet-based deep learning model that predicts the HKA angle without requiring explicit anatomical landmark annotations and to assess its accuracy and efficiency compared to conventional manual methods.

METHODS: We developed a deep learning model based on the variants of the ResNet architecture to process lower limb radiographs and predict HKA angles without explicit landmark annotations. The classification performance for the four stages of varus deformity (stage I: 0°-10°, stage II: 10°-20°, stage III: > 20°, others: genu valgum or normal alignment) was also evaluated. The model was trained and validated using a retrospective cohort of 300 knee OA patients (Kellgren-Lawrence grade 3 or higher), with horizontal flip augmentation applied to double the dataset to 600 samples, followed by fivefold cross-validation. An extended temporal validation was conducted on a separate cohort of 50 knee OA patients. The model's accuracy was assessed by calculating the mean absolute error between predicted and actual HKA angles. Additionally, the classification of varus deformity stages was conducted to evaluate the model's ability to provide clinically relevant categorizations. Time efficiency was compared between the automated model and manual measurements performed by an experienced orthopedic surgeon.

RESULTS: The ResNet-50 model achieved a bias of - 0.025° with a standard deviation of 1.422° in the retrospective cohort and a bias of - 0.008° with a standard deviation of 1.677° in the temporal validation cohort. Using the ResNet-152 model, it accurately classified the four stages of varus deformity with weighted F1-score of 0.878 and 0.859 in the retrospective and temporal validation cohorts, respectively. The automated model was 126.7 times faster than manual measurements, reducing the total time from 49.8 min to 23.6 sec for the temporal validation cohort.

CONCLUSIONS: The proposed ResNet-based model provides an efficient and accurate method for measuring HKA angles and classifying varus deformity stages without the need for extensive landmark annotations. Its high accuracy and significant improvement in time efficiency make it a valuable tool for clinical practice, potentially enhancing decision-making and workflow efficiency in the management of knee OA.

PMID:39568048 | DOI:10.1186/s13018-024-05265-y

Categories: Literature Watch

A polynomial proxy model approach to verifiable decentralized federated learning

Wed, 2024-11-20 06:00

Sci Rep. 2024 Nov 20;14(1):28786. doi: 10.1038/s41598-024-79798-x.

ABSTRACT

Decentralized Federated Learning improves data privacy and eliminates single points of failure by removing reliance on centralized storage and model aggregation in distributed computing systems. Ensuring the integrity of computations during local model training is a significant challenge, especially before sharing gradient updates from each local client. Current methods for ensuring computation integrity often involve patching local models to implement cryptographic techniques, such as Zero-Knowledge Proofs. However, this approach becomes highly complex and sometimes impractical for large-scale models that use techniques such as random dropouts to improve training convergence. These random dropouts create non-deterministic behavior, making it challenging to verify model updates under deterministic protocols. We propose ProxyZKP, a novel framework combining Zero-Knowledge Proofs with polynomial proxy models to provide computation integrity in local training to address this issue. Each local node combines a private model for online deep learning applications and a proxy model that mediates decentralized model training by exchanging gradient updates. The multivariate polynomial nature of proxy models facilitates the application of Zero-Knowledge Proofs. These proofs verify the computation integrity of updates from each node without disclosing private data. Experimental results indicate that ProxyZKP significantly reduces computational load. Specifically, ProxyZKP achieves proof generation times that are 30-50% faster compared to established methods like zk-SNARKs and Bulletproofs. This improvement is largely due to the high parallelization potential of the univariate polynomial decomposition approach. Additionally, integrating Differential Privacy into the ProxyZKP framework reduces the risk of Gradient Inversion attacks by adding calibrated noise to the gradients, while maintaining competitive model accuracy. The results demonstrate that ProxyZKP is a scalable and efficient solution for ensuring training integrity in decentralized federated learning environments, particularly in scenarios with frequent model updates and the need for strong model scalability.

PMID:39567642 | DOI:10.1038/s41598-024-79798-x

Categories: Literature Watch

Chisel plow characteristics impact on power-fuel consumption, stubble cover, and surface roughness using deep learning neural networks with sensitivity analysis

Wed, 2024-11-20 06:00

Sci Rep. 2024 Nov 20;14(1):28804. doi: 10.1038/s41598-024-80253-0.

ABSTRACT

The aim of this study was to determine the effects of different shank types, tine types, and tractor forward speeds on the power and fuel consumption of chisel plows, stubble cover, soil surface roughness, and penetration resistance. In addition, using the obtained results, the draft power, fuel consumption, soil surface roughness, and soil stubble cover rate were modeled using Deep Learning Neural Network (DLNN) architectures and sensitivity analysis of these models were performed. Four different shank types (rigid, semi-spring, spring, and vibrating) and two different tine types (conventional and winged) were used at three different tractor forward speeds (4.5, 5.4 and 6.3 km.h- 1) were tested to this end. The obtained results indicated that the maximum draft power was achieved with the rigid type shank. The highest soil surface roughness values were observed for the vibrating shank type and winged tine type. Sixteen different network architectures were conducted using deep learning neural network methods and draft power, fuel consumption, soil surface roughness, and percent stubble cover were modeled. Sensitivity analyses were performed to indicate which modeled parameters were more sensitive to the factors using the obtained models. Draft power was modeled with 97.7% accuracy using the DLNN9 network architecture. Additionally, fuel consumption and soil roughness were best modeled with the DLNN13 network architecture, R2 values for those targets were 0.929 and 0.930 respectively. According to the sensitivity analysis, draft power, fuel consumption, soil roughness, and stubble cover rate were significantly affected by changes in the physical properties of the soil.

PMID:39567614 | DOI:10.1038/s41598-024-80253-0

Categories: Literature Watch

Kangba Region of Sichuan based on swin transformer visual model research on the identification of facades of ethnic buildings

Wed, 2024-11-20 06:00

Sci Rep. 2024 Nov 20;14(1):28742. doi: 10.1038/s41598-024-78774-9.

ABSTRACT

The protection and restoration of existing buildings requires accurate acquisition of the characteristics of the building facade. The complex, diverse, and irregular distribution characteristics of the building facade components of ethnic minorities have led to a huge workload of field research, surveying, mapping, and calculation, and it is more difficult to extract its facade characteristics accurately. This study proposes a visual model based on the Swin Transformer and applies it to the graphic recognition of ethnic building elevations. The model combines the advantages of the migration learning method and deep neural network technology and is further enriched by layer normalization to improve the stability and extraction ability of model training. In the field survey of ethnic minority buildings in Kangba, Sichuan, 1100 images of local buildings were collected, including 8 different types of ethnic minority buildings. The experimental results show that compared with other mainstream deep neural network models, the Swin Transformer visual model shows excellent predictive performance to prove the effectiveness of the proposed method. This study also uses the t-sne dimension reduction method to verify the feature extraction ability of the Swin Transformer, which contributes to the protection and restoration of ethnic minority buildings, active exploration of energy conservation, digital archiving, and more. Provide theoretical and practical reference in the fields of architectural style and cultural research.

PMID:39567608 | DOI:10.1038/s41598-024-78774-9

Categories: Literature Watch

Tapping line detection and rubber tapping pose estimation for natural rubber trees based on improved YOLOv8 and RGB-D information fusion

Wed, 2024-11-20 06:00

Sci Rep. 2024 Nov 20;14(1):28717. doi: 10.1038/s41598-024-79132-5.

ABSTRACT

Tapping line detection and rubber tapping pose estimation are challenging tasks in rubber plantation environments for rubber tapping robots. This study proposed a method for tapping line detection and rubber tapping pose estimation based on improved YOLOv8 and RGB-D information fusion. Firstly, YOLOv8n was improved by introducing the CFB module into the backbone, adding an output layer into the neck, fusing the EMA attention mechanism into the neck, and modifying the loss function as NWD to realize multi-object detection and segmentation. Secondly, the trunk skeleton line was extracted by combining level set and ellipse fitting. Then, the new tapping line was located by combining edge detection and geometric analysis. Finally, the rubber tapping pose was estimated based on the trunk skeleton line and the new tapping line. The detection results from 597 test images showed the improved YOLOv8n's detection mAP0.5, segmentation mAP0.5, and model size were 81.9%, 72.9%, and 6.06 MB, respectively. The improved YOLOv8n's effect and efficiency were superior compared to other networks, and it could better detect and segment natural rubber tree image targets in different scenes. The pose estimation results from 300 new tapping lines showed the average success rate and average time consumed for rubber tapping pose estimation were 96% and 0.2818 s, respectively. The positioning errors in x, y, and z directions were 0.69 ± 0.51 mm, 0.73 ± 0.4 mm, and 1.07 ± 0.56 mm, respectively. The error angles in a, o, and n directions were 1.65° ± 0.68°, 2.53° ± 0.88°, and 2.26° ± 0.89°, respectively. Therefore, this method offers an effective solution for rubber tapping pose estimation and provides theoretical support for the development of rubber tapping robots.

PMID:39567603 | DOI:10.1038/s41598-024-79132-5

Categories: Literature Watch

A continuous pursuit dataset for online deep learning-based EEG brain-computer interface

Wed, 2024-11-20 06:00

Sci Data. 2024 Nov 20;11(1):1256. doi: 10.1038/s41597-024-04090-6.

ABSTRACT

This dataset is from an EEG brain-computer interface (BCI) study investigating the use of deep learning (DL) for online continuous pursuit (CP) BCI. In this task, subjects use Motor Imagery (MI) to control a cursor to follow a randomly moving target, instead of a single stationary target used in other traditional BCI tasks. DL methods have recently achieved promising performance in traditional BCI tasks, but most studies investigate offline data analysis using DL algorithms. This dataset consists of ~168 hours of EEG recordings from complex CP BCI experiments, collected from 28 unique human subjects over multiple sessions each, with an online DL-based decoder. The large amount of subject specific data from multiple sessions may be useful for developing new BCI decoders, especially DL methods that require large amounts of training data. By providing this dataset to the public, we hope to help facilitate the development of new or improved BCI decoding algorithms for the complex CP paradigm for continuous object control, bringing EEG-based BCIs closer to real-world applications.

PMID:39567538 | DOI:10.1038/s41597-024-04090-6

Categories: Literature Watch

Probing the evolution of fault properties during the seismic cycle with deep learning

Wed, 2024-11-20 06:00

Nat Commun. 2024 Nov 20;15(1):10025. doi: 10.1038/s41467-024-54153-w.

ABSTRACT

We use seismic waves that pass through the hypocentral region of the 2016 M6.5 Norcia earthquake together with Deep Learning (DL) to distinguish between foreshocks, aftershocks and time-to-failure (TTF). Binary and N-class models defined by TTF correctly identify seismograms in test with > 90% accuracy. We use raw seismic records as input to a 7 layer CNN model to perform the classification. Here we show that DL models successfully distinguish seismic waves pre/post mainshock in accord with lab and theoretical expectations of progressive changes in crack density prior to abrupt change at failure and gradual postseismic recovery. Performance is lower for band-pass filtered seismograms (below 10 Hz) suggesting that DL models learn from the evolution of subtle changes in elastic wave attenuation. Tests to verify that our results indeed provide a proxy for fault properties included DL models trained with the wrong mainshock time and those using seismic waves far from the Norcia mainshock; both show degraded performance. Our results demonstrate that DL models have the potential to track the evolution of fault zone properties during the seismic cycle. If this result is generalizable it could improve earthquake early warning and seismic hazard analysis.

PMID:39567514 | DOI:10.1038/s41467-024-54153-w

Categories: Literature Watch

multiDGD: A versatile deep generative model for multi-omics data

Wed, 2024-11-20 06:00

Nat Commun. 2024 Nov 20;15(1):10031. doi: 10.1038/s41467-024-53340-z.

ABSTRACT

Recent technological advancements in single-cell genomics have enabled joint profiling of gene expression and alternative modalities at unprecedented scale. Consequently, the complexity of multi-omics data sets is increasing massively. Existing models for multi-modal data are typically limited in functionality or scalability, making data integration and downstream analysis cumbersome. We present multiDGD, a scalable deep generative model providing a probabilistic framework to learn shared representations of transcriptome and chromatin accessibility. It shows outstanding performance on data reconstruction without feature selection. We demonstrate on several data sets from human and mouse that multiDGD learns well-clustered joint representations. We further find that probabilistic modeling of sample covariates enables post-hoc data integration without the need for fine-tuning. Additionally, we show that multiDGD can detect statistical associations between genes and regulatory regions conditioned on the learned representations. multiDGD is available as an scverse-compatible package on GitHub.

PMID:39567490 | DOI:10.1038/s41467-024-53340-z

Categories: Literature Watch

Automatic detection and proximity quantification of inferior alveolar nerve and mandibular third molar on cone-beam computed tomography

Wed, 2024-11-20 06:00

Clin Oral Investig. 2024 Nov 20;28(12):648. doi: 10.1007/s00784-024-05967-x.

ABSTRACT

OBJECTIVES: During mandibular third molar (MTM) extraction surgery, preoperative analysis to quantify the proximity of the MTM to the surrounding inferior alveolar nerve (IAN) is essential to minimize the risk of IAN injury. This study aims to propose an automated tool to quantitatively measure the proximity of IAN and MTM in cone-beam computed tomography (CBCT) images.

MATERIALS AND METHODS: Using the dataset including 302 CBCT scans with 546 MTMs, a deep-learning-based network was developed to support the automatic detection of the IAN, MTM, and intersection region IR. To ensure accurate proximity detection, a distance detection algorithm and a volume measurement algorithm were also developed.

RESULTS: The deep learning-based model showed encouraging segmentation accuracy of the target structures (Dice similarity coefficient: 0.9531 ± 0.0145, IAN; 0.9832 ± 0.0055, MTM; 0.8336 ± 0.0746, IR). In addition, with the application of the developed algorithms, the distance between the IAN and MTM and the volume of the IR could be equivalently detected (90% confidence interval (CI): - 0.0345-0.0014 mm, distance; - 0.0155-0.0759 mm3, volume). The total time for the IAN, MTM, and IR segmentation was 2.96 ± 0.11 s, while the accurate manual segmentation required 39.01 ± 5.89 min.

CONCLUSIONS: This study presented a novel, fast, and accurate model for the detection and proximity quantification of the IAN and MTM on CBCT.

CLINICAL RELEVANCE: This model illustrates that a deep learning network may assist surgeons in evaluating the risk of MTM extraction surgery by detecting the proximity of the IAN and MTM at a quantitative level that was previously unparalleled.

PMID:39567447 | DOI:10.1007/s00784-024-05967-x

Categories: Literature Watch

Burnout crisis in Chinese radiology: will artificial intelligence help?

Wed, 2024-11-20 06:00

Eur Radiol. 2024 Nov 20. doi: 10.1007/s00330-024-11206-4. Online ahead of print.

ABSTRACT

OBJECTIVES: To assess the correlation between the use of artificial intelligence (AI) software and burnout in the radiology departments of hospitals in China.

METHODS: This study employed a cross-sectional research design. From February to July 2024, an online survey was conducted among radiologists and technicians at 68 public hospitals in China. The survey utilized general information questionnaires, the Maslach Burnout Inventory-Human Services Survey (MBI-HSS) scale, and a custom-designed AI usage questionnaire. This study analyzed the correlation between AI software usage and occupational burnout, and general information was included as a control variable in a multiple linear regression analysis.

RESULTS: The analysis of survey data from 522 radiology staff revealed that 389 (74.5%) had used AI and that 252 (48.3%) had used it for more than 12 months. Only 133 (25.5%) had not yet adopted AI. Among the respondents, radiologists had a higher AI usage rate (82.0%) than technicians (only 59.9%). Furthermore, 344 (65.9%) of the respondents exhibited signs of burnout. The duration of AI software usage was significantly negatively correlated with overall burnout, yielding a Pearson correlation coefficient of -0.112 (p < 0.05). Multiple stepwise regression analysis revealed that salary satisfaction, night shifts, duration of AI usage, weekly working hours, having children, and professional rank were the main factors influencing occupational burnout (all p < 0.05).

CONCLUSION: AI has the potential to significantly help mitigate occupational burnout among radiology staff. This study reveals the key role that AI plays in assisting radiology staff in their work.

KEY POINTS: Questions Although we are aware that radiology staff burnout is intensifying, there is no quantitative research assessing whether artificial intelligence software can mitigate this occupational burnout. Findings The longer staff use deep learning-based artificial intelligence imaging software, the less severe their occupational burnout tends to be. This result is particularly evident among radiologists. Clinical relevance In China, radiologists and technicians experience high burnout rates. Even if there is an artificial intelligence usage controversy, encouraging the use of artificial intelligence software in radiology helps prevent and alleviate this occupational burnout.

PMID:39567429 | DOI:10.1007/s00330-024-11206-4

Categories: Literature Watch

Digital image processing to detect adaptive evolution

Wed, 2024-11-20 06:00

Mol Biol Evol. 2024 Nov 20:msae242. doi: 10.1093/molbev/msae242. Online ahead of print.

ABSTRACT

In recent years, advances in image processing and machine learning have fueled a paradigm shift in detecting genomic regions under natural selection. Early machine learning techniques employed population-genetic summary statistics as features, which focus on specific genomic patterns expected by adaptive and neutral processes. Though such engineered features are important when training data is limited, the ease at which simulated data can now be generated has led to the recent development of approaches that take in image representations of haplotype alignments and automatically extract important features using convolutional neural networks. Digital image processing methods termed α-molecules are a class of techniques for multi-scale representation of objects that can extract a diverse set of features from images. One such α-molecule method, termed wavelet decomposition, lends greater control over high-frequency components of images. Another α-molecule method, termed curvelet decomposition, is an extension of the wavelet concept that considers events occurring along curves within images. We show that application of these α-molecule techniques to extract features from image representations of haplotype alignments yield high true positive rate and accuracy to detect hard and soft selective sweep signatures from genomic data with both linear and nonlinear machine learning classifiers. Moreover, we find that such models are easy to visualize and interpret, with performance rivaling those of contemporary deep learning approaches for detecting sweeps.

PMID:39565932 | DOI:10.1093/molbev/msae242

Categories: Literature Watch

Enhancing Gout Diagnosis with Deep Learning in Dual-energy Computed Tomography: A Retrospective Analysis of Crystal and Artifact Differentiation

Wed, 2024-11-20 06:00

Rheumatology (Oxford). 2024 Nov 20:keae523. doi: 10.1093/rheumatology/keae523. Online ahead of print.

ABSTRACT

OBJECTIVES: To evaluate whether the application of deep learning (DL) could achieve high diagnostic accuracy in differentiating between green colour coding, indicative of tophi, and clumpy artifacts observed in dual-energy computed tomography (DECT) scans.

METHODS: A comprehensive analysis of 18 704 regions of interest (ROIs) extracted from green foci in DECT scans obtained from 47 patients with gout and 27 gout-free controls was performed. The ROIs were categorized into three size groups: small, medium, and large. Convolutional neural network (CNN) analysis on a per-lesion basis and support vector machine (SVM) analysis on a per-patient basis were performed. The area under the receiver operating characteristic curve, sensitivity, specificity, positive predictive value, and negative predictive value of the models were compared.

RESULTS: For small ROIs, the sensitivity and specificity of the CNN model were 81.5% and 96.1%, respectively; for medium ROIs, 82.7% and 96.1%, respectively; for large ROIs, 91.8% and 86.9%, respectively. Additionally, the DL algorithm exhibited accuracies of 88.5%, 88.6%, and 91.0% for small, medium, and large ROIs, respectively. In the per-patient analysis, the SVM approach demonstrated a sensitivity of 87.2%, a specificity of 100%, and an accuracy of 91.8% in distinguishing between patients with gout and gout-free controls.

CONCLUSION: Our study demonstrates the effectiveness of the DL algorithm in differentiating between green colour coding indicative of crystal deposition and clumpy artifacts in DECT scans. With high sensitivity, specificity, and accuracy, the utilization of DL in DECT for diagnosing gout enables precise lesion classification, facilitating early-stage diagnosis and promoting timely intervention approaches.

PMID:39565918 | DOI:10.1093/rheumatology/keae523

Categories: Literature Watch

Pages