Deep learning
Real-time sports injury monitoring system based on the deep learning algorithm
BMC Med Imaging. 2024 May 24;24(1):122. doi: 10.1186/s12880-024-01304-6.
ABSTRACT
In response to the low real-time performance and accuracy of traditional sports injury monitoring, this article conducts research on a real-time injury monitoring system using the SVM model as an example. Video detection is performed to capture human movements, followed by human joint detection. Polynomial fitting analysis is used to extract joint motion patterns, and the average of training data is calculated as a reference point. The raw data is then normalized to adjust position and direction, and dimensionality reduction is achieved through singular value decomposition to enhance processing efficiency and model training speed. A support vector machine classifier is used to classify and identify the processed data. The experimental section monitors sports injuries and investigates the accuracy of the system's monitoring. Compared to mainstream models such as Random Forest and Naive Bayes, the SVM utilized demonstrates good performance in accuracy, sensitivity, and specificity, reaching 94.2%, 92.5%, and 96.0% respectively.
PMID:38789963 | DOI:10.1186/s12880-024-01304-6
A deep learning-based radiomics model for predicting lymph node status from lung adenocarcinoma
BMC Med Imaging. 2024 May 24;24(1):121. doi: 10.1186/s12880-024-01300-w.
ABSTRACT
OBJECTIVES: At present, there are many limitations in the evaluation of lymph node metastasis of lung adenocarcinoma. Currently, there is a demand for a safe and accurate method to predict lymph node metastasis of lung cancer. In this study, radiomics was used to accurately predict the lymph node status of lung adenocarcinoma patients based on contrast-enhanced CT.
METHODS: A total of 503 cases that fulfilled the analysis requirements were gathered from two distinct hospitals. Among these, 287 patients exhibited lymph node metastasis (LNM +) while 216 patients were confirmed to be without lymph node metastasis (LNM-). Using both traditional and deep learning methods, 22,318 features were extracted from the segmented images of each patient's enhanced CT. Then, the spearman test and the least absolute shrinkage and selection operator were used to effectively reduce the dimension of the feature data, enabling us to focus on the most pertinent features and enhance the overall analysis. Finally, the classification model of lung adenocarcinoma lymph node metastasis was constructed by machine learning algorithm. The Accuracy, AUC, Specificity, Precision, Recall and F1 were used to evaluate the efficiency of the model.
RESULTS: By incorporating a comprehensively selected set of features, the extreme gradient boosting method (XGBoost) effectively distinguished the status of lymph nodes in patients with lung adenocarcinoma. The Accuracy, AUC, Specificity, Precision, Recall and F1 of the prediction model performance on the external test set were 0.765, 0.845, 0.705, 0.784, 0.811 and 0.797, respectively. Moreover, the decision curve analysis, calibration curve and confusion matrix of the model on the external test set all indicated the stability and accuracy of the model.
CONCLUSIONS: Leveraging enhanced CT images, our study introduces a noninvasive classification prediction model based on the extreme gradient boosting method. This approach exhibits remarkable precision in identifying the lymph node status of lung adenocarcinoma patients, offering a safe and accurate alternative to invasive procedures. By providing clinicians with a reliable tool for diagnosing and assessing disease progression, our method holds the potential to significantly improve patient outcomes and enhance the overall quality of clinical practice.
PMID:38789936 | DOI:10.1186/s12880-024-01300-w
VER-Net: a hybrid transfer learning model for lung cancer detection using CT scan images
BMC Med Imaging. 2024 May 24;24(1):120. doi: 10.1186/s12880-024-01238-z.
ABSTRACT
BACKGROUND: Lung cancer is the second most common cancer worldwide, with over two million new cases per year. Early identification would allow healthcare practitioners to handle it more effectively. The advancement of computer-aided detection systems significantly impacted clinical analysis and decision-making on human disease. Towards this, machine learning and deep learning techniques are successfully being applied. Due to several advantages, transfer learning has become popular for disease detection based on image data.
METHODS: In this work, we build a novel transfer learning model (VER-Net) by stacking three different transfer learning models to detect lung cancer using lung CT scan images. The model is trained to map the CT scan images with four lung cancer classes. Various measures, such as image preprocessing, data augmentation, and hyperparameter tuning, are taken to improve the efficacy of VER-Net. All the models are trained and evaluated using multiclass classifications chest CT images.
RESULTS: The experimental results confirm that VER-Net outperformed the other eight transfer learning models compared with. VER-Net scored 91%, 92%, 91%, and 91.3% when tested for accuracy, precision, recall, and F1-score, respectively. Compared to the state-of-the-art, VER-Net has better accuracy.
CONCLUSION: VER-Net is not only effectively used for lung cancer detection but may also be useful for other diseases for which CT scan images are available.
PMID:38789925 | DOI:10.1186/s12880-024-01238-z
Fully automatic quantification for hand synovitis in rheumatoid arthritis using pixel-classification-based segmentation network in DCE-MRI
Jpn J Radiol. 2024 May 24. doi: 10.1007/s11604-024-01592-6. Online ahead of print.
ABSTRACT
PURPOSE: A classification-based segmentation method is proposed to quantify synovium in rheumatoid arthritis (RA) patients using a deep learning (DL) method based on time-intensity curve (TIC) analysis in dynamic contrast-enhanced magnetic resonance imaging (DCE-MRI).
MATERIALS AND METHODS: This retrospective study analyzed a hand MR dataset of 28 RA patients (six males, mean age 53.7 years). A researcher, under expert guidance, used in-house software to delineate regions of interest (ROIs) for hand muscles, bones, and synovitis, generating a dataset with 27,255 pixels with corresponding TICs (muscle: 11,413, bone: 8502, synovitis: 7340). One experienced musculoskeletal radiologist performed ground truth segmentation of enhanced pannus in the joint bounding box on the 10th DCE phase, or around 5 min after contrast injection. Data preprocessing included median filtering for noise reduction, phase-only correlation algorithm for motion correction, and contrast-limited adaptive histogram equalization for improved image contrast and noise suppression. TIC intensity values were normalized using zero-mean normalization. A DL model with dilated causal convolution and SELU activation function was developed for enhanced pannus segmentation, tested using leave-one-out cross-validation.
RESULTS: 407 joint bounding boxes were manually segmented, with 129 synovitis masks. On the pixel-based level, the DL model achieved sensitivity of 85%, specificity of 98%, accuracy of 99% and precision of 84% for enhanced pannus segmentation, with a mean Dice score of 0.73. The false-positive rate for predicting cases without synovitis was 0.8%. DL-measured enhanced pannus volume strongly correlated with ground truth at both pixel-based (r = 0.87, p < 0.001) and patient-based levels (r = 0.84, p < 0.001). Bland-Altman analysis showed the mean difference for hand joints at the pixel-based and patient-based levels were -9.46 mm3 and -50.87 mm3, respectively.
CONCLUSION: Our DL-based DCE-MRI TIC shape analysis has the potential for automatic segmentation and quantification of enhanced synovium in the hands of RA patients.
PMID:38789911 | DOI:10.1007/s11604-024-01592-6
Skip connection information enhancement network for retinal vessel segmentation
Med Biol Eng Comput. 2024 May 25. doi: 10.1007/s11517-024-03108-w. Online ahead of print.
ABSTRACT
Many major diseases of the retina often show symptoms of lesions in the fundus of the eye. The extraction of blood vessels from retinal fundus images is essential to assist doctors. Some of the existing methods do not fully extract the detailed features of retinal images or lose some information, making it difficult to accurately segment capillaries located at the edges of the images. In this paper, we propose a multi-scale retinal vessel segmentation network (SCIE_Net) based on skip connection information enhancement. Firstly, the network processes retinal images at multiple scales to achieve network capture of features at different scales. Secondly, the feature aggregation module is proposed to aggregate the rich information of the shallow network. Further, the skip connection information enhancement module is proposed to take into account the detailed features of the shallow layer and the advanced features of the deeper network to avoid the problem of incomplete information interaction between the layers of the network. Finally, SCIE_Net achieves better vessel segmentation performance and results on the publicly available retinal image standard datasets DRIVE, CHASE_DB1, and STARE.
PMID:38789838 | DOI:10.1007/s11517-024-03108-w
Prediction of recurrence risk in endometrial cancer with multimodal deep learning
Nat Med. 2024 May 24. doi: 10.1038/s41591-024-02993-w. Online ahead of print.
ABSTRACT
Predicting distant recurrence of endometrial cancer (EC) is crucial for personalized adjuvant treatment. The current gold standard of combined pathological and molecular profiling is costly, hampering implementation. Here we developed HECTOR (histopathology-based endometrial cancer tailored outcome risk), a multimodal deep learning prognostic model using hematoxylin and eosin-stained, whole-slide images and tumor stage as input, on 2,072 patients from eight EC cohorts including the PORTEC-1/-2/-3 randomized trials. HECTOR demonstrated C-indices in internal (n = 353) and two external (n = 160 and n = 151) test sets of 0.789, 0.828 and 0.815, respectively, outperforming the current gold standard, and identified patients with markedly different outcomes (10-year distant recurrence-free probabilities of 97.0%, 77.7% and 58.1% for HECTOR low-, intermediate- and high-risk groups, respectively, by Kaplan-Meier analysis). HECTOR also predicted adjuvant chemotherapy benefit better than current methods. Morphological and genomic feature extraction identified correlates of HECTOR risk groups, some with therapeutic potential. HECTOR improves on the current gold standard and may help delivery of personalized treatment in EC.
PMID:38789645 | DOI:10.1038/s41591-024-02993-w
Surgical optomics: hyperspectral imaging and deep learning towards precision intraoperative automatic tissue recognition-results from the EX-MACHYNA trial
Surg Endosc. 2024 May 24. doi: 10.1007/s00464-024-10880-1. Online ahead of print.
ABSTRACT
BACKGROUND: Hyperspectral imaging (HSI), combined with machine learning, can help to identify characteristic tissue signatures enabling automatic tissue recognition during surgery. This study aims to develop the first HSI-based automatic abdominal tissue recognition with human data in a prospective bi-center setting.
METHODS: Data were collected from patients undergoing elective open abdominal surgery at two international tertiary referral hospitals from September 2020 to June 2021. HS images were captured at various time points throughout the surgical procedure. Resulting RGB images were annotated with 13 distinct organ labels. Convolutional Neural Networks (CNNs) were employed for the analysis, with both external and internal validation settings utilized.
RESULTS: A total of 169 patients were included, 73 (43.2%) from Strasbourg and 96 (56.8%) from Verona. The internal validation within centers combined patients from both centers into a single cohort, randomly allocated to the training (127 patients, 75.1%, 585 images) and test sets (42 patients, 24.9%, 181 images). This validation setting showed the best performance. The highest true positive rate was achieved for the skin (100%) and the liver (97%). Misclassifications included tissues with a similar embryological origin (omentum and mesentery: 32%) or with overlaying boundaries (liver and hepatic ligament: 22%). The median DICE score for ten tissue classes exceeded 80%.
CONCLUSION: To improve automatic surgical scene segmentation and to drive clinical translation, multicenter accurate HSI datasets are essential, but further work is needed to quantify the clinical value of HSI. HSI might be included in a new omics science, namely surgical optomics, which uses light to extract quantifiable tissue features during surgery.
PMID:38789623 | DOI:10.1007/s00464-024-10880-1
CodonBERT: a BERT-based architecture tailored for codon optimization using the cross-attention mechanism
Bioinformatics. 2024 May 24:btae330. doi: 10.1093/bioinformatics/btae330. Online ahead of print.
ABSTRACT
MOTIVATION: Due to the varying delivery methods of messenger RNA (mRNA) vaccines, codon optimization plays a critical role in vaccine design to improve the stability and expression of proteins in specific tissues. Considering the many-to-one relationship between synonymous codons and amino acids, the number of mRNA sequences encoding the same amino acid sequence could be enormous. Finding stable and highly expressed mRNA sequences from the vast sequence space using in silico methods can generally be viewed as a path-search problem or a machine translation problem. However, current deep learning-based methods inspired by machine translation may have some limitations, such as recurrent neural networks (RNNs), which have a weak ability to capture the long-term dependencies of codon preferences.
RESULTS: We develop a BERT-based architecture that uses the cross-attention mechanism for codon optimization. In CodonBERT, the codon sequence is randomly masked with each codon serving as a key and a value. In the meantime, the amino acid sequence is used as the query. CodonBERT was trained on high-expression transcripts from Human Protein Atlas mixed with different proportions of high codon adaptation index (CAI) codon sequences. The result showed that CodonBERT can effectively capture the long-term dependencies between codons and amino acids, suggesting that it can be used as a customized training framework for specific optimization targets.
AVAILABILITY: CodonBERT is freely available on https://github.com/FPPGroup/CodonBERT.
SUPPLEMENTARY INFORMATION: Supplementary data are available at Bioinformatics online.
PMID:38788220 | DOI:10.1093/bioinformatics/btae330
Deep learning approach for skin melanoma and benign classification using empirical wavelet decomposition
Technol Health Care. 2024 May 15. doi: 10.3233/THC-240020. Online ahead of print.
ABSTRACT
BACKGROUND: Melanoma is a malignant skin cancer that causes high mortality. Early detection of melanoma can save patients' lives. The features of the skin lesion images can be extracted using computer techniques to differentiate early between melanoma and benign skin lesions.
OBJECTIVE: A new model of empirical wavelet decomposition (EWD) based on tan hyperbolic modulated filter banks (THMFBs) (EWD-THMFBs) was used to obtain the features of skin lesion images by MATLAB software.
METHODS: The EWD-THMFBs model was compared with the empirical short-time Fourier decomposition method based on THMFBs (ESTFD-THMFBs) and the empirical Fourier decomposition method based on THMFBs (EFD-THMFBs).
RESULTS: The accuracy rates obtained for EWD-THMFBs, ESTFD-THMFBs, and EFD-THMFBs models were 100%, 98.89%, and 83.33%, respectively. The area under the curve (AUC) was 1, 0.97, and 0.91, respectively.
CONCLUSION: The EWD-THMFBs model performed best in extracting features from skin lesion images. This model can be programmed on a mobile to detect skin lesions in rural areas by a nurse before consulting a dermatologist.
PMID:38788103 | DOI:10.3233/THC-240020
Epistatic Features and Machine Learning Improve Alzheimer's Disease Risk Prediction Over Polygenic Risk Scores
J Alzheimers Dis. 2024 May 23. doi: 10.3233/JAD-230236. Online ahead of print.
ABSTRACT
BACKGROUND: Polygenic risk scores (PRS) are linear combinations of genetic markers weighted by effect size that are commonly used to predict disease risk. For complex heritable diseases such as late-onset Alzheimer's disease (LOAD), PRS models fail to capture much of the heritability. Additionally, PRS models are highly dependent on the population structure of the data on which effect sizes are assessed and have poor generalizability to new data.
OBJECTIVE: The goal of this study is to construct a paragenic risk score that, in addition to single genetic marker data used in PRS, incorporates epistatic interaction features and machine learning methods to predict risk for LOAD.
METHODS: We construct a new state-of-the-art genetic model for risk of Alzheimer's disease. Our approach innovates over PRS models in two ways: First, by directly incorporating epistatic interactions between SNP loci using an evolutionary algorithm guided by shared pathway information; and second, by estimating risk via an ensemble of non-linear machine learning models rather than a single linear model. We compare the paragenic model to several PRS models from the literature trained on the same dataset.
RESULTS: The paragenic model is significantly more accurate than the PRS models under 10-fold cross-validation, obtaining an AUC of 83% and near-clinically significant matched sensitivity/specificity of 75%. It remains significantly more accurate when evaluated on an independent holdout dataset and maintains accuracy within APOE genotype strata.
CONCLUSIONS: Paragenic models show potential for improving disease risk prediction for complex heritable diseases such as LOAD over PRS models.
PMID:38788065 | DOI:10.3233/JAD-230236
Automated Prediction of Proximal Middle Cerebral Artery Occlusions in Noncontrast Brain Computed Tomography
Stroke. 2024 Jun;55(6):1609-1618. doi: 10.1161/STROKEAHA.123.045772. Epub 2024 May 24.
ABSTRACT
BACKGROUND: Early identification of large vessel occlusion (LVO) in patients with ischemic stroke is crucial for timely interventions. We propose a machine learning-based algorithm (JLK-CTL) that uses handcrafted features from noncontrast computed tomography to predict LVO.
METHODS: We included patients with ischemic stroke who underwent concurrent noncontrast computed tomography and computed tomography angiography in seven hospitals. Patients from 5 of these hospitals, admitted between May 2011 and March 2015, were randomly divided into training and internal validation (9:1 ratio). Those from the remaining 2 hospitals, admitted between March 2021 and September 2021, were designated for external validation. From each noncontrast computed tomography scan, we extracted differences in volume, tissue density, and Hounsfield unit distribution between bihemispheric regions (striatocapsular, insula, M1-M3, and M4-M6, modified from the Alberta Stroke Program Early Computed Tomography Score). A deep learning algorithm was used to incorporate clot signs as an additional feature. Machine learning models, including ExtraTrees, random forest, extreme gradient boosting, support vector machine, and multilayer perceptron, as well as a deep learning model, were trained and evaluated. Additionally, we assessed the models' performance after incorporating the National Institutes of Health Stroke Scale scores as an additional feature.
RESULTS: Among 2919 patients, 83 were excluded. Across the training (n=2463), internal validation (n=275), and external validation (n=95) datasets, the mean ages were 68.5±12.4, 67.6±13.8, and 67.9±13.6 years, respectively. The proportions of men were 57%, 53%, and 59%, with LVO prevalences of 17.0%, 16.4%, and 26.3%, respectively. In the external validation, the ExtraTrees model achieved a robust area under the curve of 0.888 (95% CI, 0.850-0.925), with a sensitivity of 80.1% (95% CI, 72.0-88.1) and a specificity of 88.6% (95% CI, 84.7-92.5). Adding the National Institutes of Health Stroke Scale score to the ExtraTrees model increased sensitivity (from 80.1% to 92.1%) while maintaining specificity.
CONCLUSIONS: Our algorithm provides reliable predictions of LVO using noncontrast computed tomography. By enabling early LVO identification, our algorithm has the potential to expedite the stroke workflow.
PMID:38787932 | DOI:10.1161/STROKEAHA.123.045772
Automatic exudate and aneurysm segmentation in OCT images using UNET++ and hyperreflective-foci feature based bagged tree ensemble
PLoS One. 2024 May 24;19(5):e0304146. doi: 10.1371/journal.pone.0304146. eCollection 2024.
ABSTRACT
Diabetic retinopathy's signs, such as exudates (EXs) and aneurysms (ANs), initially develop from under the retinal surface detectable from optical coherence tomography (OCT) images. Detecting these signs helps ophthalmologists diagnose DR sooner. Detecting and segmenting exudates (EXs) and aneurysms (ANs) in medical images is challenging due to their small size, similarity to other hyperreflective regions, noise presence, and low background contrast. Furthermore, the scarcity of public OCT images featuring these abnormalities has limited the number of studies related to the automatic segmentation of EXs and ANs, and the reported performance of such studies has not been satisfactory. This work proposes an efficient algorithm that can automatically segment these anomalies by improving key steps in the process. The potential area where these hyper-reflective EXs and ANs occur was scoped by our method using a deep-learning U-Net++ program. From this area, the candidates for EX-AN were segmented using the adaptive thresholding method. Nine features based on appearances, locations, and shadow markers were extracted from these candidates. They were trained and tested using bagged tree ensemble classifiers to obtain only EX-AN blobs. The proposed method was tested on a collection of a public dataset comprising 80 images with hand-drawn ground truths. The experimental results showed that our method could segment EX-AN blobs with average recall, precision, and F1-measure as 87.9%, 86.1%, and 87.0%, respectively. Its F1-measure drastically outperformed two comparative methods, binary thresholding and watershed (BT-WS) and adaptive thresholding with shadow tracking (AT-ST), by 78.0% and 82.1%, respectively.
PMID:38787844 | DOI:10.1371/journal.pone.0304146
Accelerated construction of stress relief music datasets using CNN and the Mel-scaled spectrogram
PLoS One. 2024 May 24;19(5):e0300607. doi: 10.1371/journal.pone.0300607. eCollection 2024.
ABSTRACT
Listening to music is a crucial tool for relieving stress and promoting relaxation. However, the limited options available for stress-relief music do not cater to individual preferences, compromising its effectiveness. Traditional methods of curating stress-relief music rely heavily on measuring biological responses, which is time-consuming, expensive, and requires specialized measurement devices. In this paper, a deep learning approach to solve this problem is introduced that explicitly uses convolutional neural networks and provides a more efficient and economical method for generating large datasets of stress-relief music. These datasets are composed of Mel-scaled spectrograms that include essential sound elements (such as frequency, amplitude, and waveform) that can be directly extracted from the music. The trained model demonstrated a test accuracy of 98.7%, and a clinical study indicated that the model-selected music was as effective as researcher-verified music in terms of stress-relieving capacity. This paper underlines the transformative potential of deep learning in addressing the challenge of limited music options for stress relief. More importantly, the proposed method has profound implications for music therapy because it enables a more personalized approach to stress-relief music selection, offering the potential for enhanced emotional well-being.
PMID:38787824 | DOI:10.1371/journal.pone.0300607
Enhancing automated strabismus classification with limited data: Data augmentation using StyleGAN2-ADA
PLoS One. 2024 May 24;19(5):e0303355. doi: 10.1371/journal.pone.0303355. eCollection 2024.
ABSTRACT
In this study, we propose a generative data augmentation technique to overcome the challenges of severely limited data when designing a deep learning-based automated strabismus diagnosis system. We implement a generative model based on the StyleGAN2-ADA model for system design and assess strabismus classification performance using two classifiers. We evaluate the capability of our proposed method against traditional data augmentation techniques and confirm a substantial enhancement in performance. Furthermore, we conduct experiments to explore the relationship between the diagnosis agreement among ophthalmologists and the generation performance of the generative model. Beyond FID, we validate the generative samples on the classifier to establish their practicality. Through these experiments, we demonstrate that the generative model-based data augmentation improves overall quantitative performance in scenarios of extreme data scarcity and effectively mitigates overfitting issues during deep learning model training.
PMID:38787813 | DOI:10.1371/journal.pone.0303355
Multi-focused image fusion algorithm based on multi-scale hybrid attention residual network
PLoS One. 2024 May 24;19(5):e0302545. doi: 10.1371/journal.pone.0302545. eCollection 2024.
ABSTRACT
In order to improve the detection performance of image fusion in focus areas and realize end-to-end decision diagram optimization, we design a multi-focus image fusion network based on deep learning. The network is trained using unsupervised learning and a multi-scale hybrid attention residual network model is introduced to enable solving for features at different levels of the image. In the training stage, multi-scale features are extracted from two source images with different focal points using hybrid multi-scale residual blocks (MSRB), and the up-down projection module (UDP) is introduced to obtain multi-scale edge information, then the extracted features are operated to obtain deeper image features. These blocks can effectively utilize multi-scale feature information without increasing the number of parameters. The deep features of the image are extracted in its test phase, input to the spatial frequency domain to calculate and measure the activity level and obtain the initial decision map, and use post-processing techniques to eliminate the edge errors. Finally, the decision map is generated and optimized, and the final fused image is obtained by combining the optimized decision map with the source image. The comparative experiments show that our proposed model achieves better fusion performance in subjective evaluation, and the quality of the obtained fused images is more robust with richer details. The objective evaluation metrics work better and the image fusion quality is higher.
PMID:38787800 | DOI:10.1371/journal.pone.0302545
Sensitivity Decouple Learning for Image Compression Artifacts Reduction
IEEE Trans Image Process. 2024 May 24;PP. doi: 10.1109/TIP.2024.3403034. Online ahead of print.
ABSTRACT
With the benefit of deep learning techniques, recent researches have made significant progress in image compression artifacts reduction. Despite their improved performances, prevailing methods only focus on learning a mapping from the compressed image to the original one but ignore the intrinsic attributes of the given compressed images, which greatly harms the performance of downstream parsing tasks. Different from these methods, we propose to decouple the intrinsic attributes into two complementary features for artifacts reduction, i.e., the compression-insensitive features to regularize the high-level semantic representations during training and the compression-sensitive features to be aware of the compression degree. To achieve this, we first employ adversarial training to regularize the compressed and original encoded features for retaining high-level semantics, and we then develop the compression quality-aware feature encoder for compression-sensitive features. Based on these dual complementary features, we propose a Dual Awareness Guidance Network (DAGN) to utilize these awareness features as transformation guidance during the decoding phase. In our proposed DAGN, we develop a cross-feature fusion module to maintain the consistency of compression-insensitive features by fusing compression-insensitive features into the artifacts reduction baseline. Our method achieves an average 2.06 dB PSNR gains on BSD500, outperforming state-of-the-art methods, and only requires 29.7 ms to process one image on BSD500. Besides, the experimental results on LIVE1 and LIU4K also demonstrates the efficiency, effectiveness, and superiority of the proposed method in terms of quantitative metrics, visual quality, and downstream machine vision tasks.
PMID:38787669 | DOI:10.1109/TIP.2024.3403034
RawECGNet: Deep Learning Generalization for Atrial Fibrillation Detection From the Raw ECG
IEEE J Biomed Health Inform. 2024 May 24;PP. doi: 10.1109/JBHI.2024.3404877. Online ahead of print.
ABSTRACT
INTRODUCTION: Deep learning models for detecting episodes of atrial fibrillation (AF) using rhythm information in long-term ambulatory ECG recordings have shown high performance. However, the rhythm-based approach does not take advantage of the morphological information conveyed by the different ECG waveforms, particularly the f-waves. As a result, the performance of such models may be inherently limited.
METHODS: To address this limitation, we have developed a deep learning model, named RawECGNet, to detect episodes of AF and atrial flutter (AFl) using the raw, single-lead ECG. We compare the generalization performance of RawECGNet on two external data sets that account for distribution shifts in geography, ethnicity, and lead position. RawECGNet is further benchmarked against a state-of-the-art deep learning model, named ArNet2, which utilizes rhythm information as input.
RESULTS: Using RawECGNet, the results for the different leads in the external test sets in terms of the F1 score were 0.91-0.94 in RBDB and 0.93 in SHDB, compared to 0.89-0.91 in RBDB and 0.91 in SHDB for ArNet2. The results highlight RawECGNet as a high-performance, generalizable algorithm for detection of AF and AFl episodes, exploiting information on both rhythm and morphology.
PMID:38787663 | DOI:10.1109/JBHI.2024.3404877
Correction: Displacement detection with sub-pixel accuracy and high spatial resolution using deep learning
J Med Ultrason (2001). 2024 May 24. doi: 10.1007/s10396-024-01460-w. Online ahead of print.
NO ABSTRACT
PMID:38787517 | DOI:10.1007/s10396-024-01460-w
A Deep Learning Approach for Chromium Detection and Characterization from Soil Hyperspectral Data
Toxics. 2024 May 11;12(5):357. doi: 10.3390/toxics12050357.
ABSTRACT
High levels of chromium (Cr) in soil pose a significant threat to both humans and the environment. Laboratory-based chemical analysis methods for Cr are time consuming and expensive; thus, there is an urgent need for a more efficient method for detecting Cr in soil. In this study, a deep neural network (DNN) approach was applied to the Land Use and Cover Area frame Survey (LUCAS) dataset to develop a hyperspectral soil Cr content prediction model with good generalizability and accuracy. The optimal DNN model was constructed by optimizing the spectral preprocessing methods and DNN hyperparameters, which achieved good predictive performance for Cr detection, with a correlation coefficient value of 0.79 on the testing set. Four important hyperspectral bands with strong Cr sensitivity (400-439, 1364-1422, 1862-1934, and 2158-2499 nm) were identified by permutation importance and local interpretable model-agnostic explanations. Soil iron oxide and clay mineral content were found to be important factors influencing soil Cr content. The findings of this study provide a feasible method for rapidly determining soil Cr content from hyperspectral data, which can be further refined and applied to large-scale Cr detection in the future.
PMID:38787136 | DOI:10.3390/toxics12050357
A Review of Artificial Intelligence in Breast Imaging
Tomography. 2024 May 9;10(5):705-726. doi: 10.3390/tomography10050055.
ABSTRACT
With the increasing dominance of artificial intelligence (AI) techniques, the important prospects for their application have extended to various medical fields, including domains such as in vitro diagnosis, intelligent rehabilitation, medical imaging, and prognosis. Breast cancer is a common malignancy that critically affects women's physical and mental health. Early breast cancer screening-through mammography, ultrasound, or magnetic resonance imaging (MRI)-can substantially improve the prognosis for breast cancer patients. AI applications have shown excellent performance in various image recognition tasks, and their use in breast cancer screening has been explored in numerous studies. This paper introduces relevant AI techniques and their applications in the field of medical imaging of the breast (mammography and ultrasound), specifically in terms of identifying, segmenting, and classifying lesions; assessing breast cancer risk; and improving image quality. Focusing on medical imaging for breast cancer, this paper also reviews related challenges and prospects for AI.
PMID:38787015 | DOI:10.3390/tomography10050055