Deep learning
Deep learning automation of radiographic patterns for hallux valgus diagnosis
World J Orthop. 2024 Feb 18;15(2):105-109. doi: 10.5312/wjo.v15.i2.105. eCollection 2024 Feb 18.
ABSTRACT
Artificial intelligence (AI) and deep learning are becoming increasingly powerful tools in diagnostic and radiographic medicine. Deep learning has already been utilized for automated detection of pneumonia from chest radiographs, diabetic retinopathy, breast cancer, skin carcinoma classification, and metastatic lymphadenopathy detection, with diagnostic reliability akin to medical experts. In the World Journal of Orthopedics article, the authors apply an automated and AI-assisted technique to determine the hallux valgus angle (HVA) for assessing HV foot deformity. With the U-net neural network, the authors constructed an algorithm for pattern recognition of HV foot deformity from anteroposterior high-resolution radiographs. The performance of the deep learning algorithm was compared to expert clinician manual performance and assessed alongside clinician-clinician variability. The authors found that the AI tool was sufficient in assessing HVA and proposed the system as an instrument to augment clinical efficiency. Though further sophistication is needed to establish automated algorithms for more complicated foot pathologies, this work adds to the growing evidence supporting AI as a powerful diagnostic tool.
PMID:38464350 | PMC:PMC10921175 | DOI:10.5312/wjo.v15.i2.105
Ribonanza: deep learning of RNA structure through dual crowdsourcing
bioRxiv [Preprint]. 2024 Feb 27:2024.02.24.581671. doi: 10.1101/2024.02.24.581671.
ABSTRACT
Prediction of RNA structure from sequence remains an unsolved problem, and progress has been slowed by a paucity of experimental data. Here, we present Ribonanza, a dataset of chemical mapping measurements on two million diverse RNA sequences collected through Eterna and other crowdsourced initiatives. Ribonanza measurements enabled solicitation, training, and prospective evaluation of diverse deep neural networks through a Kaggle challenge, followed by distillation into a single, self-contained model called RibonanzaNet. When fine tuned on auxiliary datasets, RibonanzaNet achieves state-of-the-art performance in modeling experimental sequence dropout, RNA hydrolytic degradation, and RNA secondary structure, with implications for modeling RNA tertiary structure.
PMID:38464325 | PMC:PMC10925082 | DOI:10.1101/2024.02.24.581671
Deep-NCA: A deep learning methodology for performing noncompartmental analysis of pharmacokinetic data
CPT Pharmacometrics Syst Pharmacol. 2024 Mar 11. doi: 10.1002/psp4.13124. Online ahead of print.
ABSTRACT
Noncompartmental analysis (NCA) is a model-independent approach for assessing pharmacokinetics (PKs). Although the existing NCA algorithms are very well-established and widely utilized, they suffer from low accuracies in the setting of sparse PK samples. In response, we developed Deep-NCA, a deep learning (DL) model to improve the prediction of key noncompartmental PK parameters. Our methodology utilizes synthetic PK data for model training and uses an innovative patient-specific normalization method for data preprocessing. Deep-NCA demonstrated adequate performance across six previously unseen simulated drugs under multiple dosing, showcasing effective generalization. Compared to traditional NCA, Deep-NCA exhibited superior performance for sparse PK data. This study advances the application of DL to PK studies and introduces an effective method for handling sparse PK data. With further validation and refinement, Deep-NCA could significantly enhance the efficiency of drug development by providing more accurate NCA estimates while requiring fewer PK samples.
PMID:38465417 | DOI:10.1002/psp4.13124
Precise and Rapid Whole-Head Segmentation from Magnetic Resonance Images of Older Adults using Deep Learning
Imaging Neurosci (Camb). 2024 Mar;2. doi: 10.1162/imag_a_00090. Epub 2024 Feb 13.
ABSTRACT
Whole-head segmentation from Magnetic Resonance Images (MRI) establishes the foundation for individualized computational models using finite element method (FEM). This foundation paves the path for computer-aided solutions in fields, particularly in non-invasive brain stimulation. Most current automatic head segmentation tools are developed using healthy young adults. Thus, they may neglect the older population that is more prone to age-related structural decline such as brain atrophy. In this work, we present a new deep learning method called GRACE, which stands for General, Rapid, And Comprehensive whole-hEad tissue segmentation. GRACE is trained and validated on a novel dataset that consists of 177 manually corrected MR-derived reference segmentations that have undergone meticulous manual review. Each T1-weighted MRI volume is segmented into 11 tissue types, including white matter, grey matter, eyes, cerebrospinal fluid, air, blood vessel, cancellous bone, cortical bone, skin, fat, and muscle. To the best of our knowledge, this work contains the largest manually corrected dataset to date in terms of number of MRIs and segmented tissues. GRACE outperforms five freely available software tools and a traditional 3D U-Net on a five-tissue segmentation task. On this task, GRACE achieves an average Hausdorff Distance of 0.21, which exceeds the runner-up at an average Hausdorff Distance of 0.36. GRACE can segment a whole-head MRI in about 3 seconds, while the fastest software tool takes about 3 minutes. In summary, GRACE segments a spectrum of tissue types from older adults T1-MRI scans at favorable accuracy and speed. The trained GRACE model is optimized on older adult heads to enable high-precision modeling in age-related brain disorders. To support open science, the GRACE code and trained weights are made available online and open to the research community at https://github.com/lab-smile/GRACE.
PMID:38465203 | PMC:PMC10922731 | DOI:10.1162/imag_a_00090
Vascular changes of the choroid and their correlations with visual acuity in diabetic retinopathy
Front Endocrinol (Lausanne). 2024 Feb 23;15:1327325. doi: 10.3389/fendo.2024.1327325. eCollection 2024.
ABSTRACT
OBJECTIVE: To investigate changes in the choroidal vasculature and their correlations with visual acuity in diabetic retinopathy (DR).
METHODS: The cohort was composed of 225 eyes from 225 subjects, including 60 eyes from 60 subjects with healthy control, 55 eyes from 55 subjects without DR, 46 eyes from 46 subjects with nonproliferative diabetic retinopathy (NPDR), 21 eyes from 21 subjects with proliferative diabetic retinopathy (PDR), and 43 eyes from 43 subjects with clinically significant macular edema (CSME). Swept-source optical coherence tomography (SS-OCT) was used to image the eyes with a 12-mm radial line scan protocol. The parameters for 6-mm diameters of region centered on the macular fovea were analyzed. Initially, a custom deep learning algorithm based on a modified residual U-Net architecture was utilized for choroidal boundary segmentation. Subsequently, the SS-OCT image was binarized and the Niblack-based automatic local threshold algorithm was employed to calibrate subfoveal choroidal thickness (SFCT), luminal area (LA), and stromal area (SA) by determining the distance between the two boundaries. Finally, the ratio of LA and total choroidal area (SA + LA) was defined as the choroidal vascularity index (CVI). The choroidal parameters in five groups were compared, and correlations of the choroidal parameters with age, gender, duration of diabetes mellitus (DM), glycated hemoglobin (HbA1c), fasting blood sugar, SFCT and best-corrected visual acuity (BCVA) were analyzed.
RESULTS: The CVI, SFCT, LA, and SA values of patients with DR were found to be significantly lower compared to both healthy patients and patients without DR (P < 0.05). The SFCT was significantly higher in NPDR group compared to the No DR group (P < 0.001). Additionally, the SFCT was lower in the PDR group compared to the NPDR group (P = 0.014). Furthermore, there was a gradual decrease in CVI with progression of diabetic retinopathy, reaching its lowest value in the PDR group. However, the CVI of the CSME group exhibited a marginally closer proximity to that of the NPDR group. The multivariate regression analysis revealed a positive correlation between CVI and the duration of DM as well as LA (P < 0.05). The results of both univariate and multivariate regression analyses demonstrated a significant positive correlation between CVI and BCVA (P = 0.003).
CONCLUSION: Choroidal vascular alterations, especially decreased CVI, occurred in patients with DR. The CVI decreased with duration of DM and was correlated with visual impairment, indicating that the CVI might be a reliable imaging biomarker to monitor the progression of DR.
PMID:38464970 | PMC:PMC10920230 | DOI:10.3389/fendo.2024.1327325
Artificial-intelligence-driven measurements of brain metastases' response to SRS compare favorably with current manual standards of assessment
Neurooncol Adv. 2024 Feb 19;6(1):vdae015. doi: 10.1093/noajnl/vdae015. eCollection 2024 Jan-Dec.
ABSTRACT
BACKGROUND: Evaluation of treatment response for brain metastases (BMs) following stereotactic radiosurgery (SRS) becomes complex as the number of treated BMs increases. This study uses artificial intelligence (AI) to track BMs after SRS and validates its output compared with manual measurements.
METHODS: Patients with BMs who received at least one course of SRS and followed up with MRI scans were retrospectively identified. A tool for automated detection, segmentation, and tracking of intracranial metastases on longitudinal imaging, MEtastasis Tracking with Repeated Observations (METRO), was applied to the dataset. The longest three-dimensional (3D) diameter identified with METRO was compared with manual measurements of maximum axial BM diameter, and their correlation was analyzed. Change in size of the measured BM identified with METRO after SRS treatment was used to classify BMs as responding, or not responding, to treatment, and its accuracy was determined relative to manual measurements.
RESULTS: From 71 patients, 176 BMs were identified and measured with METRO and manual methods. Based on a one-to-one correlation analysis, the correlation coefficient was R2 = 0.76 (P = .0001). Using modified BM response classifications of BM change in size, the longest 3D diameter data identified with METRO had a sensitivity of 0.72 and a specificity of 0.95 in identifying lesions that responded to SRS, when using manual axial diameter measurements as the ground truth.
CONCLUSIONS: Using AI to automatically measure and track BM volumes following SRS treatment, this study showed a strong correlation between AI-driven measurements and the current clinically used method: manual axial diameter measurements.
PMID:38464949 | PMC:PMC10924534 | DOI:10.1093/noajnl/vdae015
Application of Artificial Intelligence to the Monitoring of Medication Adherence for Tuberculosis Treatment in Africa: Algorithm Development and Validation
JMIR AI. 2023 Jan-Dec;2(1):e40167. doi: 10.2196/40167. Epub 2023 Feb 23.
ABSTRACT
BACKGROUND: Artificial intelligence (AI) applications based on advanced deep learning methods in image recognition tasks can increase efficiency in the monitoring of medication adherence through automation. AI has sparsely been evaluated for the monitoring of medication adherence in clinical settings. However, AI has the potential to transform the way health care is delivered even in limited-resource settings such as Africa.
OBJECTIVE: We aimed to pilot the development of a deep learning model for simple binary classification and confirmation of proper medication adherence to enhance efficiency in the use of video monitoring of patients in tuberculosis treatment.
METHODS: We used a secondary data set of 861 video images of medication intake that were collected from consenting adult patients with tuberculosis in an institutional review board-approved study evaluating video-observed therapy in Uganda. The video images were processed through a series of steps to prepare them for use in a training model. First, we annotated videos using a specific protocol to eliminate those with poor quality. After the initial annotation step, 497 videos had sufficient quality for training the models. Among them, 405 were positive samples, whereas 92 were negative samples. With some preprocessing techniques, we obtained 160 frames with a size of 224 × 224 in each video. We used a deep learning framework that leveraged 4 convolutional neural networks models to extract visual features from the video frames and automatically perform binary classification of adherence or nonadherence. We evaluated the diagnostic properties of the different models using sensitivity, specificity, F1-score, and precision. The area under the curve (AUC) was used to assess the discriminative performance and the speed per video review as a metric for model efficiency. We conducted a 5-fold internal cross-validation to determine the diagnostic and discriminative performance of the models. We did not conduct external validation due to a lack of publicly available data sets with specific medication intake video frames.
RESULTS: Diagnostic properties and discriminative performance from internal cross-validation were moderate to high in the binary classification tasks with 4 selected automated deep learning models. The sensitivity ranged from 92.8 to 95.8%, specificity from 43.5 to 55.4%, F1-score from 0.91 to 0.92, precision from 88% to 90.1%, and AUC from 0.78 to 0.85. The 3D ResNet model had the highest precision, AUC, and speed.
CONCLUSIONS: All 4 deep learning models showed comparable diagnostic properties and discriminative performance. The findings serve as a reasonable proof of concept to support the potential application of AI in the binary classification of video frames to predict medication adherence.
PMID:38464947 | PMC:PMC10923555 | DOI:10.2196/40167
MGA-NET: MULTI-SCALE GUIDED ATTENTION MODELS FOR AN AUTOMATED DIAGNOSIS OF IDIOPATHIC PULMONARY FIBROSIS (IPF)
Proc IEEE Int Symp Biomed Imaging. 2021 Apr;2021:1777-1780. doi: 10.1109/isbi48211.2021.9433956. Epub 2021 May 25.
ABSTRACT
We propose a Multi-scale, domain knowledge-Guided Attention model (MGA-Net) for a weakly supervised problem - disease diagnosis with only coarse scan-level labels. The use of guided attention models encourages the deep learning-based diagnosis model to focus on the area of interests (in our case, lung parenchyma), at different resolutions, in an end-to-end manner. The research interest is to diagnose subjects with idiopathic pulmonary fibrosis (IPF) among subjects with interstitial lung disease (ILD) using an axial chest high resolution computed tomography (HRCT) scan. Our dataset contains 279 IPF patients and 423 non-IPF ILD patients. The network's performance was evaluated by the area under the receiver operating characteristic curve (AUC) with standard errors (SE) using stratified five-fold cross validation. We observe that without attention modules, the IPF diagnosis model performs unsatisfactorily (AUC±SE =0.690 ± 0.194); by including unguided attention module, the IPF diagnosis model reaches satisfactory performance (AUC±SE =0.956±0.040), but lack explainability; when including only guided high- or medium- resolution attention, the learned attention maps highlight the lung areas but the AUC decreases; when including both high- and medium- resolution attention, the model reaches the highest AUC among all experiments (AUC± SE =0.971 ±0.021) and the estimated attention maps concentrate on the regions of interests for this task. Our results suggest that, for a weakly supervised task, MGA-Net can utilize the population-level domain knowledge to guide the training of the network in an end-to-end manner, which increases both model accuracy and explainability.
PMID:38464881 | PMC:PMC10924672 | DOI:10.1109/isbi48211.2021.9433956
Automatic recognition of depression based on audio and video: A review
World J Psychiatry. 2024 Feb 19;14(2):225-233. doi: 10.5498/wjp.v14.i2.225. eCollection 2024 Feb 19.
ABSTRACT
Depression is a common mental health disorder. With current depression detection methods, specialized physicians often engage in conversations and physiological examinations based on standardized scales as auxiliary measures for depression assessment. Non-biological markers-typically classified as verbal or non-verbal and deemed crucial evaluation criteria for depression-have not been effectively utilized. Specialized physicians usually require extensive training and experience to capture changes in these features. Advancements in deep learning technology have provided technical support for capturing non-biological markers. Several researchers have proposed automatic depression estimation (ADE) systems based on sounds and videos to assist physicians in capturing these features and conducting depression screening. This article summarizes commonly used public datasets and recent research on audio- and video-based ADE based on three perspectives: Datasets, deficiencies in existing research, and future development directions.
PMID:38464777 | PMC:PMC10921287 | DOI:10.5498/wjp.v14.i2.225
High-resolution density assessment assisted by deep learning of Dendrophyllia cornigera (Lamarck, 1816) and Phakellia ventilabrum (Linnaeus, 1767) in rocky circalittoral shelf of Bay of Biscay
PeerJ. 2024 Mar 7;12:e17080. doi: 10.7717/peerj.17080. eCollection 2024.
ABSTRACT
This study presents a novel approach to high-resolution density distribution mapping of two key species of the 1170 "Reefs" habitat, Dendrophyllia cornigera and Phakellia ventilabrum, in the Bay of Biscay using deep learning models. The main objective of this study was to establish a pipeline based on deep learning models to extract species density data from raw images obtained by a remotely operated towed vehicle (ROTV). Different object detection models were evaluated and compared in various shelf zones at the head of submarine canyon systems using metrics such as precision, recall, and F1 score. The best-performing model, YOLOv8, was selected for generating density maps of the two species at a high spatial resolution. The study also generated synthetic images to augment the training data and assess the generalization capacity of the models. The proposed approach provides a cost-effective and non-invasive method for monitoring and assessing the status of these important reef-building species and their habitats. The results have important implications for the management and protection of the 1170 habitat in Spain and other marine ecosystems worldwide. These results highlight the potential of deep learning to improve efficiency and accuracy in monitoring vulnerable marine ecosystems, allowing informed decisions to be made that can have a positive impact on marine conservation.
PMID:38464748 | PMC:PMC10924775 | DOI:10.7717/peerj.17080
MCP-Net: Inter-frame Motion Correction with Patlak Regularization for Whole-body Dynamic PET
Med Image Comput Comput Assist Interv. 2022 Sep;13434:163-172. doi: 10.1007/978-3-031-16440-8_16. Epub 2022 Sep 16.
ABSTRACT
Inter-frame patient motion introduces spatial misalignment and degrades parametric imaging in whole-body dynamic positron emission tomography (PET). Most current deep learning inter-frame motion correction works consider only the image registration problem, ignoring tracer kinetics. We propose an inter-frame Motion Correction framework with Patlak regularization (MCP-Net) to directly optimize the Patlak fitting error and further improve model performance. The MCP-Net contains three modules: a motion estimation module consisting of a multiple-frame 3-D U-Net with a convolutional long short-term memory layer combined at the bottleneck; an image warping module that performs spatial transformation; and an analytical Patlak module that estimates Patlak fitting with the motion-corrected frames and the individual input function. A Patlak loss penalization term using mean squared percentage fitting error is introduced to the loss function in addition to image similarity measurement and displacement gradient loss. Following motion correction, the parametric images were generated by standard Patlak analysis. Compared with both traditional and deep learning benchmarks, our network further corrected the residual spatial mismatch in the dynamic frames, improved the spatial alignment of Patlak Ki/Vb images, and reduced normalized fitting error. With the utilization of tracer dynamics and enhanced network performance, MCP-Net has the potential for further improving the quantitative accuracy of dynamic PET. Our code is released at https://github.com/gxq1998/MCP-Net.
PMID:38464686 | PMC:PMC10923180 | DOI:10.1007/978-3-031-16440-8_16
Deep Learning Combined with Radiologist's Intervention Achieves Accurate Segmentation of Hepatocellular Carcinoma in Dual-Phase Magnetic Resonance Images
Biomed Res Int. 2024 Feb 28;2024:9267554. doi: 10.1155/2024/9267554. eCollection 2024.
ABSTRACT
PURPOSE: Segmentation of hepatocellular carcinoma (HCC) is crucial; however, manual segmentation is subjective and time-consuming. Accurate and automatic lesion contouring for HCC is desirable in clinical practice. In response to this need, our study introduced a segmentation approach for HCC combining deep convolutional neural networks (DCNNs) and radiologist intervention in magnetic resonance imaging (MRI). We sought to design a segmentation method with a deep learning method that automatically segments using manual location information for moderately experienced radiologists. In addition, we verified the viability of this method to assist radiologists in accurate and fast lesion segmentation.
METHOD: In our study, we developed a semiautomatic approach for segmenting HCC using DCNN in conjunction with radiologist intervention in dual-phase gadolinium-ethoxybenzyl-diethylenetriamine penta-acetic acid- (Gd-EOB-DTPA-) enhanced MRI. We developed a DCNN and deep fusion network (DFN) trained on full-size images, namely, DCNN-F and DFN-F. Furthermore, DFN was applied to the image blocks containing tumor lesions that were roughly contoured by a radiologist with 10 years of experience in abdominal MRI, and this method was named DFN-R. Another radiologist with five years of experience (moderate experience) performed tumor lesion contouring for comparison with our proposed methods. The ground truth image was contoured by an experienced radiologist and reviewed by an independent experienced radiologist.
RESULTS: The mean DSC of DCNN-F, DFN-F, and DFN-R was 0.69 ± 0.20 (median, 0.72), 0.74 ± 0.21 (median, 0.77), and 0.83 ± 0.13 (median, 0.88), respectively. The mean DSC of the segmentation by the radiologist with moderate experience was 0.79 ± 0.11 (median, 0.83), which was lower than the performance of DFN-R.
CONCLUSIONS: Deep learning using dual-phase MRI shows great potential for HCC lesion segmentation. The radiologist-aided semiautomated method (DFN-R) achieved improved performance compared to manual contouring by the radiologist with moderate experience, although the difference was not statistically significant.
PMID:38464681 | PMC:PMC10923620 | DOI:10.1155/2024/9267554
Multi input-Multi output 3D CNN for dementia severity assessment with incomplete multimodal data
Artif Intell Med. 2024 Mar;149:102774. doi: 10.1016/j.artmed.2024.102774. Epub 2024 Jan 24.
ABSTRACT
Alzheimer's Disease is the most common cause of dementia, whose progression spans in different stages, from very mild cognitive impairment to mild and severe conditions. In clinical trials, Magnetic Resonance Imaging (MRI) and Positron Emission Tomography (PET) are mostly used for the early diagnosis of neurodegenerative disorders since they provide volumetric and metabolic function information of the brain, respectively. In recent years, Deep Learning (DL) has been employed in medical imaging with promising results. Moreover, the use of the deep neural networks, especially Convolutional Neural Networks (CNNs), has also enabled the development of DL-based solutions in domains characterized by the need of leveraging information coming from multiple data sources, raising the Multimodal Deep Learning (MDL). In this paper, we conduct a systematic analysis of MDL approaches for dementia severity assessment exploiting MRI and PET scans. We propose a Multi Input-Multi Output 3D CNN whose training iterations change according to the characteristic of the input as it is able to handle incomplete acquisitions, in which one image modality is missed. Experiments performed on OASIS-3 dataset show the satisfactory results of the implemented network, which outperforms approaches exploiting both single image modality and different MDL fusion techniques.
PMID:38462278 | DOI:10.1016/j.artmed.2024.102774
A clinically actionable and explainable real-time risk assessment framework for stroke-associated pneumonia
Artif Intell Med. 2024 Mar;149:102772. doi: 10.1016/j.artmed.2024.102772. Epub 2024 Jan 20.
ABSTRACT
The current medical practice is more responsive rather than proactive, despite the widely recognized value of early disease detection, including improving the quality of care and reducing medical costs. One of the cornerstones of early disease detection is clinically actionable predictions, where predictions are expected to be accurate, stable, real-time and interpretable. As an example, we used stroke-associated pneumonia (SAP), setting up a transformer-encoder-based model that analyzes highly heterogeneous electronic health records in real-time. The model was proven accurate and stable on an independent test set. In addition, it issued at least one warning for 98.6 % of SAP patients, and on average, its alerts were ahead of physician diagnoses by 2.71 days. We applied Integrated Gradient to glean the model's reasoning process. Supplementing the risk scores, the model highlighted critical historical events on patients' trajectories, which were shown to have high clinical relevance.
PMID:38462273 | DOI:10.1016/j.artmed.2024.102772
Multi-task reconstruction network for synthetic diffusion kurtosis imaging: Predicting neoadjuvant chemoradiotherapy response in locally advanced rectal cancer
Eur J Radiol. 2024 Mar 2;174:111402. doi: 10.1016/j.ejrad.2024.111402. Online ahead of print.
ABSTRACT
PURPOSE: To assess the feasibility and clinical value of synthetic diffusion kurtosis imaging (DKI) generated from diffusion weighted imaging (DWI) through multi-task reconstruction network (MTR-Net) for tumor response prediction in patients with locally advanced rectal cancer (LARC).
METHODS: In this retrospective study, 120 eligible patients with LARC were enrolled and randomly divided into training and testing datasets with a 7:3 ratio. The MTR-Net was developed for reconstructing Dapp and Kapp images from apparent diffusion coefficient (ADC) images. Tumor regions were manually segmented on both true and synthetic DKI images. The synthetic image quality and manual segmentation agreement were quantitatively assessed. The support vector machine (SVM) classifier was used to construct radiomics models based on the true and synthetic DKI images for pathological complete response (pCR) prediction. The prediction performance for the models was evaluated by the receiver operating characteristic (ROC) curve analysis.
RESULTS: The mean squared error (MSE), peak signal-to-noise ratio (PSNR), and structural similarity index measure (SSIM) for tumor regions were 0.212, 24.278, and 0.853, respectively, for the synthetic Dapp images and 0.516, 24.883, and 0.804, respectively, for the synthetic Kapp images. The Dice similarity coefficient (DSC), positive predictive value (PPV), sensitivity (SEN), and Hausdorff distance (HD) for the manually segmented tumor regions were 0.786, 0.844, 0.755, and 0.582, respectively. For predicting pCR, the true and synthetic DKI-based radiomics models achieved area under the curve (AUC) values of 0.825 and 0.807 in the testing datasets, respectively.
CONCLUSIONS: Generating synthetic DKI images from DWI images using MTR-Net is feasible, and the efficiency of synthetic DKI images in predicting pCR is comparable to that of true DKI images.
PMID:38461737 | DOI:10.1016/j.ejrad.2024.111402
A systematic review of the application of deep learning techniques in the physiotherapeutic therapy of musculoskeletal pathologies
Comput Biol Med. 2024 Jan 29;172:108082. doi: 10.1016/j.compbiomed.2024.108082. Online ahead of print.
ABSTRACT
Physiotherapy is a critical area of healthcare that involves the assessment and treatment of physical disabilities and injuries. The use of Artificial Intelligence (AI) in physiotherapy has gained significant attention due to its potential to enhance the accuracy and effectiveness of clinical decision-making and treatment outcomes. Nevertheless, it is still a rather innovative field of application of these techniques and there is a need to find what aspects are highly developed and what possible job niches can be exploited. This systematic review aims to evaluate the current state of research on the use of a particular AI called deep learning models in physiotherapy and identify the key trends, challenges, and opportunities in this field. The findings of this review, conducted following the PRISMA guidelines, provide valuable insights for researchers and clinicians. The most relevant databases included were PubMed, Web of Science, Scopus, Astrophysics Data System, and Central Citation Export. Inclusion and exclusion criteria were established to determine which items would be considered for further review. These criteria were used to screen the items during the first and second peer review processes. A set of quality criteria was developed to select the papers obtained after the second screening. Finally, of the 214 initial papers, 23 studies were selected. From our analysis of the selected articles, we can draw the following conclusions: Concerning deep learning models, innovation is primarily seen in the adoption of hybrid models, with convolutional models being extensively utilized. In terms of data, it's unsurprising that body signals and images are predominantly used. However, texts and structured data present promising avenues for groundbreaking work in the field. Additionally, medical tests that involve the collection of 3D images, Functional Movement Screening, or thermographies emerge as novel areas to explore applications within the scope of physiotherapy.
PMID:38461697 | DOI:10.1016/j.compbiomed.2024.108082
Transformer based multi-modal MRI fusion for prediction of post-menstrual age and neonatal brain development analysis
Med Image Anal. 2024 Mar 7;94:103140. doi: 10.1016/j.media.2024.103140. Online ahead of print.
ABSTRACT
The brain development during the perinatal period is characterized by rapid changes in both structure and function, which have significant impact on the cognitive and behavioral abilities later in life. Accurate assessment of brain age is a crucial indicator for brain development maturity and can help predict the risk of neonatal pathology. However, evaluating neonatal brains using magnetic resonance imaging (MRI) is challenging due to its complexity, multi-dimension, and noise with subtle alterations. In this paper, we propose a multi-modal deep learning framework based on transformers for precise post-menstrual age (PMA) estimation and brain development analysis using T2-weighted structural MRI (T2-sMRI) and diffusion MRI (dMRI) data. First, we build a two-stream dense network to learn modality-specific features from T2-sMRI and dMRI of brain individually. Then, a transformer module based on self-attention mechanism integrates these features for PMA prediction and preterm/term classification. Finally, saliency maps on brain templates are used to enhance the interpretability of results. Our method is evaluated on the multi-modal MRI dataset of the developing Human Connectome Project (dHCP), which contains 592 neonates, including 478 term-born and 114 preterm-born subjects. The results demonstrate that our method achieves a 0.5-week mean absolute error (MAE) in PMA estimation for term-born subjects. Notably, preterm-born subjects exhibit delayed brain development, worsening with increasing prematurity. Our method also achieves 95% accuracy in classification of term-born and preterm-born subjects, revealing significant group differences.
PMID:38461655 | DOI:10.1016/j.media.2024.103140
Reduction of Motion Artifacts in Liver MRI Using Deep Learning with High-pass Filtering
Nihon Hoshasen Gijutsu Gakkai Zasshi. 2024 Mar 11. doi: 10.6009/jjrt.2024-1408. Online ahead of print.
ABSTRACT
PURPOSE: To investigate whether deep learning with high-pass filtering can be used to effectively reduce motion artifacts in magnetic resonance (MR) images of the liver.
METHODS: The subjects were 69 patients who underwent liver MR examination at our hospital. Simulated motion artifact images (SMAIs) were created from non-artifact images (NAIs) and used for deep learning. Structural similarity index measure (SSIM) and contrast ratio (CR) were used to verify the effect of reducing motion artifacts in motion artifact reduction image (MARI) output from the obtained deep learning model. In the visual assessment, reduction of motion artifacts and image sharpness were evaluated between motion artifact images (MAIs) and MARIs.
RESULTS: The SSIM values were 0.882 on the MARIs and 0.869 on the SMAIs. There was no statistically significant difference in CR between NAIs and MARIs. The visual assessment showed that MARIs had reduced motion artifacts and improved sharpness compared to MAIs.
CONCLUSION: The learning model in this study is indicated to be reduced motion artifacts without decreasing the sharpness of liver MR images.
PMID:38462509 | DOI:10.6009/jjrt.2024-1408
Preliminary study on automatic quantification and grading of leopard spots fundus based on deep learning technology
Zhonghua Yan Ke Za Zhi. 2024 Mar 11;60(3):257-264. doi: 10.3760/cma.j.cn112142-20231210-00281.
ABSTRACT
Objective: To achieve automatic segmentation, quantification, and grading of different regions of leopard spots fundus (FT) using deep learning technology. The analysis includes exploring the correlation between novel quantitative indicators, leopard spot fundus grades, and various systemic and ocular parameters. Methods: This was a cross-sectional study. The data were sourced from the Beijing Eye Study, a population-based longitudinal study. In 2001, a group of individuals aged 40 and above were surveyed in five urban communities in Haidian District and three rural communities in Daxing District of Beijing. A follow-up was conducted in 2011. This study included individuals aged 50 and above who participated in the second 5-year follow-up in 2011, considering only the data from the right eye. Color fundus images centered on the macula of the right eye were input into the leopard spot segmentation model and macular detection network. Using the macular center as the origin, with inner circle diameters of 1 mm, 3 mm, and outer circle diameter of 6 mm, fine segmentation of the fundus was achieved. This allowed the calculation of the leopard spot density (FTD) and leopard spot grade for each region. Further analyses of the differences in ocular and systemic parameters among different regions' FTD and leopard spot grades were conducted. The participants were categorized into three refractive types based on equivalent spherical power (SE): myopia (SE<-0.25 D), emmetropia (-0.25 D≤SE≤0.25 D), and hyperopia (SE>0.25 D). Based on axial length, the participants were divided into groups with axial length<24 mm, 24-26 mm, and>26 mm for the analysis of different types of FTD. Statistical analyses were performed using one-way analysis of variance, Kruskal-Wallis test, Bonferroni test, and Spearman correlation analysis. Results: The study included 3 369 participants (3 369 eyes) with an average age of (63.9±10.6) years; among them, 1 886 were female (56.0%) and 1, 483 were male (64.0%). The overall FTD for all eyes was 0.060 (0.016, 0.163); inner circle FTD was 0.000 (0.000, 0.025); middle circle FTD was 0.030 (0.000, 0.130); outer circle FTD was 0.055 (0.009, 0.171). The results of the univariate analysis indicated that FTD in various regions was correlated with axial length (overall: r=0.38, P<0.001; inner circle: r=0.31, P<0.001; middle circle: r=0.36, P<0.001; outer circle: r=0.39, P<0.001), subfoveal choroidal thickness (SFCT) (overall: r=-0.69, P<0.001; inner circle: r=-0.57, P<0.001; middle circle: r=-0.68, P<0.001; outer circle: r=-0.72, P<0.001), age (overall: r=0.34, P<0.001; inner circle: r=0.30, P<0.001; middle circle: r=0.31, P<0.001; outer circle: r=0.35, P<0.001), gender (overall: r=-0.11, P<0.001; inner circle: r=-0.04, P<0.001; middle circle: r=-0.07, P<0.001; outer circle: r=-0.11, P<0.001), SE (overall: r=-0.20; P<0.001; inner circle: r=-0.19, P<0.001; middle circle: r=-0.20, P<0.001; outer circle: r=-0.20, P<0.001), uncorrected visual acuity (overall: r=-0.18, P<0.001; inner circle: r=-0.26, P<0.001; middle circle: r=-0.24, P<0.001; outer circle: r=-0.22, P<0.001), and body mass index (BMI) (overall: r=-0.11, P<0.001; inner circle: r=-0.13, P<0.001; middle circle: r=-0.14, P<0.001; outer circle: r=-0.13, P<0.001). Further multivariate analysis results indicated that different region FTD was correlated with axial length (overall: β=0.020, P<0.001; inner circle: β=-0.022, P<0.001; middle circle: β=0.027, P<0.001; outer circle: β=0.022, P<0.001), SFCT (overall: β=-0.001, P<0.001; inner circle: β=-0.001, P<0.001; middle circle: β=-0.001, P<0.001; outer circle: β=-0.001, P<0.001), and age (overall: β=0.002, P<0.001; inner circle: β=0.001, P<0.001; middle circle: β=0.002, P<0.001; outer circle: β=0.002, P<0.001). The distribution of overall (H=56.76, P<0.001), inner circle (H=72.22, P<0.001), middle circle (H=75.83, P<0.001), and outer circle (H=70.34, P<0.001) FTD differed significantly among different refractive types. The distribution of overall (H=373.15, P<0.001), inner circle (H=367.67, P<0.001), middle circle (H=389.14, P<0.001), and outer circle (H=386.89, P<0.001) FTD differed significantly among different axial length groups. Furthermore, comparing various levels of FTD with systemic and ocular parameters, significant differences were found in axial length (F=142.85, P<0.001) and SFCT (F=530.46, P<0.001). Conclusions: The use of deep learning technology enables automatic segmentation and quantification of different regions of theFT, as well as preliminary grading. Different region FTD is significantly correlated with axial length, SFCT, and age. Individuals with older age, myopia, and longer axial length tend to have higher FTD and more advanced FT grades.
PMID:38462374 | DOI:10.3760/cma.j.cn112142-20231210-00281
An in silico scheme for optimizing the enzymatic acquisition of natural biologically active peptides based on machine learning and virtual digestion
Anal Chim Acta. 2024 Apr 15;1298:342419. doi: 10.1016/j.aca.2024.342419. Epub 2024 Feb 26.
ABSTRACT
BACKGROUND: As a potential natural active substance, natural biologically active peptides (NBAPs) are recently attracting increasing attention. The traditional proteolysis methods of obtaining effective NBAPs are considerably vexing, especially since multiple proteases can be used, which blocks the exploration of available NBAPs. Although the development of virtual digesting brings some degree of convenience, the activity of the obtained peptides remains unclear, which would still not allow efficient access to the NBAPs. It is necessary to develop an efficient and accurate strategy for acquiring NBAPs.
RESULTS: A new in silico scheme named SSA-LSTM-VD, which combines a sparrow search algorithm-long short-term memory (SSA-LSTM) deep learning and virtually digested, was presented to optimize the proteolysis acquisition of NBAPs. Therein, SSA-LSTM reached the highest Efficiency value reached 98.00 % compared to traditional machine learning algorithms, and basic LSTM algorithm. SSA-LSTM was trained to predict the activity of peptides in the proteins virtually digested results, obtain the percentage of target active peptide, and select the appropriate protease for the actual experiment. As an application, SSA-LSTM was employed to predict the percentage of neuroprotective peptides in the virtual digested result of walnut protein, and trypsin was ultimately found to possess the highest value (85.29 %). The walnut protein was digested by trypsin (WPTrH) and the peptide sequence obtained was analyzed closely matches the theoretical neuroprotective peptide. More importantly, the neuroprotective effects of WPTrH had been demonstrated in nerve damage mouse models.
SIGNIFICANCE: The proposed SSA-LSTM-VD in this paper makes the acquisition of NBAPs efficient and accurate. The approach combines deep learning and virtually digested skillfully. Utilizing the SSA-LSTM-VD based strategy holds promise for discovering and developing peptides with neuroprotective properties or other desired biological activities.
PMID:38462343 | DOI:10.1016/j.aca.2024.342419