Deep learning

Automated and explainable machine learning for monitoring lipid and protein oxidative damage in mutton using hyperspectral imaging

Sat, 2025-03-01 06:00

Food Res Int. 2025 Feb;203:115905. doi: 10.1016/j.foodres.2025.115905. Epub 2025 Feb 1.

ABSTRACT

Current detection methods for lipid and protein oxidation using hyperspectral imaging (HSI) in conjunction with machine learning (ML) necessitate the involvement of data scientists and domain experts to adjust the model architecture and tune hyperparameters. Additionally, prediction models lack explainability in the predictive outcomes and decision-making process. In this study, ML, automated machine learning (AutoML) and automated deep learning (AutoDL) models were developed for visible near-infrared HSI of mutton samples treated with different freeze-thaw cycles to evaluate the feasibility of building prediction models for lipid and protein oxidation without manual intervention. SHapley Additive exPlanations (SHAP) were utilized to explain the prediction models. The results showed that the AutoDL attained the effective prediction models for lipid oxidation (R2p = 0.9021, RMSEP = 0.0542 mg/kg, RPD = 3.3624) and protein oxidation (R2p = 0.8805, RMSEP = 3.8065 nmol/mg, RPD = 3.0789). AutoML driven stacked ensembles further improved the generalization ability of the models, predicting lipid and protein oxidation with R2p of 0.9237 and 0.9347. The important wavelengths identified through SHAP closely align with the results obtained from spectral analysis, and the analysis also determined the magnitude and direction of the impact of these important wavelengths on the model outputs. Finally, changes in lipid and protein oxidation of mutton in different freeze-thaw cycles were visualized. The research indicated that the combination of HSI, AutoML and SHAP may generate high-quality explainable models without human assistance for monitoring lipid and protein oxidative damage in mutton.

PMID:40022412 | DOI:10.1016/j.foodres.2025.115905

Categories: Literature Watch

A robust deep learning model for predicting green tea moisture content during fixation using near-infrared spectroscopy: Integration of multi-scale feature fusion and attention mechanisms

Sat, 2025-03-01 06:00

Food Res Int. 2025 Feb;203:115874. doi: 10.1016/j.foodres.2025.115874. Epub 2025 Jan 30.

ABSTRACT

Fixation is a critical step in green tea processing, and the moisture content of the leaves after fixation is a key indicator of the fixation quality. Near-infrared spectroscopy (NIRS)-based moisture detection technology is often applied in the tea processing industry. However, temperature fluctuations during processing can cause changes in the NIRS curves, which in turn affect the accuracy of moisture prediction models based on the spectral data. To address this challenge, NIRS data were collected from samples at various stages of fixation and at different temperatures, and a novel deep learning network (DiSENet) was proposed, which integrates multi-scale feature fusion and attention mechanisms. Using a global modeling approach, the proposed method achieved a coefficient of determination (RP2) of 0.781 for moisture content prediction, with a root mean square error (RMSEP) of 1.720 % and a residual predictive deviation (RPD) of 2.148. On the dataset constructed for this study, DiSENet demonstrated superior predictive accuracy compared to the spectral correction methods of external parameter orthogonalization (EPO) and generalized least squares weighting (GLSW), as well as traditional global modeling methods such as partial least squares regression (PLSR) and support vector regression (SVR). This approach effectively corrects spectral interferences caused by temperature variations, thereby enhancing the accuracy of moisture content prediction. Thus, it offers a reliable solution for real-time, non-destructive moisture detection during tea processing.

PMID:40022390 | DOI:10.1016/j.foodres.2025.115874

Categories: Literature Watch

Discrimination of unsound soybeans using hyperspectral imaging: A deep learning method based on dual-channel feature fusion strategy and attention mechanism

Sat, 2025-03-01 06:00

Food Res Int. 2025 Feb;203:115810. doi: 10.1016/j.foodres.2025.115810. Epub 2025 Jan 22.

ABSTRACT

The application of high-level data fusion in the detection of agricultural products still presents a significant challenge. In this study, dual-channel feature fusion model (DCFFM) with attention mechanism was proposed to optimize the utilization of both one-dimensional spectral data and two-dimensional image data in the hyperspectral images for achieving high-level data fusion. A comparative analysis of support vector machine (SVM), convolutional neural network (CNN) with DCFFM, demonstrated that DCFFM exhibited superior results, achieving the accuracy, precision, recall, specificity, and F1-score of 95.13 %, 95.49 %, 94.83 %, 98.97 %, 95.12 % in the visible and near-infrared (Vis-NIR), and 94.00 %, 94.43 %, 94.16 %, 98.67 %, 94.27 % in the short-wave infrared (SWIR). This also indicated that Vis-NIR was more suitable for identifying unsound soybeans than SWIR. Furthermore, visualization was employed to demonstrate classification outcomes, thereby illustrating the generalization capacity of DCFFM through model inversion. In summary, this study is to explore a modeling framework that is capable of the comprehensive acquisition of spectra and images in the hyperspectral images, allowing for high-level data fusion, thereby achieving enhanced levels of accuracy.

PMID:40022337 | DOI:10.1016/j.foodres.2025.115810

Categories: Literature Watch

Preoperative diagnosis of meningioma sinus invasion based on MRI radiomics and deep learning: a multicenter study

Fri, 2025-02-28 06:00

Cancer Imaging. 2025 Feb 28;25(1):20. doi: 10.1186/s40644-025-00845-5.

ABSTRACT

OBJECTIVE: Exploring the construction of a fusion model that combines radiomics and deep learning (DL) features is of great significance for the precise preoperative diagnosis of meningioma sinus invasion.

MATERIALS AND METHODS: This study retrospectively collected data from 601 patients with meningioma confirmed by surgical pathology. For each patient, 3948 radiomics features, 12,288 VGG features, 6144 ResNet features, and 3072 DenseNet features were extracted from MRI images. Thus, univariate logistic regression, correlation analysis, and the Boruta algorithm were applied for further feature dimension reduction, selecting radiomics and DL features highly associated with meningioma sinus invasion. Finally, diagnosis models were constructed using the random forest (RF) algorithm. Additionally, the diagnostic performance of different models was evaluated using receiver operating characteristic (ROC) curves, and AUC values of different models were compared using the DeLong test.

RESULTS: Ultimately, 21 features highly associated with meningioma sinus invasion were selected, including 6 radiomics features, 2 VGG features, 7 ResNet features, and 6 DenseNet features. Based on these features, five models were constructed: the radiomics model, VGG model, ResNet model, DenseNet model, and DL-radiomics (DLR) fusion model. This fusion model demonstrated superior diagnostic performance, with AUC values of 0.818, 0.814, and 0.769 in the training set, internal validation set, and independent external validation set, respectively. Furthermore, the results of the DeLong test indicated that there were significant differences between the fusion model and both the radiomics model and the VGG model (p < 0.05).

CONCLUSIONS: The fusion model combining radiomics and DL features exhibits superior diagnostic performance in preoperative diagnosis of meningioma sinus invasion. It is expected to become a powerful tool for clinical surgical plan selection and patient prognosis assessment.

PMID:40022261 | DOI:10.1186/s40644-025-00845-5

Categories: Literature Watch

LETSmix: a spatially informed and learning-based domain adaptation method for cell-type deconvolution in spatial transcriptomics

Fri, 2025-02-28 06:00

Genome Med. 2025 Feb 28;17(1):16. doi: 10.1186/s13073-025-01442-8.

ABSTRACT

Spatial transcriptomics (ST) enables the study of gene expression in spatial context, but many ST technologies face challenges due to limited resolution, leading to cell mixtures at each spot. We present LETSmix to deconvolve cell types by integrating spatial correlations through a tailored LETS filter, which leverages layer annotations, expression similarities, image texture features, and spatial coordinates to refine ST data. Additionally, LETSmix employs a mixup-augmented domain adaptation strategy to address discrepancies between ST and reference single-cell RNA sequencing data. Comprehensive evaluations across diverse ST platforms and tissue types demonstrate its high accuracy in estimating cell-type proportions and spatial patterns, surpassing existing methods (URL: https://github.com/ZhanYangen/LETSmix ).

PMID:40022231 | DOI:10.1186/s13073-025-01442-8

Categories: Literature Watch

Automatic gait EVENT detection in older adults during perturbed walking

Fri, 2025-02-28 06:00

J Neuroeng Rehabil. 2025 Feb 28;22(1):40. doi: 10.1186/s12984-025-01560-9.

ABSTRACT

Accurate detection of gait events in older adults, particularly during perturbed walking, is essential for evaluating balance control and fall risk. Traditional force plate-based methods often face limitations in perturbed walking scenarios due to the difficulty in landing cleanly on the force plates. Subsequently, previous studies have not addressed gait event automatic detection methods for perturbed walking. This study introduces an automated gait event detection method using a bidirectional gated recurrent unit (Bi-GRU) model, leveraging ground reaction force, joint angles, and marker data, for both regular and perturbed walking scenarios from 307 healthy older adults. Our marker-based model achieved over 97% accuracy with a mean error of less than 14 ms in detecting touchdown (TD) and liftoff (LO) events for both walking scenarios. The results highlight the efficacy of kinematic approaches, demonstrating their potential in gait event detection for clinical settings. When integrated with wearable sensors or computer vision techniques, these methods enable real-time, precise monitoring of gait patterns, which is helpful for applying personalized programs for fall prevention. This work takes a significant step forward in automated gait analysis for perturbed walking, offering a reliable method for evaluating gait patterns, balance control, and fall risk in clinical settings.

PMID:40022199 | DOI:10.1186/s12984-025-01560-9

Categories: Literature Watch

A computational spectrometer for the visible, near, and mid-infrared enabled by a single-spinning film encoder

Fri, 2025-02-28 06:00

Commun Eng. 2025 Feb 28;4(1):37. doi: 10.1038/s44172-025-00379-5.

ABSTRACT

Computational spectrometers enable low-cost, in-situ, and rapid spectral analysis, with applications in chemistry, biology, and environmental science. Traditional filter-based spectral encoding approaches typically use filter arrays, complicating the manufacturing process and hindering device consistency. Here we propose a computational spectrometer spanning visible to mid-infrared by combining the Single-Spinning Film Encoder (SSFE) with a deep learning-based reconstruction algorithm. Optimization through particle swarm optimization (PSO) allows for low-correlation and high-complexity spectral responses under different polarizations and spinning angles. The spectrometer demonstrates single-peak resolutions of 0.5 nm, 2 nm, 10 nm, and dual-peak resolutions of 3 nm, 6 nm, 20 nm for the visible, near, and mid-infrared wavelength ranges. Experimentally, it shows an average MSE of 1.05 × 10⁻³ for narrowband spectral reconstruction in the visible wavelength range, with average center-wavelength and linewidth errors of 0.61 nm and 0.56 nm. Additionally, it achieves an overall 81.38% precision for the classification of 220 chemical compounds, showcasing its potential for compact, cost-effective spectroscopic solutions.

PMID:40021937 | DOI:10.1038/s44172-025-00379-5

Categories: Literature Watch

Software defect prediction based on residual/shuffle network optimized by upgraded fish migration optimization algorithm

Fri, 2025-02-28 06:00

Sci Rep. 2025 Feb 28;15(1):7201. doi: 10.1038/s41598-025-91784-5.

ABSTRACT

The study introduces a new method for predicting software defects based on Residual/Shuffle (RS) Networks and an enhanced version of Fish Migration Optimization (UFMO). The overall contribution is to improve the accuracy, and reduce the manual effort needed. The originality of this work rests in the synergic use of deep learning and metaheuristics to train the software code for extraction of semantic and structural properties. The model is tested on a variety of open-source projects, yielding an average accuracy of 93% and surpassing the performance of the state-of-the-art models. The results indicate an overall increase in the precision (78-98%), recall (71-98%), F-measure (72-96%), and Area Under the Curve (AUC) (78-99%). The proposed model is simple and efficient and proves to be effective in identifying potential defects, consequently decreasing the chance of missing these defects and improving the overall quality of the software as opposed to existing approaches. However, the analysis is limited to open-source projects and warrants further evaluation on proprietary software. The study enables a robust and efficient tool for developers. This approach can revolutionize software development practices in order to use artificial intelligence to solve difficult issues presented in software. The model offers high accuracy to reduce the software development cost, which can improve user satisfaction, and enhance the overall quality of software being developed.

PMID:40021906 | DOI:10.1038/s41598-025-91784-5

Categories: Literature Watch

Exploring the application of deep learning methods for polygenic risk score estimation

Fri, 2025-02-28 06:00

Biomed Phys Eng Express. 2025 Feb 28. doi: 10.1088/2057-1976/adbb71. Online ahead of print.

ABSTRACT

&#xD;Polygenic risk scores (PRS) summarise genetic information into a single number with clinical and research uses. Machine learning (ML) has revolutionised multiple fields, however, the impact of ML on PRSs has been less significant. We explore how ML can improve the generation of PRSs.&#xD;Methods:&#xD;We train ML models on known PRSs using UK Biobank data. We explore whether the models can recreate human programmed PRSs, including using a single model to generate multiple PRSs, and ML difficulties in PRS generation. We investigate how ML can compensate for missing data and constraints on performance.&#xD;Results:&#xD;We demonstrate almost perfect generation of multiple PRSs with little loss of performance with reduced quantity of training data. For an example set of missing SNPs the MLP produces predictions that enable separation of cases from population samples with an area under the receiver operating characteristic curve of 0.847 (95% CI: 0.828-0.864) compared to 0.798 (95% CI: 0.779-0.818) for the PRS.&#xD;Conclusions:&#xD;ML can accurately generate PRSs, including with one model for multiple PRSs. The models are transferable and have high longevity. With certain missing SNPs the ML models can improve on PRS generation. Further improvements likely require use of additional input data.&#xD.

PMID:40020248 | DOI:10.1088/2057-1976/adbb71

Categories: Literature Watch

Derivation of an artificial intelligence-based electrocardiographic model for the detection of acute coronary occlusive myocardial infarction

Fri, 2025-02-28 06:00

Arch Cardiol Mex. 2025 Feb 28. doi: 10.24875/ACM.24000195. Online ahead of print.

ABSTRACT

OBJECTIVES: We aimed to assess the performance of an artificial intelligence-electrocardiogram (AI-ECG)-based model capable of detecting acute coronary occlusion myocardial infarction (ACOMI) in the setting of patients with acute coronary syndrome (ACS).

METHODS: This was a prospective, observational, longitudinal, and single-center study including patients with the initial diagnosis of ACS (both ST-elevation acute myocardial infarction [STEMI] & non-ST-segment elevation myocardial infarction [NSTEMI]). To train the deep learning model in recognizing ACOMI, manual digitization of a patient's ECG was conducted using smartphone cameras of varying quality. We relied on the use of convolutional neural networks as the AI models for the classification of ECG examples. ECGs were also independently evaluated by two expert cardiologists blinded to clinical outcomes; each was asked to determine (a) whether the patient had a STEMI, based on universal criteria or (b) if STEMI criteria were not met, to identify any other ECG finding suggestive of ACOMI. ACOMI was defined by coronary angiography findings meeting any of the following three criteria: (a) total thrombotic occlusion, (b) TIMI thrombus grade 2 or higher + TIMI grade flow 1 or less, or (c) the presence of a subocclusive lesion (> 95% angiographic stenosis) with TIMI grade flow < 3. Patients were classified into four groups: STEMI + ACOMI, NSTEMI + ACOMI, STEMI + non-ACOMI, and NSTEMI + non-ACOMI.

RESULTS: For the primary objective of the study, AI outperformed human experts in both NSTEMI and STEMI, with an area under the curve (AUC) of 0.86 (95% confidence interval [CI] 0.75-0.98) for identifying ACOMI, compared with ECG experts using their experience (AUC: 0.33, 95% CI 0.17-0.49) or under universal STEMI criteria (AUC: 0.50, 95% CI 0.35-0.54), (p value for AUC receiver operating characteristic comparison < 0.001). The AI model demonstrated a PPV of 0.84 and an NPV of 1.0.

CONCLUSION: Our AI-ECG model demonstrated a higher diagnostic precision for the detection of ACOMI compared with experts and the use of STEMI criteria. Further research and external validation are needed to understand the role of AI-based models in the setting of ACS.

PMID:40020200 | DOI:10.24875/ACM.24000195

Categories: Literature Watch

Phantom-metasurface cooperative system trained by a deep learning network driven by a bound state for a magnetic resonance-enhanced system

Fri, 2025-02-28 06:00

Opt Lett. 2025 Mar 1;50(5):1723-1726. doi: 10.1364/OL.546727.

ABSTRACT

With the development of medical imaging technology, magnetic resonance imaging (MRI) has become an important tool for diagnosing and monitoring a variety of diseases. However, traditional MRI techniques are limited in terms of imaging speed and resolution. In this study, we developed an efficient body mode metasurface composite MRI enhancement system based on deep learning network training and realized the design and control of metasurface in the MHz band. Firstly, forward neural network is used to predict the electromagnetic response characteristics quickly. On this basis, the network is reverse-designed and the structural parameters of the metasurface are predicted. The experimental results show that the combination of deep neural network and electromagnetic metasurface significantly improves the design efficiency of metasurface and has great application potential in the MRI system.

PMID:40020024 | DOI:10.1364/OL.546727

Categories: Literature Watch

Physics-driven deep learning for high-fidelity photon-detection ghost imaging

Fri, 2025-02-28 06:00

Opt Lett. 2025 Mar 1;50(5):1719-1722. doi: 10.1364/OL.541330.

ABSTRACT

Single-photon detection has significant potential in the field of imaging due to its high sensitivity and has been widely applied across various domains. However, achieving high spatial and depth resolution through scattering media remains challenging because of the limitations of low light intensity, high background noise, and inherent time jitter of the detector. This paper proposes a physics-driven, learning-based photon-detection ghost imaging method to address these challenges. By co-designing the computational ghost imaging system and the network, we integrate imaging and reconstruction more closely to surpass the physical resolution limitations. Fringe patterns are employed to encode the depth information of the object into different channels of an image cube. A specialized depth fusion network with attention mechanisms is then designed to extract inter-depth correlation features, enabling super-resolution reconstruction at 256 × 256 pixels. Experimental results demonstrate that the proposed method presents superior imaging performance across various scenarios, offering a more compact and cost-effective alternative for photon-detection imaging.

PMID:40020023 | DOI:10.1364/OL.541330

Categories: Literature Watch

Comparing the performance of a large language model and naive human interviewers in interviewing children about a witnessed mock-event

Fri, 2025-02-28 06:00

PLoS One. 2025 Feb 28;20(2):e0316317. doi: 10.1371/journal.pone.0316317. eCollection 2025.

ABSTRACT

PURPOSE: The present study compared the performance of a Large Language Model (LLM; ChatGPT) and human interviewers in interviewing children about a mock-event they witnessed.

METHODS: Children aged 6-8 (N = 78) were randomly assigned to the LLM (n = 40) or the human interviewer condition (n = 38). In the experiment, the children were asked to watch a video filmed by the researchers that depicted behavior including elements that could be misinterpreted as abusive in other contexts, and then answer questions posed by either an LLM (presented by a human researcher) or a human interviewer.

RESULTS: Irrespective of condition, recommended (vs. not recommended) questions elicited more correct information. The LLM posed fewer questions overall, but no difference in the proportion of the questions recommended by the literature. There were no differences between the LLM and human interviewers in unique correct information elicited but questions posed by LLM (vs. humans) elicited more unique correct information per question. LLM (vs. humans) also elicited less false information overall, but there was no difference in false information elicited per question.

CONCLUSIONS: The findings show that the LLM was competent in formulating questions that adhere to best practice guidelines while human interviewers asked more questions following up on the child responses in trying to find out what the children had witnessed. The results indicate LLMs could possibly be used to support child investigative interviewers. However, substantial further investigation is warranted to ascertain the utility of LLMs in more realistic investigative interview settings.

PMID:40019879 | DOI:10.1371/journal.pone.0316317

Categories: Literature Watch

MultiKD-DTA: Enhancing Drug-Target Affinity Prediction Through Multiscale Feature Extraction

Fri, 2025-02-28 06:00

Interdiscip Sci. 2025 Feb 28. doi: 10.1007/s12539-025-00697-4. Online ahead of print.

ABSTRACT

The discovery and development of novel pharmaceutical agents is characterized by high costs, lengthy timelines, and significant safety concerns. Traditional drug discovery involves pharmacologists manually screening drug molecules against protein targets, focusing on binding within protein cavities. However, this manual process is slow and inherently limited. Given these constraints, the use of deep learning techniques to predict drug-target interaction (DTI) affinities is both significant and promising for future applications. This paper introduces an innovative deep learning architecture designed to enhance the prediction of DTI affinities. The model ingeniously combines graph neural networks, pre-trained large-scale protein models, and attention mechanisms to improve performance. In this framework, molecular structures are represented as graphs and processed through graph neural networks and multiscale convolutional networks to facilitate feature extraction. Simultaneously, protein sequences are encoded using pre-trained ESM-2 large models and processed with bidirectional long short-term memory networks. Subsequently, the molecular and protein embeddings derived from these processes are integrated within a fusion module to compute affinity scores. Experimental results demonstrate that our proposed model outperforms existing methods on two publicly available datasets.

PMID:40019659 | DOI:10.1007/s12539-025-00697-4

Categories: Literature Watch

A novel approach for estimating postmortem intervals under varying temperature conditions using pathology images and artificial intelligence models

Fri, 2025-02-28 06:00

Int J Legal Med. 2025 Feb 28. doi: 10.1007/s00414-025-03447-9. Online ahead of print.

ABSTRACT

Estimating the postmortem interval (PMI) is a critical yet complex task in forensic investigations, with accurate and timely determination playing a key role in case resolution and legal outcomes. Traditional methods often suffer from environmental variability and subjective biases, emphasizing the need for more reliable and objective approaches. In this study, we present a novel predictive model for PMI estimation, introduced here for the first time, that leverages pathological tissue images and artificial intelligence (AI). The model is designed to perform under three temperature conditions: 25 °C, 37 °C, and 4 °C. Using a ResNet50 neural network, patch-level images were analyzed to extract deep learning-derived features, which were integrated with machine learning algorithms for whole slide image (WSI) classification. The model achieved strong performance, with micro and macro AUC values of at least 0.949 at the patch-level and 0.800 at the WSI-level in both training and testing sets. In external validation, micro and macro AUC values at the patch-level exceeded 0.960. These results highlight the potential of AI to improve the accuracy and efficiency of PMI estimation. As AI technology continues to advance, this approach holds promise for enhancing forensic investigations and supporting more precise case resolutions.

PMID:40019556 | DOI:10.1007/s00414-025-03447-9

Categories: Literature Watch

Artificial intelligence in otorhinolaryngology: current trends and application areas

Fri, 2025-02-28 06:00

Eur Arch Otorhinolaryngol. 2025 Feb 17. doi: 10.1007/s00405-025-09272-5. Online ahead of print.

ABSTRACT

PURPOSE: This study aims to perform a bibliometric analysis of scientific research on the use of artificial intelligence (AI) in the field of Otorhinolaryngology (ORL), with a specific focus on identifying emerging AI trend topics within this discipline.

METHODS: A total of 498 articles on AI in ORL, published between 1982 and 2024, were retrieved from the Web of Science database. Various bibliometric techniques, including trend keyword analysis and factor analysis, were applied to analyze the data.

RESULTS: The most prolific journal was the European Archives of Oto-Rhino-Laryngology (n = 67). The USA (n = 200) and China (n = 61) were the most productive countries in AI-related ORL research. The most productive institutions were Harvard University / Harvard Medical School (n = 71). The leading authors in this field were Lechien JR. (n = 18) and Rameau A. (n = 17). The most frequently used keywords in the AI research were cochlear implant, head and neck cancer, magnetic resonance imaging (MRI), hearing loss, patient education, diagnosis, radiomics, surgery, hearing aids, laryngology ve otitis media. Recent trends in otorhinolaryngology research reflect a dynamic focus, progressing from hearing-related technologies such as hearing aids and cochlear implants in earlier years, to diagnostic innovations like audiometry, psychoacoustics, and narrow band imaging. The emphasis has recently shifted toward advanced applications of MRI, radiomics, and computed tomography (CT) for conditions such as head and neck cancer, chronic rhinosinusitis, laryngology, and otitis media. Additionally, increasing attention has been given to patient education, quality of life, and prognosis, underscoring a holistic approach to diagnosis, surgery, and treatment in otorhinolaryngology.

CONCLUSION: AI has significantly impacted the field of ORL, especially in diagnostic imaging and therapeutic planning. With advancements in MRI and CT-based technologies, AI has proven to enhance disease detection and management. The future of AI in ORL suggests a promising path toward improving clinical decision-making, patient care, and healthcare efficiency.

PMID:40019544 | DOI:10.1007/s00405-025-09272-5

Categories: Literature Watch

Pd-Modified Microneedle Array Sensor Integration with Deep Learning for Predicting Silica Aerogel Properties in Real Time

Fri, 2025-02-28 06:00

ACS Appl Mater Interfaces. 2025 Feb 28. doi: 10.1021/acsami.4c17680. Online ahead of print.

ABSTRACT

The continuous global effort to predict material properties through artificial intelligence has predominantly focused on utilizing material stoichiometry or structures in deep learning models. This study aims to predict material properties using electrochemical impedance data, along with frequency and time parameters, that can be obtained during processing stages. The target material, silica aerogel, is widely recognized for its lightweight structure and excellent insulating properties, which are attributed to its large surface area and pore size. However, production is often delayed due to the prolonged aging process. Real-time prediction of material properties during processing can significantly enhance process optimization and monitoring. In this study, we developed a system to predict the physical properties of silica aerogel, specifically pore diameter, pore volume, and surface area. This system integrates a 3 × 3 array Pd/Au sensor, which exhibits high sensitivity to varying pH levels during aerogel synthesis and is capable of acquiring a large data set (impedance, frequency, time) in real-time. The collected data is then processed through a deep neural network algorithm. Because the system is trained with data obtained during the processing stage, it enables real-time predictions of the critical properties of silica aerogel, thus facilitating process optimization and monitoring. The final performance evaluation demonstrated an optimal alignment between true and predicted values for silica aerogel properties, with a mean absolute percentage error of approximately 0.9%. This approach holds great promise for significantly improving the efficiency and effectiveness of silica aerogel production by providing accurate real-time predictions.

PMID:40019213 | DOI:10.1021/acsami.4c17680

Categories: Literature Watch

Quantifying Facial Gestures Using Deep Learning in a New World Monkey

Fri, 2025-02-28 06:00

Am J Primatol. 2025 Mar;87(3):e70013. doi: 10.1002/ajp.70013.

ABSTRACT

Facial gestures are a crucial component of primate multimodal communication. However, current methodologies for extracting facial data from video recordings are labor-intensive and prone to human subjectivity. Although automatic tools for this task are still in their infancy, deep learning techniques are revolutionizing animal behavior research. This study explores the distinctiveness of facial gestures in cotton-top tamarins, quantified using markerless pose estimation algorithms. From footage of captive individuals, we extracted and manually labeled frames to develop a model that can recognize a custom set of landmarks positioned on the face of the target species. The trained model predicted landmark positions and subsequently transformed them into distance matrices representing landmarks' spatial distributions within each frame. We employed three competitive machine learning classifiers to assess the ability to automatically discriminate facial configurations that cooccur with vocal emissions and are associated with different behavioral contexts. Initial analysis showed correct classification rates exceeding 80%, suggesting that voiced facial configurations are highly distinctive from unvoiced ones. Our findings also demonstrated varying context specificity of facial gestures, with the highest classification accuracy observed during yawning, social activity, and resting. This study highlights the potential of markerless pose estimation for advancing the study of primate multimodal communication, even in challenging species such as cotton-top tamarins. The ability to automatically distinguish facial gestures in different behavioral contexts represents a critical step in developing automated tools for extracting behavioral cues from raw video data.

PMID:40019116 | DOI:10.1002/ajp.70013

Categories: Literature Watch

Deep learning for named entity recognition in Turkish radiology reports

Fri, 2025-02-28 06:00

Diagn Interv Radiol. 2025 Feb 28. doi: 10.4274/dir.2025.243100. Online ahead of print.

ABSTRACT

PURPOSE: The primary objective of this research is to enhance the accuracy and efficiency of information extraction from radiology reports. In addressing this objective, the study aims to develop and evaluate a deep learning framework for named entity recognition (NER).

METHODS: We used a synthetic dataset of 1,056 Turkish radiology reports created and labeled by the radiologists in our research team. Due to privacy concerns, actual patient data could not be used; however, the synthetic reports closely mimic genuine reports in structure and content. We employed the four-stage DYGIE++ model for the experiments. First, we performed token encoding using four bidirectional encoder representations from transformers (BERT) models: BERTurk, BioBERTurk, PubMedBERT, and XLM-RoBERTa. Second, we introduced adaptive span enumeration, considering the word count of a sentence in Turkish. Third, we adopted span graph propagation to generate a multidirectional graph crucial for coreference resolution. Finally, we used a two-layered feed-forward neural network to classify the named entity.

RESULTS: The experiments conducted on the labeled dataset showcase the approach's effectiveness. The study achieved an F1 score of 80.1 for the NER task, with the BioBERTurk model, which is pre-trained on Turkish Wikipedia, radiology reports, and biomedical texts, proving to be the most effective of the four BERT models used in the experiment.

CONCLUSION: We show how different dataset labels affect the model's performance. The results demonstrate the model's ability to handle the intricacies of Turkish radiology reports, providing a detailed analysis of precision, recall, and F1 scores for each label. Additionally, this study compares its findings with related research in other languages.

CLINICAL SIGNIFICANCE: Our approach provides clinicians with more precise and comprehensive insights to improve patient care by extracting relevant information from radiology reports. This innovation in information extraction streamlines the diagnostic process and helps expedite patient treatment decisions.

PMID:40018795 | DOI:10.4274/dir.2025.243100

Categories: Literature Watch

Diagnostic accuracy of convolutional neural network algorithms to distinguish gastrointestinal obstruction on conventional radiographs in a pediatric population

Fri, 2025-02-28 06:00

Diagn Interv Radiol. 2025 Feb 28. doi: 10.4274/dir.2025.242950. Online ahead of print.

ABSTRACT

PURPOSE: Gastrointestinal (GI) dilatations are frequently observed in radiographs of pediatric patients who visit emergency departments with acute symptoms such as vomiting, pain, constipation, or diarrhea. Timely and accurate differentiation of whether there is an obstruction requiring surgery in these patients is crucial to prevent complications such as necrosis and perforation, which can lead to death. In this study, we aimed to use convolutional neural network (CNN) models to differentiate healthy children with normal intestinal gas distribution in abdominal radiographs from those with GI dilatation or obstruction. We also aimed to distinguish patients with obstruction requiring surgery and those with other GI dilatation or ileus.

METHODS: Abdominal radiographs of patients with a surgical, clinical, and/or laboratory diagnosis of GI diseases with GI dilatation were retrieved from our institution's Picture Archiving and Communication System archive. Additionally, abdominal radiographs performed to detect abnormalities other than GI disorders were collected to form a control group. The images were labeled with three tags according to their groups: surgically-corrected dilatation (SD), inflammatory/infectious dilatation (ID), and normal. To determine the impact of standardizing the imaging area on the model's performance, an additional dataset was created by applying an automated cropping process. Five CNN models with proven success in image analysis (ResNet50, InceptionResNetV2, Xception, EfficientNetV2L, and ConvNeXtXLarge) were trained, validated, and tested using transfer learning.

RESULTS: A total of 540 normal, 298 SD, and 314 ID were used in this study. In the differentiation between normal and abnormal images, the highest accuracy rates were achieved with ResNet50 (93.3%) and InceptionResNetV2 (90.6%) CNN models. Then, after using automated cropping preprocessing, the highest accuracy rates were achieved with ConvNeXtXLarge (96.9%), ResNet50 (95.5%), and InceptionResNetV2 (95.5%). The highest accuracy in the differentiation between SD and ID was achieved with EfficientNetV2L (94.6%).

CONCLUSION: Deep learning models can be integrated into radiographs located in the emergency departments as a decision support system with high accuracy rates in pediatric GI obstructions by immediately alerting the physicians about abnormal radiographs and possible etiologies.

CLINICAL SIGNIFICANCE: This paper describes a novel area of utilization of well-known deep learning algorithm models. Although some studies in the literature have shown the efficiency of CNN models in identifying small bowel obstruction with high accuracy for the adult population or some specific diseases, our study is unique for the pediatric population and for evaluating the requirement of surgical versus medical treatment.

PMID:40018794 | DOI:10.4274/dir.2025.242950

Categories: Literature Watch

Pages