Deep learning

Unveiling pathology-related predictive uncertainty of glomerular lesion recognition using prototype learning

Thu, 2025-01-02 06:00

J Biomed Inform. 2024 Dec 31:104745. doi: 10.1016/j.jbi.2024.104745. Online ahead of print.

ABSTRACT

OBJECTIVE: Recognizing glomerular lesions is essential in diagnosing chronic kidney disease. However, deep learning faces challenges due to the lesion heterogeneity, superposition, progression, and tissue incompleteness, leading to uncertainty in model predictions. Therefore, it is crucial to analyze pathology-related predictive uncertainty in glomerular lesion recognition and unveil its relationship with pathological properties and its impact on model performance.

METHODS: This paper presents a novel framework for pathology-related predictive uncertainty analysis towards glomerular lesion recognition, including prototype learning based predictive uncertainty estimation, pathology-characterized correlation analysis and weight-redistributed prediction rectification. The prototype learning based predictive uncertainty estimation includes deep prototyping, affinity embedding, and multi-dimensional uncertainty fusion. The pathology-characterized correlation analysis is the first to use expert-based and learning- based approach to construct the pathology-related characterization of lesions and tissues. The weight-redistributed prediction rectification module performs reweighting- based lesion recognition.

RESULTS: To validate the performance, extensive experiments were conducted. Based on the Spearman and Pearson correlation analysis the proposed framework enables more efficient correlation analysis, and strong correlation with pathology-related characterization can be achieved (c index > 0.6 and p < 0.01). Furthermore, the prediction rectification module demonstrated improved lesion recognition performance across most metrics, with enhancements of up to 6.36 %.

CONCLUSION: The proposed predictive uncertainty analysis in glomerular lesion recognition offers a valuable approach for assessing computational pathology's predictive uncertainty from a pathology-related perspective.

SIGNIFICANCE: The paper provides a solution for pathology-related predictive uncertainty estimation in algorithm development and clinical practice.

PMID:39746430 | DOI:10.1016/j.jbi.2024.104745

Categories: Literature Watch

Malignancy risk stratification for pulmonary nodules: comparing a deep learning approach to multiparametric statistical models in different disease groups

Thu, 2025-01-02 06:00

Eur Radiol. 2025 Jan 2. doi: 10.1007/s00330-024-11256-8. Online ahead of print.

ABSTRACT

OBJECTIVES: Incidentally detected pulmonary nodules present a challenge in clinical routine with demand for reliable support systems for risk classification. We aimed to evaluate the performance of the lung-cancer-prediction-convolutional-neural-network (LCP-CNN), a deep learning-based approach, in comparison to multiparametric statistical methods (Brock model and Lung-RADS®) for risk classification of nodules in cohorts with different risk profiles and underlying pulmonary diseases.

MATERIALS AND METHODS: Retrospective analysis was conducted on non-contrast and contrast-enhanced CT scans containing pulmonary nodules measuring 5-30 mm. Ground truth was defined by histology or follow-up stability. The final analysis was performed on 297 patients with 422 eligible nodules, of which 105 nodules were malignant. Classification performance of the LCP-CNN, Brock model, and Lung-RADS® was evaluated in terms of diagnostic accuracy measurements including ROC-analysis for different subcohorts (total, screening, emphysema, and interstitial lung disease).

RESULTS: LCP-CNN demonstrated superior performance compared to the Brock model in total and screening cohorts (AUC 0.92 (95% CI: 0.89-0.94) and 0.93 (95% CI: 0.89-0.96)). Superior sensitivity of LCP-CNN was demonstrated compared to the Brock model and Lung-RADS® in total, screening, and emphysema cohorts for a risk threshold of 5%. Superior sensitivity of LCP-CNN was also shown across all disease groups compared to the Brock model at a threshold of 65%, compared to Lung-RADS® sensitivity was better or equal. No significant differences in the performance of LCP-CNN were found between subcohorts.

CONCLUSION: This study offers further evidence of the potential to integrate deep learning-based decision support systems into pulmonary nodule classification workflows, irrespective of the individual patient risk profile and underlying pulmonary disease.

KEY POINTS: Question Is a deep-learning approach (LCP-CNN) superior to multiparametric models (Brock model, Lung-RADS®) in classifying pulmonary nodule risk across varied patient profiles? Findings LCP-CNN shows superior performance in risk classification of pulmonary nodules compared to multiparametric models with no significant impact on risk profiles and structural pulmonary diseases. Clinical relevance LCP-CNN offers efficiency and accuracy, addressing limitations of traditional models, such as variations in manual measurements or lack of patient data, while producing robust results. Such approaches may therefore impact clinical work by complementing or even replacing current approaches.

PMID:39747589 | DOI:10.1007/s00330-024-11256-8

Categories: Literature Watch

SAILOR: perceptual anchoring for robotic cognitive architectures

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):113. doi: 10.1038/s41598-024-84071-2.

ABSTRACT

Symbolic anchoring is an important topic in robotics, as it enables robots to obtain symbolic knowledge from the perceptual information acquired through their sensors and maintain the link between that knowledge and the sensory data. In cognitive-based robots, this process of transforming sub-symbolic data generated by sensors to obtain and maintain symbolic knowledge is still an open problem. To address this issue, this paper presents SAILOR, a framework for symbolic anchoring integrated into ROS 2. SAILOR aims to maintain the link between symbolic data and perceptual data in real robots over time. It provides a semantic world modeling approach using two deep learning-based sub-symbolic robotic skills: object recognition and matching function. The object recognition skill allows the robot to recognize and identify objects in its environment, while the matching function enables the robot to decide if new perceptual data corresponds to existing symbolic data. This paper describes the proposed method and the development of the framework, as well as its integration in MERLIN2 (a hybrid cognitive architecture fully functional in robots running ROS 2) and the validation of SAILOR using public datasets and a real-world scenario.

PMID:39747469 | DOI:10.1038/s41598-024-84071-2

Categories: Literature Watch

Deep learning-based discovery of compounds for blood pressure lowering effects

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):54. doi: 10.1038/s41598-024-83924-0.

ABSTRACT

The hypotensive side effects caused by drugs during their use have been a vexing issue. Recent studies have found that deep learning can effectively predict the biological activity of compounds by mining patterns and rules in the data, providing a potential solution for identifying drug side effects. In this study, we established a deep learning-based predictive model, utilizing a data set comprised of compounds known to either elevate or lower blood pressure. Subsequently, the trained model was used to predict the blood pressure-lowering effects of 26,000 compounds. Based on the predicted results, we randomly selected 50 molecules for validation and compared them with literature reports. The results showed that the predictions for 30 molecules were consistent with literature reports, with known antihypertensive drugs such as reserpine, guanethidine, and mecamylamine ranking at the top. We further selected 10 of these molecules and 3 related protein targets for molecular docking, and the docking results indirectly confirmed the model's accuracy. Ultimately, we discovered and validated that salaprinol significantly inhibits ACE1 activity and lowers canine blood pressure. In summary, we have established a highly accurate activity prediction model and confirmed its accuracy in predicting potential blood pressure-lowering compounds, which is expected to help patients avoid hypotensive side effects during clinical medication and also provide significant assistance in the discovery of antihypertensive drugs.

PMID:39747442 | DOI:10.1038/s41598-024-83924-0

Categories: Literature Watch

A novel deep synthesis-based insider intrusion detection (DS-IID) model for malicious insiders and AI-generated threats

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):207. doi: 10.1038/s41598-024-84673-w.

ABSTRACT

Insider threats pose a significant challenge to IT security, particularly with the rise of generative AI technologies, which can create convincing fake user profiles and mimic legitimate behaviors. Traditional intrusion detection systems struggle to differentiate between real and AI-generated activities, creating vulnerabilities in detecting malicious insiders. To address this challenge, this paper introduces a novel Deep Synthesis Insider Intrusion Detection (DS-IID) model. The model employs deep feature synthesis to automatically generate detailed user profiles from event data and utilizes binary deep learning for accurate threat identification. The DS-IID model addresses three key issues: it (i) detects malicious insiders using supervised learning, (ii) evaluates the effectiveness of generative algorithms in replicating real user profiles, and (iii) distinguishes between real and synthetic abnormal user profiles. To handle imbalanced data, the model uses on-the-fly weighted random sampling. Tested on the CERT insider threat dataset, the DS-IID achieved 97% accuracy and an AUC of 0.99. Moreover, the model demonstrates strong performance in differentiating real from AI-generated (synthetic) threats, achieving over 99% accuracy on optimally generated data. While primarily evaluated on synthetic datasets, the high accuracy of the DS-IID model suggests its potential as a valuable tool for real-world cybersecurity applications.

PMID:39747424 | DOI:10.1038/s41598-024-84673-w

Categories: Literature Watch

A lightweight weed detection model for cotton fields based on an improved YOLOv8n

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):457. doi: 10.1038/s41598-024-84748-8.

ABSTRACT

In modern agriculture, the proliferation of weeds in cotton fields poses a significant threat to the healthy growth and yield of crops. Therefore, efficient detection and control of cotton field weeds are of paramount importance. In recent years, deep learning models have shown great potential in the detection of cotton field weeds, achieving high-precision weed recognition. However, existing deep learning models, despite their high accuracy, often have complex computations and high resource consumption, making them difficult to apply in practical scenarios. To address this issue, developing efficient and lightweight detection methods for weed recognition in cotton fields is crucial for effective weed control. This study proposes the YOLO-Weed Nano algorithm based on the improved YOLOv8n model. First, the Depthwise Separable Convolution (DSC) structure is used to improve the HGNetV2 network, creating the DS_HGNetV2 network to replace the backbone of the YOLOv8n model. Secondly, the Bidirectional Feature Pyramid Network (BiFPN) is introduced to enhance the feature fusion layer, further optimizing the model's ability to recognize weed features in complex backgrounds. Finally, a lightweight detection head, LiteDetect, suitable for the BiFPN structure, is designed to streamline the model structure and reduce computational load. Experimental results show that compared to the original YOLOv8n model, YOLO-Weed Nano improves mAP by 1%, while reducing the number of parameters, computation, and weights by 63.8%, 42%, and 60.7%, respectively.

PMID:39747358 | DOI:10.1038/s41598-024-84748-8

Categories: Literature Watch

Drug discovery and mechanism prediction with explainable graph neural networks

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):179. doi: 10.1038/s41598-024-83090-3.

ABSTRACT

Apprehension of drug action mechanism is paramount for drug response prediction and precision medicine. The unprecedented development of machine learning and deep learning algorithms has expedited the drug response prediction research. However, existing methods mainly focus on forward encoding of drugs, which is to obtain an accurate prediction of the response levels, but omitted to decipher the reaction mechanism between drug molecules and genes. We propose the eXplainable Graph-based Drug response Prediction (XGDP) approach that achieves a precise drug response prediction and reveals the comprehensive mechanism of action between drugs and their targets. XGDP represents drugs with molecular graphs, which naturally preserve the structural information of molecules and a Graph Neural Network module is applied to learn the latent features of molecules. Gene expression data from cancer cell lines are incorporated and processed by a Convolutional Neural Network module. A couple of deep learning attribution algorithms are leveraged to interpret interactions between drug molecular features and genes. We demonstrate that XGDP not only enhances the prediction accuracy compared to pioneering works but is also capable of capturing the salient functional groups of drugs and interactions with significant genes of cancer cells.

PMID:39747341 | DOI:10.1038/s41598-024-83090-3

Categories: Literature Watch

Personalized tourism recommendation model based on temporal multilayer sequential neural network

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):382. doi: 10.1038/s41598-024-84581-z.

ABSTRACT

Personalized tourism has recently become an increasingly popular mode of travel. Effective personalized route recommendations must consider numerous complex factors, including the vast historical trajectory of tourism, individual traveler preferences, and real-time environmental conditions. However, the large temporal and spatial spans of trajectory data pose significant challenges to achieving high relevance and accuracy in personalized route recommendation systems. This study addresses these challenges by proposing a personalized tourism route recommendation model, the Temporal Multilayer Sequential Neural Network (TMS-Net). The fixed-length trajectory segmentation method designed in TMS-Net can adaptively adjust the segmentation length of tourist trajectories, effectively addressing the issue of large spatiotemporal spans by integrating tourist behavior characteristics and route complexity. The self-attention mechanism incorporating relative positional information enhances the model's ability to capture the relationships between different paths within a tourism route by merging position encoding and distance information. Additionally, the multilayer Long Short-Term Memory neural network module, built through hierarchical time series modeling, deeply captures the complex temporal dependencies in travel routes, improving the relevance of the recommendation results and the ability to recognize long-duration travel behaviors. The TMS-Net model was trained on over six million trajectory data points from Chengdu City, Sichuan Province, spanning January 2016 to December 2022. The experimental results indicated that the optimal trajectory segmentation interval ranged from 0.8 to 1.2 h. The model achieved a recommendation accuracy of 88.6% and a Haversine distance error of 1.23, demonstrating its ability to accurately identify tourist points of interest and provide highly relevant recommendations. This study demonstrates the potential of TMS-Net to improve personalized tourism experiences significantly and offers new methodological insights for personalized travel recommendations.

PMID:39747325 | DOI:10.1038/s41598-024-84581-z

Categories: Literature Watch

The usefulness of automated high frequency ultrasound image analysis in atopic dermatitis staging

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):163. doi: 10.1038/s41598-024-84051-6.

ABSTRACT

The last decades have brought an interest in ultrasound applications in dermatology. Especially in the case of atopic dermatitis, where the formation of a subepidermal low echogenic band (SLEB) may serve as an independent indicator of the effects of treatment, the use of ultrasound is of particular interest. This study proposes and evaluates the computer-aided diagnosis method for assessing atopic dermatitis (AD). The fully automated image processing framework combines advanced machine learning techniques for fast, reliable, and repeatable HFUS image analysis, supporting clinical decisions. The proposed methodology comprises accurate SLEB segmentation followed by a classification step. The data set includes 20 MHz images of 80 patients diagnosed with AD according to Hanifin and Rajka criteria, which were evaluated before and after treatment. The ground true labels- clinical evaluation based on Investigator Global Assessment index (IGA score) together with ultrasound skin examination was performed. For reliable analysis, in further experiments, two experts annotated the HFUS images twice in two-week intervals. The analysis aimed to verify whether the fully automated method can classify the HFUS images at the expert level. The Dice coefficient values for segmentation reached 0.908 for SLEB and 0.936 for the entry echo layer. The accuracy of SLEB presence detection results (IGA0) is equal to 98% and slightly outperforms the experts' assessment, which reaches 96%. The overall accuracy of the AD assessment was equal to 69% (Cohen's kappa 0.78) and was comparable with the experts' assessment, ranging between 64% and 70% (Cohen's kappa 0.73-0.79). The results indicate that the automated method can be applied to AD assessment, and its combination with standard diagnosis may benefit repeatable analysis and a better understanding of the processes that take place within the skin and aid treatment monitoring.

PMID:39747292 | DOI:10.1038/s41598-024-84051-6

Categories: Literature Watch

Varying pixel resolution significantly improves deep learning-based carotid plaque histology segmentation

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):139. doi: 10.1038/s41598-024-83948-6.

ABSTRACT

Carotid plaques-the buildup of cholesterol, calcium, cellular debris, and fibrous tissues in carotid arteries-can rupture, release microemboli into the cerebral vasculature and cause strokes. The likelihood of a plaque rupturing is thought to be associated with its composition (i.e. lipid, calcium, hemorrhage and inflammatory cell content) and the mechanical properties of the plaque. Automating and digitizing histopathological images of these plaques into tissue specific (lipid and calcified) regions can help us compare histologic findings to in vivo imaging and thereby enable us to optimize medical treatments or interventions for patients based on the composition of plaques. Lack of public datasets and the hypocellular nature of plaques have made applying deep learning to this task difficult. To address this, we sampled 1944 regions of interests from 323 whole slide images and drastically varied their pixel resolution from [Formula: see text] to [Formula: see text] as we anticipated that varying the pixel resolution of histology images can provide neural networks more 'context' that pathologists also rely on. We were able to train Mask R-CNN using regions of interests with varied pixel resolution, with a [Formula: see text] increase in pixel accuracy versus training with patches. The model achieved F1 scores of [Formula: see text] for calcified regions, [Formula: see text] for lipid core with fibrinous material and cholesterol crystals, and [Formula: see text] for fibrous regions, as well as a pixel accuracy of [Formula: see text]. While the F1 score was not calculated for lumen, qualitative results illustrate the model's ability to predict lumen. Hemorrhage was excluded as a class since only one out of 34 carotid endarterectomy specimens had sufficient hemorrhage for annotation.

PMID:39747244 | DOI:10.1038/s41598-024-83948-6

Categories: Literature Watch

A deep learning method based on multi-scale fusion for noise-resistant coal-gangue recognition

Thu, 2025-01-02 06:00

Sci Rep. 2025 Jan 2;15(1):101. doi: 10.1038/s41598-024-83604-z.

ABSTRACT

Coal-gangue recognition technology plays an important role in the intelligent realization of integrated working faces and coal quality improvement. However, the existing methods are easily affected by high dust, noise, and other disturbances, resulting in unstable recognition results that make it difficult to meet the needs of industrial applications. To realize accurate recognition of coal-gangue in noisy environments, this paper proposes an end-to-end multi-scale feature fusion convolutional neural network (MCNN-BILSTM) based gangue recognition method, which can automatically learn and fuse complementary information from multiple signal components of vibration signals. It combines traditional filtering methods and the idea of multi-scale learning, which can expand the breadth and depth of the feature learning process. the breadth and depth of the feature learning process. Moreover, to strengthen the expression of key features, a feature weighting method based on the attention mechanism is combined to give adaptive weights to different features. Finally, the experimental platform of a tail beam of coal-gangue impact hydraulic support is built, and several comparative experiments are carried out. The comprehensive comparison experiments show that the method shows strong adaptability, robustness, and noise resistance under various complex noise environments, and is suitable for complex practical industrial sites.

PMID:39747222 | DOI:10.1038/s41598-024-83604-z

Categories: Literature Watch

Two-Dimensional Transition Metal Dichalcogenides: A Theory and Simulation Perspective

Thu, 2025-01-02 06:00

Chem Rev. 2025 Jan 2. doi: 10.1021/acs.chemrev.4c00628. Online ahead of print.

ABSTRACT

Two-dimensional transition metal dichalcogenides (2D TMDs) are a promising class of functional materials for fundamental physics explorations and applications in next-generation electronics, catalysis, quantum technologies, and energy-related fields. Theory and simulations have played a pivotal role in recent advancements, from understanding physical properties and discovering new materials to elucidating synthesis processes and designing novel devices. The key has been developments in ab initio theory, deep learning, molecular dynamics, high-throughput computations, and multiscale methods. This review focuses on how theory and simulations have contributed to recent progress in 2D TMDs research, particularly in understanding properties of twisted moiré-based TMDs, predicting exotic quantum phases in TMD monolayers and heterostructures, understanding nucleation and growth processes in TMD synthesis, and comprehending electron transport and characteristics of different contacts in potential devices based on TMD heterostructures. The notable achievements provided by theory and simulations are highlighted, along with the challenges that need to be addressed. Although 2D TMDs have demonstrated potential and prototype devices have been created, we conclude by highlighting research areas that demand the most attention and how theory and simulation might address them and aid in attaining the true potential of 2D TMDs toward commercial device realizations.

PMID:39746214 | DOI:10.1021/acs.chemrev.4c00628

Categories: Literature Watch

Leveraging Artificial Intelligence/Machine Learning Models to Identify Potential Palliative Care Beneficiaries: A Systematic Review

Thu, 2025-01-02 06:00

J Gerontol Nurs. 2025 Jan;51(1):7-14. doi: 10.3928/00989134-20241210-01. Epub 2025 Jan 1.

ABSTRACT

PURPOSE: The current review examined the application of artificial intelligence (AI) and machine learning (ML) techniques in palliative care, specifically focusing on models used to identify potential beneficiaries of palliative services among individuals with chronic and terminal illnesses.

METHODS: A systematic review was conducted across four electronic databases. Five studies met inclusion criteria, all of which applied AI/ML models to predict outcomes relevant to palliative care, such as mortality or the need for services.

RESULTS: Of 1,504 studies screened, five studies used supervised ML algorithms, whereas one used natural language processing with a deep learning model to identify potential palliative care candidates. The most common AI/ML algorithms included neural network-based models, logistic regression, and tree-based models.

CONCLUSION: AI and ML models offer promising avenues for identifying palliative care beneficiaries. As AI continues to evolve, its potential to reshape palliative care through early identification is significant, providing opportunities for timely and targeted care interventions. [Journal of Gerontological Nursing, 51(1), 7-14.].

PMID:39746126 | DOI:10.3928/00989134-20241210-01

Categories: Literature Watch

Incident duration prediction through integration of uncertainty and risk factor evaluation: A San Francisco incidents case study

Thu, 2025-01-02 06:00

PLoS One. 2025 Jan 2;20(1):e0316289. doi: 10.1371/journal.pone.0316289. eCollection 2025.

ABSTRACT

Predicting incident duration and understanding incident types are essential in traffic management for resource optimization and disruption minimization. Precise predictions enable the efficient deployment of response teams and strategic traffic rerouting, leading to reduced congestion and enhanced safety. Furthermore, an in-depth understanding of incident types helps in implementing preventive measures and formulating strategies to alleviate their influence on road networks. In this paper, we present a comprehensive framework for accurately predicting incident duration, with a particular emphasis on the critical role of street conditions and locations as major incident triggers. To demonstrate the effectiveness of our framework, we performed an in-depth case study using a dataset from San Francisco. We introduce a novel feature called "Risk" derived from the Risk Priority Number (RPN) concept, highlighting the significance of the incident location in both incident occurrence and prediction. Additionally, we propose a refined incident categorization through fuzzy clustering methods, delineating a unique policy for identifying boundary clusters that necessitate further modeling and testing under varying scenarios. Each cluster undergoes a Multiple Criteria Decision-Making (MCDM) process to gain deeper insights into their distinctions and provide valuable managerial insights. Finally, we employ both traditional Machine Learning (ML) and Deep Learning (DL) models to perform classification and regression tasks. Specifically, incidents residing in boundary clusters are predicted utilizing the scenarios outlined in this study. Through a rigorous analysis of feature importance using top-performing predictive models, we identify the "Risk" factor as a critical determinant of incident duration. Moreover, variables such as distance, humidity, and hour demonstrate significant influence, further enhancing the predictive power of the proposed model.

PMID:39746103 | DOI:10.1371/journal.pone.0316289

Categories: Literature Watch

An investigation of feature reduction, transferability, and generalization in AWID datasets for secure Wi-Fi networks

Thu, 2025-01-02 06:00

PLoS One. 2025 Jan 2;20(1):e0306747. doi: 10.1371/journal.pone.0306747. eCollection 2025.

ABSTRACT

The widespread use of wireless networks to transfer an enormous amount of sensitive information has caused a plethora of vulnerabilities and privacy issues. The management frames, particularly authentication and association frames, are vulnerable to cyberattacks and it is a significant concern. Existing research in Wi-Fi attack detection focused on obtaining high detection accuracy while neglecting modern traffic and attack scenarios such as key reinstallation or unauthorized decryption attacks. This study proposed a novel approach using the AWID 3 dataset for cyberattack detection. The retained features were analyzed to assess their transferability, creating a lightweight and cost-effective model. A decision tree with a recursive feature elimination method was implemented for the extraction of the reduced features subset, and an additional feature wlan_radio.signal_dbm was used in combination with the extracted feature subset. Several deep learning and machine learning models were implemented, where DT and CNN achieved promising classification results. Further, feature transferability and generalizability were evaluated, and their detection performance was analyzed across different network versions where CNN outperformed other classification models. The practical implications of this research are crucial for the secure automation of wireless intrusion detection frameworks and tools in personal and enterprise paradigms.

PMID:39746088 | DOI:10.1371/journal.pone.0306747

Categories: Literature Watch

Artificial intelligence in dentistry: Assessing the informational quality of YouTube videos

Thu, 2025-01-02 06:00

PLoS One. 2025 Jan 2;20(1):e0316635. doi: 10.1371/journal.pone.0316635. eCollection 2025.

ABSTRACT

BACKGROUND AND PURPOSE: The most widely used social media platform for video content is YouTubeTM. The present study evaluated the quality of information on YouTubeTM on artificial intelligence (AI) in dentistry.

METHODS: This cross-sectional study used YouTubeTM (https://www.youtube.com) for searching videos. The terms used for the search were "artificial intelligence in dentistry," "machine learning in dental care," and "deep learning in dentistry." The accuracy and reliability of the information source were assessed using the DISCERN score. The quality of the videos was evaluated using the modified Global Quality Score (mGQS) and the Journal of the American Medical Association (JAMA) score.

RESULTS: The analysis of 91 YouTube™ videos on AI in dentistry revealed insights into video characteristics, content, and quality. On average, videos were 22.45 minutes and received 1715.58 views and 23.79 likes. The topics were mainly centered on general dentistry (66%), with radiology (18%), orthodontics (9%), prosthodontics (4%), and implants (3%). DISCERN and mGQS scores were higher for videos uploaded by healthcare professionals and educational content videos(P<0.05). DISCERN exhibited a strong correlation (0.75) with the video source and with JAMA (0.77). The correlation of the video's content and mGQS, was 0.66 indicated moderate correlation.

CONCLUSION: YouTube™ has informative and moderately reliable videos on AI in dentistry. Dental students, dentists and patients can use these videos to learn and educate about artificial intelligence in dentistry. Professionals should upload more videos to enhance the reliability of the content.

PMID:39746083 | DOI:10.1371/journal.pone.0316635

Categories: Literature Watch

A phase transition in diffusion models reveals the hierarchical nature of data

Thu, 2025-01-02 06:00

Proc Natl Acad Sci U S A. 2025 Jan 7;122(1):e2408799121. doi: 10.1073/pnas.2408799121. Epub 2025 Jan 2.

ABSTRACT

Understanding the structure of real data is paramount in advancing modern deep-learning methodologies. Natural data such as images are believed to be composed of features organized in a hierarchical and combinatorial manner, which neural networks capture during learning. Recent advancements show that diffusion models can generate high-quality images, hinting at their ability to capture this underlying compositional structure. We study this phenomenon in a hierarchical generative model of data. We find that the backward diffusion process acting after a time t is governed by a phase transition at some threshold time, where the probability of reconstructing high-level features, like the class of an image, suddenly drops. Instead, the reconstruction of low-level features, such as specific details of an image, evolves smoothly across the whole diffusion process. This result implies that at times beyond the transition, the class has changed, but the generated sample may still be composed of low-level elements of the initial image. We validate these theoretical insights through numerical experiments on class-unconditional ImageNet diffusion models. Our analysis characterizes the relationship between time and scale in diffusion models and puts forward generative models as powerful tools to model combinatorial data properties.

PMID:39746044 | DOI:10.1073/pnas.2408799121

Categories: Literature Watch

Exploring happiness factors with explainable ensemble learning in a global pandemic

Thu, 2025-01-02 06:00

PLoS One. 2025 Jan 2;20(1):e0313276. doi: 10.1371/journal.pone.0313276. eCollection 2025.

ABSTRACT

Happiness is a state of contentment, joy, and fulfillment, arising from relationships, accomplishments, and inner peace, leading to well-being and positivity. The greatest happiness principle posits that morality is determined by pleasure, aiming for a society where individuals are content and free from suffering. While happiness factors vary, some are universally recognized. The World Happiness Report (WHR), published annually, includes data on 'GDP per capita', 'social support', 'life expectancy', 'freedom to make life choices', 'generosity', and 'perceptions of corruption'. This paper predicts happiness scores using Machine Learning (ML), Deep Learning (DL), and ensemble ML and DL algorithms and examines the impact of individual variables on the happiness index. We also show the impact of COVID-19 pandemic on the happiness features. We design two ensemble ML and DL models using blending and stacking ensemble techniques, namely, Blending RGMLL, which combines Ridge Regression (RR), Gradient Boosting (GB), Multilayer Perceptron (MLP), Long Short-Term Memory (LSTM), and Linear Regression (LR), and Stacking LRGR, which combines LR, Random Forest (RF), GB, and RR. Among the trained models, Blending RGMLL demonstrates the highest predictive accuracy with R2 of 85%, MSE of 0.15, and RMSE of 0.38. We employ Explainable Artificial Intelligence (XAI) techniques to uncover changes in happiness indices, variable importance, and the impact of the COVID-19 pandemic on happiness. The study utilizes an open dataset from the WHR, covering 156 countries from 2018 to 2023. Our findings indicate that 'GDP per capita' is the most critical indicator of happiness score (HS), while 'social support' and 'healthy life expectancy' are also important features before and after the pandemic. However, during the pandemic, 'social support' emerged as the most important indicator, followed by 'healthy life expectancy' and 'GDP per capita', because social support is the prime necessity in the pandemic situation. The outcome of this research helps people understand the impact of these features on increasing the HS and provides guidelines on how happiness can be maintain during unwanted situations. Future research will explore advanced methods and include other related features with real-time monitoring for more comprehensive insights.

PMID:39746025 | DOI:10.1371/journal.pone.0313276

Categories: Literature Watch

Predicting learning achievement using ensemble learning with result explanation

Thu, 2025-01-02 06:00

PLoS One. 2025 Jan 2;20(1):e0312124. doi: 10.1371/journal.pone.0312124. eCollection 2025.

ABSTRACT

Predicting learning achievement is a crucial strategy to address high dropout rates. However, existing prediction models often exhibit biases, limiting their accuracy. Moreover, the lack of interpretability in current machine learning methods restricts their practical application in education. To overcome these challenges, this research combines the strengths of various machine learning algorithms to design a robust model that performs well across multiple metrics, and uses interpretability analysis to elucidate the prediction results. This study introduces a predictive framework for learning achievement based on ensemble learning techniques. Specifically, six distinct machine learning models are utilized to establish a base learner, with logistic regression serving as the meta learner to construct an ensemble model for predicting learning achievement. The SHapley Additive exPlanation (SHAP) model is then employed to explain the prediction results. Through the experiments on XuetangX dataset, the effectiveness of the proposed model is verified. The proposed model outperforms traditional machine learning and deep learning model in terms of prediction accuracy. The results demonstrate that the ensemble learning-based predictive framework significantly outperforms traditional machine learning methods. Through feature importance analysis, the SHAP method enhances model interpretability and improves the reliability of the prediction results, enabling more personalized interventions to support students.

PMID:39745993 | DOI:10.1371/journal.pone.0312124

Categories: Literature Watch

Deep Learning-Based SD-OCT Layer Segmentation Quantifies Outer Retina Changes in Patients With Biallelic RPE65 Mutations Undergoing Gene Therapy

Thu, 2025-01-02 06:00

Invest Ophthalmol Vis Sci. 2025 Jan 2;66(1):5. doi: 10.1167/iovs.66.1.5.

ABSTRACT

PURPOSE: To quantify outer retina structural changes and define novel biomarkers of inherited retinal degeneration associated with biallelic mutations in RPE65 (RPE65-IRD) in patients before and after subretinal gene augmentation therapy with voretigene neparvovec (Luxturna).

METHODS: Application of advanced deep learning for automated retinal layer segmentation, specifically tailored for RPE65-IRD. Quantification of five novel biomarkers for the ellipsoid zone (EZ): thickness, granularity, reflectivity, and intensity. Estimation of the EZarea in single and volume scans was performed with optimized segmentation boundaries. The control group was age similar and without significant refractive error. Spherical equivalent refraction and ocular length were evaluated in all patients.

RESULTS: We observed significant differences in the structural analysis of EZ biomarkers in 22 patients with RPE65-IRD compared with 94 healthy controls. Relative EZ intensities were already reduced in pediatric eyes. Reductions of EZ local granularity and EZ thickness were only significant in adult eyes. Distances of the outer plexiform layer, external limiting membrane, and Bruch's membrane to EZ were reduced at all ages. EZ diameter and area were better preserved in pediatric eyes undergoing therapy with voretigene neparvovec and in patients with a milder phenotype.

CONCLUSIONS: Automated quantitative analysis of biomarkers within EZ visualizes distinct structural differences in the outer retina of patients including treatment-related effects. The automated approach using deep learning strategies allows big data analysis for distinct forms of inherited retinal degeneration. Limitations include a small dataset and potential effects on OCT scans from myopia at least -5 diopters, the latter considered nonsignificant for outer retinal layers.

PMID:39745677 | DOI:10.1167/iovs.66.1.5

Categories: Literature Watch

Pages