Deep learning
Video dataset for the detection of safe and unsafe behaviours in workplaces
Data Brief. 2024 Aug 3;56:110791. doi: 10.1016/j.dib.2024.110791. eCollection 2024 Oct.
ABSTRACT
Real-time detection of safe and unsafe behaviours in production facilities is very important to prevent these behaviours before they occur. In this context, this study presents a high-resolution video-based dataset of safe and unsafe behaviours obtained from a closed production facility for use in occupational accident prevention. The dataset was collected from the security cameras of a production facility operating in an organised industrial zone in Eskişehir, Turkey, in November and December 2022, after obtaining the necessary permissions from company officials and employees. A total of 8 behaviour classes, 4 classes of safe and 4 classes of unsafe behaviours, were identified for the dataset and 691 video clips containing these behaviours were obtained. The video clips created for the dataset are in MP4 format at 1920×1080 pixels and 24 frames per second. In the dataset, the safe behaviour classes are Safe Walkway, Authorized Intervention, Closed Panel Cover and Safe Carrying, while the unsafe behaviour classes are Safe Walkway Violation, Unauthorized Intervention, Opened Panel Cover and Carrying Overload with Forklift.
PMID:39224505 | PMC:PMC11367630 | DOI:10.1016/j.dib.2024.110791
The analysis of teaching quality evaluation for the college sports dance by convolutional neural network model and deep learning
Heliyon. 2024 Aug 9;10(16):e36067. doi: 10.1016/j.heliyon.2024.e36067. eCollection 2024 Aug 30.
ABSTRACT
This study aims to comprehensively analyze and evaluate the quality of college physical dance education using Convolutional Neural Network (CNN) models and deep learning methods. The study introduces a teaching quality evaluation (TQE) model based on one-dimensional CNN, addressing issues such as subjectivity and inconsistent evaluation criteria in traditional assessment methods. By constructing a comprehensive TQE system comprising 24 evaluation indicators, this study innovatively applies deep learning technology to quantitatively assess the quality of physical dance education. This TQE model processes one-dimensional evaluation data by extracting local features through convolutional layers, reducing dimensions via pooling layers, and feeding feature vectors into a classifier through fully connected layers to achieve an overall assessment of teaching quality. Experimental results demonstrate that after 150 iterations of training and validation on the TQE model, convergence is achieved, with mean squared error (MSE) decreasing to 0.0015 and 0.0216 on the training and validation sets, respectively. Comparatively, the TQE model exhibits significantly lower MSE on the training, validation, and test sets compared to the Back-Propagation Neural Network, accompanied by a higher R2 value, indicating superior accuracy and performance in data fitting. Further analysis on robustness, parameter sensitivity, multi-scenario adaptability, and long-term learning capabilities reveals the TQE model's strong resilience and stability in managing noisy data, varying parameter configurations, diverse teaching contexts, and extended time-series data. In practical applications, the TQE model is implemented in physical dance courses at X College to evaluate teaching quality and guide improvement strategies for instructors, resulting in notable enhancements in teaching quality and student satisfaction. In conclusion, this study offers a comprehensive evaluation of university physical dance education quality through a multidimensional assessment system and the application of the 1D-CNN model. It introduces a novel and effective approach to assessing teaching quality, providing a scientific foundation and practical guidance for future educational advancements.
PMID:39224395 | PMC:PMC11367140 | DOI:10.1016/j.heliyon.2024.e36067
Leveraging electrocardiography signals for deep learning-driven cardiovascular disease classification model
Heliyon. 2024 Aug 5;10(16):e35621. doi: 10.1016/j.heliyon.2024.e35621. eCollection 2024 Aug 30.
ABSTRACT
Electrocardiography (ECG) is the most non-invasive diagnostic tool for cardiovascular diseases (CVDs). Automatic analysis of ECG signals assists in accurately and rapidly detecting life-threatening arrhythmias like atrioventricular blockage, atrial fibrillation, ventricular tachycardia, etc. The ECG recognition models need to utilize algorithms to detect various kinds of waveforms in the ECG and identify complicated relationships over time. However, the high variability of wave morphology among patients and noise are challenging issues. Physicians frequently utilize automated ECG abnormality recognition models to classify long-term ECG signals. Recently, deep learning (DL) models can be used to achieve enhanced ECG recognition accuracy in the healthcare decision making system. In this aspect, this study introduces an automated DL enabled ECG signal recognition (ADL-ECGSR) technique for CVD detection and classification. The ADL-ECGSR technique employs three most important subprocesses: pre-processed, feature extraction, parameter tuning, and classification. Besides, the ADL-ECGSR technique involves the design of a bidirectional long short-term memory (BiLSTM) based feature extractor, and the Adamax optimizer is utilized to optimize the trained method of the BiLSTM model. Finally, the dragonfly algorithm (DFA) with a stacked sparse autoencoder (SSAE) module is applied to recognize and classify EEG signals. An extensive range of simulations occur on benchmark PTB-XL datasets to validate the enhanced ECG recognition efficiency. The comparative analysis of the ADL-ECGSR methodology showed a remarkable performance of 91.24 % on the existing methods.
PMID:39224246 | PMC:PMC11367027 | DOI:10.1016/j.heliyon.2024.e35621
Transformer-based deep learning networks for fault detection, classification, and location prediction in transmission lines
Network. 2024 Sep 3:1-21. doi: 10.1080/0954898X.2024.2393746. Online ahead of print.
ABSTRACT
Fault detection, classification, and location prediction are crucial for maintaining the stability and reliability of modern power systems, reducing economic losses, and enhancing system protection sensitivity. This paper presents a novel Hierarchical Deep Learning Approach (HDLA) for accurate and efficient fault diagnosis in transmission lines. HDLA leverages two-stage transformer-based classification and regression models to perform Fault Detection (FD), Fault Type Classification (FTC), and Fault Location Prediction (FLP) directly from synchronized raw three-phase current and voltage samples. By bypassing the need for feature extraction, HDLA significantly reduces computational complexity while achieving superior performance compared to existing deep learning methods. The efficacy of HDLA is validated on a comprehensive dataset encompassing various fault scenarios with diverse types, locations, resistances, inception angles, and noise levels. The results demonstrate significant improvements in accuracy, recall, precision, and F1-score metrics for classification, and Mean Absolute Errors (MAEs) and Root Mean Square Errors (RMSEs) for prediction, showcasing the effectiveness of HDLA for real-time fault diagnosis in power systems.
PMID:39224075 | DOI:10.1080/0954898X.2024.2393746
Robust mosquito species identification from diverse body and wing images using deep learning
Parasit Vectors. 2024 Sep 2;17(1):372. doi: 10.1186/s13071-024-06459-3.
ABSTRACT
Mosquito-borne diseases are a major global health threat. Traditional morphological or molecular methods for identifying mosquito species often require specialized expertise or expensive laboratory equipment. The use of convolutional neural networks (CNNs) to identify mosquito species based on images may offer a promising alternative, but their practical implementation often remains limited. This study explores the applicability of CNNs in classifying mosquito species. It compares the efficacy of body and wing depictions across three image collection methods: a smartphone, macro-lens attached to a smartphone and a professional stereomicroscope. The study included 796 specimens of four morphologically similar Aedes species, Aedes aegypti, Ae. albopictus, Ae. koreicus and Ae. japonicus japonicus. The findings of this study indicate that CNN models demonstrate superior performance in wing-based classification 87.6% (95% CI: 84.2-91.0) compared to body-based classification 78.9% (95% CI: 77.7-80.0). Nevertheless, there are notable limitations of CNNs as they perform reliably across multiple devices only when trained specifically on those devices, resulting in an average decline of mean accuracy by 14%, even with extensive image augmentation. Additionally, we also estimate the required training data volume for effective classification, noting a reduced requirement for wing-based classification compared to body-based methods. Our study underscores the viability of both body and wing classification methods for mosquito species identification while emphasizing the need to address practical constraints in developing accessible classification systems.
PMID:39223629 | DOI:10.1186/s13071-024-06459-3
Evaluation of the invasiveness of pure ground-glass nodules based on dual-head ResNet technique
BMC Cancer. 2024 Sep 2;24(1):1080. doi: 10.1186/s12885-024-12823-4.
ABSTRACT
OBJECTIVE: To intelligently evaluate the invasiveness of pure ground-glass nodules with multiple classifications using deep learning.
METHODS: pGGNs in 1136 patients were pathologically confirmed as lung precursor lesions [atypical adenomatous hyperplasia (AAH) and adenocarcinoma in situ (AIS)], minimally invasive adenocarcinoma (MIA), or invasive adenocarcinoma (IAC). Four different models [EfficientNet-b0 2D, dual-head ResNet_3D, a 3D model combining three features (3D_3F), and a 3D model combining 19 features (3D_19F)] were constructed to evaluate the invasiveness of pGGNs using the EfficientNet and ResNet networks. The Obuchowski index was used to evaluate the differences in diagnostic efficiency among the four models.
RESULTS: The patients with pGGNs (360 men, 776 women; mean age, 54.63 ± 12.36 years) included 235 cases of AAH + AIS, 332 cases of MIA, and 569 cases of IAC. In the validation group, the areas under the curve in detecting the invasiveness of pGGNs as a three-category classification (AAH + AIS, MIA, IAC) were 0.8008, 0.8090, 0.8165, and 0.8158 for EfficientNet-b0 2D, dual-head ResNet_3D, 3D_3F, and 3D_19F, respectively, whereas the accuracies were 0.6422, 0.6158, 0.651, and 0.6364, respectively. The Obuchowski index revealed no significant differences in the diagnostic performance of the four models.
CONCLUSIONS: The dual-head ResNet_3D_3F model had the highest diagnostic efficiency for evaluating the invasiveness of pGGNs in the four models.
PMID:39223592 | DOI:10.1186/s12885-024-12823-4
Comparison of deep learning models to detect crossbites on 2D intraoral photographs
Head Face Med. 2024 Sep 2;20(1):45. doi: 10.1186/s13005-024-00448-8.
ABSTRACT
BACKGROUND: To support dentists with limited experience, this study trained and compared six convolutional neural networks to detect crossbites and classify non-crossbite, frontal, and lateral crossbites using 2D intraoral photographs.
METHODS: Based on 676 photographs from 311 orthodontic patients, six convolutional neural network models were trained and compared to classify (1) non-crossbite vs. crossbite and (2) non-crossbite vs. lateral crossbite vs. frontal crossbite. The trained models comprised DenseNet, EfficientNet, MobileNet, ResNet18, ResNet50, and Xception.
FINDINGS: Among the models, Xception showed the highest accuracy (98.57%) in the test dataset for classifying non-crossbite vs. crossbite images. When additionally distinguishing between lateral and frontal crossbites, average accuracy decreased with the DenseNet architecture achieving the highest accuracy among the models with 91.43% in the test dataset.
CONCLUSIONS: Convolutional neural networks show high potential in processing clinical photographs and detecting crossbites. This study provides initial insights into how deep learning models can be used for orthodontic diagnosis of malocclusions based on intraoral 2D photographs.
PMID:39223562 | DOI:10.1186/s13005-024-00448-8
Revolutionizing breast ultrasound diagnostics with EfficientNet-B7 and Explainable AI
BMC Med Imaging. 2024 Sep 2;24(1):230. doi: 10.1186/s12880-024-01404-3.
ABSTRACT
Breast cancer is a leading cause of mortality among women globally, necessitating precise classification of breast ultrasound images for early diagnosis and treatment. Traditional methods using CNN architectures such as VGG, ResNet, and DenseNet, though somewhat effective, often struggle with class imbalances and subtle texture variations, leading to reduced accuracy for minority classes such as malignant tumors. To address these issues, we propose a methodology that leverages EfficientNet-B7, a scalable CNN architecture, combined with advanced data augmentation techniques to enhance minority class representation and improve model robustness. Our approach involves fine-tuning EfficientNet-B7 on the BUSI dataset, implementing RandomHorizontalFlip, RandomRotation, and ColorJitter to balance the dataset and improve model robustness. The training process includes early stopping to prevent overfitting and optimize performance metrics. Additionally, we integrate Explainable AI (XAI) techniques, such as Grad-CAM, to enhance the interpretability and transparency of the model's predictions, providing visual and quantitative insights into the features and regions of ultrasound images influencing classification outcomes. Our model achieves a classification accuracy of 99.14%, significantly outperforming existing CNN-based approaches in breast ultrasound image classification. The incorporation of XAI techniques enhances our understanding of the model's decision-making process, thereby increasing its reliability and facilitating clinical adoption. This comprehensive framework offers a robust and interpretable tool for the early detection and diagnosis of breast cancer, advancing the capabilities of automated diagnostic systems and supporting clinical decision-making processes.
PMID:39223507 | DOI:10.1186/s12880-024-01404-3
PCP-GC-LM: single-sequence-based protein contact prediction using dual graph convolutional neural network and convolutional neural network
BMC Bioinformatics. 2024 Sep 2;25(1):287. doi: 10.1186/s12859-024-05914-3.
ABSTRACT
BACKGROUND: Recently, the process of evolution information and the deep learning network has promoted the improvement of protein contact prediction methods. Nevertheless, still remain some bottleneck: (1) One of the bottlenecks is the prediction of orphans and other fewer evolution information proteins. (2) The other bottleneck is the method of predicting single-sequence-based proteins mainly focuses on selecting protein sequence features and tuning the neural network architecture, However, while the deeper neural networks improve prediction accuracy, there is still the problem of increasing the computational burden. Compared with other neural networks in the field of protein prediction, the graph neural network has the following advantages: due to the advantage of revealing the topology structure via graph neural network and being able to take advantage of the hierarchical structure and local connectivity of graph neural networks has certain advantages in capturing the features of different levels of abstraction in protein molecules. When using protein sequence and structure information for joint training, the dependencies between the two kinds of information can be better captured. And it can process protein molecular structures of different lengths and shapes, while traditional neural networks need to convert proteins into fixed-size vectors or matrices for processing.
RESULTS: Here, we propose a single-sequence-based protein contact map predictor PCP-GC-LM, with dual-level graph neural networks and convolution networks. Our method performs better with other single-sequence-based predictors in different independent tests. In addition, to verify the validity of our method against complex protein structures, we will also compare it with other methods in two homodimers protein test sets (DeepHomo test dataset and CASP-CAPRI target dataset). Furthermore, we also perform ablation experiments to demonstrate the necessity of a dual graph network. In all, our framework presents new modules to accurately predict inter-chain contact maps in protein and it's also useful to analyze interactions in other types of protein complexes.
PMID:39223474 | DOI:10.1186/s12859-024-05914-3
Fractional differentiation based image enhancement for automatic detection of malignant melanoma
BMC Med Imaging. 2024 Sep 2;24(1):231. doi: 10.1186/s12880-024-01400-7.
ABSTRACT
Recent improvements in artificial intelligence and computer vision make it possible to automatically detect abnormalities in medical images. Skin lesions are one broad class of them. There are types of lesions that cause skin cancer, again with several types. Melanoma is one of the deadliest types of skin cancer. Its early diagnosis is at utmost importance. The treatments are greatly aided with artificial intelligence by the quick and precise diagnosis of these conditions. The identification and delineation of boundaries inside skin lesions have shown promise when using the basic image processing approaches for edge detection. Further enhancements regarding edge detections are possible. In this paper, the use of fractional differentiation for improved edge detection is explored on the application of skin lesion detection. A framework based on fractional differential filters for edge detection in skin lesion images is proposed that can improve automatic detection rate of malignant melanoma. The derived images are used to enhance the input images. Obtained images then undergo a classification process based on deep learning. A well-studied dataset of HAM10000 is used in the experiments. The system achieves 81.04% accuracy with EfficientNet model using the proposed fractional derivative based enhancements whereas accuracies are around 77.94% when using original images. In almost all the experiments, the enhanced images improved the accuracy. The results show that the proposed method improves the recognition performance.
PMID:39223468 | DOI:10.1186/s12880-024-01400-7
Integrating Contrastive Learning and Cycle Generative Adversarial Networks for Non-invasive Fetal ECG Extraction
Pediatr Cardiol. 2024 Sep 2. doi: 10.1007/s00246-024-03633-3. Online ahead of print.
ABSTRACT
Fetal electrocardiogram (FECG) contains crucial information about the fetus during pregnancy, making the extraction of FECG signal essential for monitoring fetal health. However, extracting FECG signal from abdominal electrocardiogram (AECG) poses several challenges: (1) FECG signal is often contaminated by noise, and (2) FECG signal is frequently overshadowed by high-amplitude maternal electrocardiogram (MECG). To address these issues and enhance the accuracy of signal extraction, this paper proposes an improved Cycle Generative Adversarial Networks (CycleGAN) with integrated contrastive learning for FECG signal extraction. The model introduces a dual-attention mechanism in the generator of the generative adversarial network, incorporating a multi-head self-attention (MSA) module and a channel-wise self-attention (CSA) module to enhance the quality of generated signals. Additionally, a contrastive triplet loss is integrated into the CycleGAN loss function, optimizing training to increase the similarity between the extracted FECG signal and the scalp fetal electrocardiogram. The proposed method is evaluated using the ADFECG dataset and the PCDB dataset both from the Physionet. In terms of signal extraction quality, Mean Squared Error is reduced to 0.036, Mean Absolute Error (MAE) to 0.009, and Pearson Correlation Coefficient reaches 0.924. When validating the model performance, Structural Similarity Index achieves 95.54%, Peak Signal-to-Noise Ratio (PSNR) reaches 38.87 dB, and R-squared (R2) attains 95.12%. Furthermore, the positive predictive value (PPV), sensitivity (SEN) and F1-score for QRS wave cluster detection on the ADFECG dataset also reached 99.56%, 99.43% and 99.50%, respectively. On the PCDB dataset, the positive predictive value (PPV), sensitivity (SEN) and F1-score for QRS wave cluster detection also reached 98.24%, 98.60% and 98.42%, respectively. All of them are higher than other methods. Therefore, the proposed model has important applications in effective monitoring of fetal health during pregnancy.
PMID:39223338 | DOI:10.1007/s00246-024-03633-3
Enabling target-aware molecule generation to follow multi objectives with Pareto MCTS
Commun Biol. 2024 Sep 2;7(1):1074. doi: 10.1038/s42003-024-06746-w.
ABSTRACT
Target-aware drug discovery has greatly accelerated the drug discovery process to design small-molecule ligands with high binding affinity to disease-related protein targets. Conditioned on targeted proteins, previous works utilize various kinds of deep generative models and have shown great potential in generating molecules with strong protein-ligand binding interactions. However, beyond binding affinity, effective drug molecules must manifest other essential properties such as high drug-likeness, which are not explicitly addressed by current target-aware generative methods. In this article, aiming to bridge the gap of multi-objective target-aware molecule generation in the field of deep learning-based drug discovery, we propose ParetoDrug, a Pareto Monte Carlo Tree Search (MCTS) generation algorithm. ParetoDrug searches molecules on the Pareto Front in chemical space using MCTS to enable synchronous optimization of multiple properties. Specifically, ParetoDrug utilizes pretrained atom-by-atom autoregressive generative models for the exploration guidance to desired molecules during MCTS searching. Besides, when selecting the next atom symbol, a scheme named ParetoPUCT is proposed to balance exploration and exploitation. Benchmark experiments and case studies demonstrate that ParetoDrug is highly effective in traversing the large and complex chemical space to discover novel compounds with satisfactory binding affinities and drug-like properties for various multi-objective target-aware drug discovery tasks.
PMID:39223327 | DOI:10.1038/s42003-024-06746-w
Impact of acquisition area on deep-learning-based glaucoma detection in different plexuses in OCTA
Sci Rep. 2024 Sep 2;14(1):20414. doi: 10.1038/s41598-024-71235-3.
ABSTRACT
Glaucoma is a group of neurodegenerative diseases that can lead to irreversible blindness. Yet, the progression can be slowed down if diagnosed and treated early enough. Optical coherence tomography angiography (OCTA) can non-invasively provide valuable information about the retinal microcirculation that has shown to be correlated with the onset of the disease. The vessel density (VD) is the most commonly used biomarker to quantify this vascular information. However, different studies showed that there is a great impact of the acquisition area on the performance of the VD to distinguish between glaucoma patients and a healthy control group. It also seems that the separate capillary plexuses are differently affected by the disease and therefore also influence the results. So in this study we investigate the impact of the acquisition area (3 × 3 mm macular scan, 6.44 × 6.4 mm macular scan, 6 × 6 mm optic nerve head (ONH) scan) and the different plexuses on the machine-learning-based distinction between glaucoma patients and healthy controls. The results yielded that the 6 × 6 mm ONH show the best performance over all plexuses. Moreover the deep learning-based approach outperforms the VD as a biomarker on every acquisition area and plexus. In addition to that, it also performs better than traditional biomarkers obtained from the OCT scans that are used in the clinical routine for diagnosis and progression tracking of glaucoma. Consequently, OCTA scans of the ONH might be a useful addition to OCT when studying glaucoma.
PMID:39223266 | DOI:10.1038/s41598-024-71235-3
Correction: Free access via computational cloud to deep learning-based EEG assessment in neonatal hypoxic-ischemic encephalopathy: revolutionary opportunities to overcome health disparities
Pediatr Res. 2024 Sep 2. doi: 10.1038/s41390-024-03539-z. Online ahead of print.
NO ABSTRACT
PMID:39223254 | DOI:10.1038/s41390-024-03539-z
SCINet: A Segmentation and Classification Interaction CNN Method for Arteriosclerotic Retinopathy Grading
Interdiscip Sci. 2024 Sep 2. doi: 10.1007/s12539-024-00650-x. Online ahead of print.
ABSTRACT
As a common disease, cardiovascular and cerebrovascular diseases pose a great harm threat to human wellness. Even using advanced and comprehensive treatment methods, there is still a high mortality rate. Arteriosclerosis, as an important factor reflecting the severity of cardiovascular and cerebrovascular diseases, is imperative to detect the arteriosclerotic retinopathy. However, the detection of arteriosclerosis retinopathy requires expensive and time-consuming manual evaluation, while end-to-end deep learning detection methods also need interpretable design to high light task-related features. Considering the importance of automatic arteriosclerotic retinopathy grading, we propose a segmentation and classification interaction network (SCINet). We propose a segmentation and classification interaction architecture for grading arteriosclerotic retinopathy. After IterNet is used to segment retinal vessel from original fundus images, the backbone feature extractor roughly extracts features from the segmented and original fundus arteriosclerosis images and further enhances them through the vessel aware module. The last classifier module generates fundus arteriosclerosis grading results. Specifically, the vessel aware module is designed to highlight the important areal vessel features segmented from original images by attention mechanism, thereby achieving information interaction. The attention mechanism selectively learns the vessel features of segmentation region information under the proposed interactive architecture, which leads to reweighting the extracted features and enhances significant feature information. Extensive experiments have confirmed the effect of our model. SCINet has the best performance on the task of arteriosclerotic retinopathy grading. Additionally, the CNN method is scalable to similar tasks by incorporating segmented images as auxiliary information.
PMID:39222258 | DOI:10.1007/s12539-024-00650-x
Reconstruction residual network with a fused spatial-channel attention mechanism for automatically classifying diabetic foot ulcer
Phys Eng Sci Med. 2024 Sep 2. doi: 10.1007/s13246-024-01472-3. Online ahead of print.
ABSTRACT
Diabetic foot ulcer (DFU) is a common chronic complication of diabetes. This complication is characterized by the formation of ulcers that are difficult to heal on the skin of the foot. Ulcers can negatively affect patients' quality of life, and improperly treated lesions can result in amputation and even death. Traditionally, the severity and type of foot ulcers are determined by doctors through visual observations and on the basis of their clinical experience; however, this subjective evaluation can lead to misjudgments. In addition, quantitative methods have been developed for classifying and scoring are therefore time-consuming and labor-intensive. In this paper, we propose a reconstruction residual network with a fused spatial-channel attention mechanism (FARRNet) for automatically classifying DFUs. The use of pseudo-labeling and Data augmentation as a pre-processing technique can overcome problems caused by data imbalance and small sample size. The developed model's attention was enhanced using a spatial channel attention (SPCA) module that incorporates spatial and channel attention mechanisms. A reconstruction mechanism was incorporated into the developed residual network to improve its feature extraction ability for achieving better classification. The performance of the proposed model was compared with that of state-of-the-art models and those in the DFUC Grand Challenge. When applied to the DFUC Grand Challenge, the proposed method outperforms other state-of-the-art schemes in terms of accuracy, as evaluated using 5-fold cross-validation and the following metrics: macro-average F1-score, AUC, Recall, and Precision. FARRNet achieved the F1-score of 60.81%, AUC of 87.37%, Recall of 61.04%, and Precision of 61.56%. Therefore, the proposed model is more suitable for use in medical diagnosis environments with embedded devices and limited computing resources. The proposed model can assist patients in initial identifications of ulcer wounds, thereby helping them to obtain timely treatment.
PMID:39222215 | DOI:10.1007/s13246-024-01472-3
Predicting agricultural and meteorological droughts using Holt Winter Conventional 2D-Long Short-Term Memory (HW-Conv2DLSTM)
Environ Monit Assess. 2024 Sep 2;196(10):875. doi: 10.1007/s10661-024-13063-6.
ABSTRACT
Drought is an extended shortage of rainfall resulting in water scarcity and affecting a region's social and economic conditions through environmental deterioration. Its adverse environmental effects can be minimised by timely prediction. Drought detection uses only ground observation stations, but satellite-based supervision scans huge land mass stretches and offers highly effective monitoring. This paper puts forward a novel drought monitoring system using satellite imagery by considering the effects of droughts that devastated agriculture in Thanjavur district, Tamil Nadu, between 2000 and 2022. The proposed method uses Holt Winter Conventional 2D-Long Short-Term Memory (HW-Conv2DLSTM) to forecast meteorological and agricultural droughts. It employs Climate Hazards Group InfraRed Precipitation with Station (CHIRPS) data precipitation index datasets, MODIS 11A1 temperature index, and MODIS 13Q1 vegetation index. It extracts the time series data from satellite images using trend and seasonal patterns and smoothens them using Holt Winter alpha, beta, and gamma parameters. Finally, an effective drought prediction procedure is developed using Conv2D-LSTM to calculate the spatiotemporal correlation amongst drought indices. The HW-Conv2DLSTM offers a better R2 value of 0.97. It holds promise as an effective computer-assisted strategy to predict droughts and maintain agricultural productivity, which is vital to feed the ever-increasing human population.
PMID:39222153 | DOI:10.1007/s10661-024-13063-6
Integration of Animal Behavioral Assessment and Convolutional Neural Network to Study Wasabi-Alcohol Taste-Smell Interaction
J Vis Exp. 2024 Aug 16;(210). doi: 10.3791/66981.
ABSTRACT
The commercial wasabi pastes commonly used for food preparation contain a homologous compound of chemosensory isothiocyanates (ITCs) that elicit an irritating sensation upon consumption. The impact of sniffing dietary alcoholic beverages on the sensation of wasabi spiciness has never been studied. While most sensory evaluation studies focus on individual food and beverages separately, there is a lack of research on the olfactory study of sniffing liquor while consuming wasabi. Here, a methodology is developed that combines the use of an animal behavioral study and a convolutional neural network to analyze the facial expressions of mice when they simultaneously sniff liquor and consume wasabi. The results demonstrate that the trained and validated deep learning model recognizes 29% of the images depicting co-treatment of wasabi and alcohol belonging to the class of the wasabi-negative liquor-positive group without the need for prior training materials filtering. Statistical analysis of mouse grimace scale scores obtained from the selected video frame images reveals a significant difference (P < 0.01) between the presence and absence of liquor. This finding suggests that dietary alcoholic beverages might have a diminishing effect on the wasabi-elicited reactions in mice. This combinatory methodology holds potential for individual ITC compound screening and sensory analyses of spirit components in the future. However, further study is required to investigate the underlying mechanism of alcohol-induced suppression of wasabi pungency.
PMID:39221929 | DOI:10.3791/66981
A novel Skin lesion prediction and classification technique: ViT-GradCAM
Skin Res Technol. 2024 Sep;30(9):e70040. doi: 10.1111/srt.70040.
ABSTRACT
BACKGROUND: Skin cancer is one of the highly occurring diseases in human life. Early detection and treatment are the prime and necessary points to reduce the malignancy of infections. Deep learning techniques are supplementary tools to assist clinical experts in detecting and localizing skin lesions. Vision transformers (ViT) based on image segmentation classification using multiple classes provide fairly accurate detection and are gaining more popularity due to legitimate multiclass prediction capabilities.
MATERIALS AND METHODS: In this research, we propose a new ViT Gradient-Weighted Class Activation Mapping (GradCAM) based architecture named ViT-GradCAM for detecting and classifying skin lesions by spreading ratio on the lesion's surface area. The proposed system is trained and validated using a HAM 10000 dataset by studying seven skin lesions. The database comprises 10 015 dermatoscopic images of varied sizes. The data preprocessing and data augmentation techniques are applied to overcome the class imbalance issues and improve the model's performance.
RESULT: The proposed algorithm is based on ViT models that classify the dermatoscopic images into seven classes with an accuracy of 97.28%, precision of 98.51, recall of 95.2%, and an F1 score of 94.6, respectively. The proposed ViT-GradCAM obtains better and more accurate detection and classification than other state-of-the-art deep learning-based skin lesion detection models. The architecture of ViT-GradCAM is extensively visualized to highlight the actual pixels in essential regions associated with skin-specific pathologies.
CONCLUSION: This research proposes an alternate solution to overcome the challenges of detecting and classifying skin lesions using ViTs and GradCAM, which play a significant role in detecting and classifying skin lesions accurately rather than relying solely on deep learning models.
PMID:39221858 | DOI:10.1111/srt.70040
A Decade of Computational Mass Spectrometry from Reference Spectra to Deep Learning
Chimia (Aarau). 2024 Aug 21;78(7-8):525-530. doi: 10.2533/chimia.2024.525.
ABSTRACT
Computational methods are playing an increasingly important role as a complement to conventional data evaluation methods in analytical chemistry, and particularly mass spectrometry. Computational mass spectrometry (CompMS) is the application of computational methods on mass spectrometry data. Herein, advances in CompMS for small molecule chemistry are discussed in the areas of spectral libraries, spectrum prediction, and tentative structure identification (annotation): Automatic spectrum curation is facilitating the expansion of openly available spectral libraries, a crucial resource both for compound annotation directly and as a resource for machine learning algorithms. Spectrum prediction and molecular fingerprint prediction have emerged as two key approaches to compound annotation. For both, multiple methods based on classical machine learning and deep learning have been developed. Driven by advances in deep learning-based generative chemistry, de novo structure generation from fragment spectra is emerging as a new field of research. This review highlights key publications in these fields, including our approaches RMassBank (automatic spectrum curation) and MSNovelist (de novo structure generation).
PMID:39221848 | DOI:10.2533/chimia.2024.525