Deep learning
Generative adversarial networks for anonymous acneic face dataset generation
PLoS One. 2024 Apr 16;19(4):e0297958. doi: 10.1371/journal.pone.0297958. eCollection 2024.
ABSTRACT
It is well known that the performance of any classification model is effective if the dataset used for the training process and the test process satisfy some specific requirements. In other words, the more the dataset size is large, balanced, and representative, the more one can trust the proposed model's effectiveness and, consequently, the obtained results. Unfortunately, large-size anonymous datasets are generally not publicly available in biomedical applications, especially those dealing with pathological human face images. This concern makes using deep-learning-based approaches challenging to deploy and difficult to reproduce or verify some published results. In this paper, we propose an efficient method to generate a realistic anonymous synthetic dataset of human faces, focusing on attributes related to acne disorders at three distinct levels of severity (Mild, Moderate, and Severe). Notably, our approach initiates from a small dataset of facial acne images, leveraging generative techniques to augment and diversify the dataset, ensuring comprehensive coverage of acne severity levels while maintaining anonymity and realism in the synthetic data. Therefore, a specific hierarchy StyleGAN-based algorithm trained at distinct levels is considered. Moreover, the utilization of generative adversarial networks for augmentation offers a means to circumvent potential privacy or legal concerns associated with acquiring medical datasets. This is attributed to the synthetic nature of the generated data, where no actual subjects are present, thereby ensuring compliance with privacy regulations and legal considerations. To evaluate the performance of the proposed scheme, we consider a CNN-based classification system, trained using the generated synthetic acneic face images and tested using authentic face images. Consequently, we show that an accuracy of 97.6% is achieved using InceptionResNetv2. As a result, this work allows the scientific community to employ the generated synthetic dataset for any data processing application without restrictions on legal or ethical concerns. Moreover, this approach can also be extended to other applications requiring the generation of synthetic medical images.
PMID:38625866 | DOI:10.1371/journal.pone.0297958
Barriers and Potential Solutions to Glaucoma Screening in the Developing World: A Review
J Glaucoma. 2024 Apr 17. doi: 10.1097/IJG.0000000000002404. Online ahead of print.
ABSTRACT
PURPOSE: Glaucoma is a leading public health concern globally and its detection and management are way more complex and challenging in the developing world. This review article discusses barriers to glaucoma screening in developing countries from the perspective of different key stakeholders and proposes solutions.
METHODS/RESULTS: A literature search was carried out in the electronic catalogs of PubMed, Medline, and Cochrane database of systematic reviews to find studies that focused on barriers and enablers to glaucoma screening. The authors' interpretations were tabulated as descriptive and qualitative data and presented concisely from the point of view of key stakeholders such as the patients and their relatives, care providers, and system/governing bodies. Key barriers to glaucoma care identified are lack of awareness, poor accessibility to ophthalmic centers, inadequately trained human resources, unsatisfactory infrastructure, and non-availability of financially viable screening programs. Educating care providers as well as the public, providing care closer to where people live, and developing cost-effective screening strategies are needed to ensure proper identification of glaucoma patients in developing countries.
CONCLUSIONS: The logistics of glaucoma detection and management are complex. Hence, glaucoma detection programs should be implemented only when facilities for glaucoma management are in place. Understanding the importance of glaucoma screening and its future implications, addressing the various roadblocks, empowering and efficiently implementing the existing strategies, and incorporating novel ones using Artificial Intelligence (AI) and Deep Learning (DL) will help in establishing a robust glaucoma screening program in developing countries.
PMID:38625838 | DOI:10.1097/IJG.0000000000002404
Automatic detection of scalp high-frequency oscillations based on deep learning
IEEE Trans Neural Syst Rehabil Eng. 2024 Apr 16;PP. doi: 10.1109/TNSRE.2024.3389010. Online ahead of print.
ABSTRACT
OBJECTIVE: Scalp high-frequency oscillations (sHFOs) are a promising non-invasive biomarker of epilepsy. However, the visual marking of sHFOs is a time-consuming and subjective process, existing automatic detectors based on single-dimensional analysis have difficulty with accurately eliminating artifacts and thus do not provide sufficient reliability to meet clinical needs. Therefore, we propose a high-performance sHFOs detector based on a deep learning algorithm.
METHODS: An initial detection module was designed to extract candidate high-frequency oscillations. Then, one-dimensional (1D) and two-dimensional (2D) deep learning models were designed, respectively. Finally, the weighted voting method is used to combine the outputs of the two model.
RESULTS: In experiments, the precision, recall, specificity and F1-score were 83.44%, 83.60%, 96.61% and 83.42%, respectively, on average and the kappa coefficient was 80.02%. In addition, the proposed detector showed a stable performance on multi-centre datasets.
CONCLUSIONS: Our sHFOs detector demonstrated high robustness and generalisation ability, which indicates its potential applicability as a clinical assistance tool.
SIGNIFICANCE: The proposed sHFOs detector achieves an accurate and robust method via deep learning algorithm.
PMID:38625771 | DOI:10.1109/TNSRE.2024.3389010
Deep Location Soft-Embedding-Based Network with Regional Scoring for Mammogram Classification
IEEE Trans Med Imaging. 2024 Apr 16;PP. doi: 10.1109/TMI.2024.3389661. Online ahead of print.
ABSTRACT
Early detection and treatment of breast cancer can significantly reduce patient mortality, and mammogram is an effective method for early screening. Computer-aided diagnosis (CAD) of mammography based on deep learning can assist radiologists in making more objective and accurate judgments. However, existing methods often depend on datasets with manual segmentation annotations. In addition, due to the large image sizes and small lesion proportions, many methods that do not use region of interest (ROI) mostly rely on multi-scale and multi-feature fusion models. These shortcomings increase the labor, money, and computational overhead of applying the model. Therefore, a deep location soft-embedding-based network with regional scoring (DLSEN-RS) is proposed. DLSEN-RS is an end-to-end mammography image classification method containing only one feature extractor and relies on positional embedding (PE) and aggregation pooling (AP) modules to locate lesion areas without bounding boxes, transfer learning, or multi-stage training. In particular, the introduced PE and AP modules exhibit versatility across various CNN models and improve the model's tumor localization and diagnostic accuracy for mammography images. Experiments are conducted on published INbreast and CBIS-DDSM datasets, and compared to previous state-of-the-art mammographic image classification methods, DLSEN-RS performed satisfactorily.
PMID:38625766 | DOI:10.1109/TMI.2024.3389661
Adaptiveness for Online Learning: Conceptualising 'Online Learning Dexterity' from Higher Education Students' Experiences
N Z J Educ Stud. 2023 May 13:1-19. doi: 10.1007/s40841-023-00287-2. Online ahead of print.
ABSTRACT
Online learning dexterity, or the ability to effortlessly adapt to online learning situations, has become critical since the COVID-19 pandemic, but its processes are not well-understood. Using grounded theory, this study develops a paradigm model of online learning dexterity from semi-structured interviews with 32 undergraduate and postgraduate students from a university in New Zealand. Through students' online learning experiences during the pandemic from 2020 to 2021, online learning dexterity is found to be how students make online learning 'just as good' as face-to-face learning by creating and adjusting five learning manoeuvres according to developing online learning circumstances. Undergraduates and postgraduates re-use familiar study strategies as deep learning manoeuvres, but undergraduates restrict support-seeking manoeuvres to lecturers. Technical problems with online systems and poor course organisation by lecturers affected learning productivity, resulting in the need for more time optimisation manoeuvres. Social support helped students activate persistence manoeuvres to sustain online class attendance. However, undergraduates had more problems sustaining interest and engagement during class as they were not as proficient with using learning presence manoeuvres as postgraduates enrolled in distance learning programmes. The theoretical and practical significance of online learning dexterity for post-pandemic higher education is discussed.
PMID:38625124 | PMC:PMC10182339 | DOI:10.1007/s40841-023-00287-2
Clinical Utility of a CT-based AI Prognostic Model for Segmentectomy in Non-Small Cell Lung Cancer
Radiology. 2024 Apr;311(1):e231793. doi: 10.1148/radiol.231793.
ABSTRACT
Background Currently, no tool exists for risk stratification in patients undergoing segmentectomy for non-small cell lung cancer (NSCLC). Purpose To develop and validate a deep learning (DL) prognostic model using preoperative CT scans and clinical and radiologic information for risk stratification in patients with clinical stage IA NSCLC undergoing segmentectomy. Materials and Methods In this single-center retrospective study, transfer learning of a pretrained model was performed for survival prediction in patients with clinical stage IA NSCLC who underwent lobectomy from January 2008 to March 2017. The internal set was divided into training, validation, and testing sets based on the assignments from the pretraining set. The model was tested on an independent test set of patients with clinical stage IA NSCLC who underwent segmentectomy from January 2010 to December 2017. Its prognostic performance was analyzed using the time-dependent area under the receiver operating characteristic curve (AUC), sensitivity, and specificity for freedom from recurrence (FFR) at 2 and 4 years and lung cancer-specific survival and overall survival at 4 and 6 years. The model sensitivity and specificity were compared with those of the Japan Clinical Oncology Group (JCOG) eligibility criteria for sublobar resection. Results The pretraining set included 1756 patients. Transfer learning was performed in an internal set of 730 patients (median age, 63 years [IQR, 56-70 years]; 366 male), and the segmentectomy test set included 222 patients (median age, 65 years [IQR, 58-71 years]; 114 male). The model performance for 2-year FFR was as follows: AUC, 0.86 (95% CI: 0.76, 0.96); sensitivity, 87.4% (7.17 of 8.21 patients; 95% CI: 59.4, 100); and specificity, 66.7% (136 of 204 patients; 95% CI: 60.2, 72.8). The model showed higher sensitivity for FFR than the JCOG criteria (87.4% vs 37.6% [3.08 of 8.21 patients], P = .02), with similar specificity. Conclusion The CT-based DL model identified patients at high risk among those with clinical stage IA NSCLC who underwent segmentectomy, outperforming the JCOG criteria. © RSNA, 2024 Supplemental material is available for this article.
PMID:38625008 | DOI:10.1148/radiol.231793
Multitask Adversarial Networks Based on Extensive Nonlinear Spiking Neuron Models
Int J Neural Syst. 2024 Apr 17:2450032. doi: 10.1142/S0129065724500321. Online ahead of print.
ABSTRACT
Deep learning technology has been successfully used in Chest X-ray (CXR) images of COVID-19 patients. However, due to the characteristics of COVID-19 pneumonia and X-ray imaging, the deep learning methods still face many challenges, such as lower imaging quality, fewer training samples, complex radiological features and irregular shapes. To address these challenges, this study first introduces an extensive NSNP-like neuron model, and then proposes a multitask adversarial network architecture based on ENSNP-like neurons for chest X-ray images of COVID-19, called MAE-Net. The MAE-Net serves two tasks: (i) converting low-quality CXR images to high-quality images; (ii) classifying CXR images of COVID-19. The adversarial architecture of MAE-Net uses two generators and two discriminators, and two new loss functions have been introduced to guide the optimization of the network. The MAE-Net is tested on four benchmark COVID-19 CXR image datasets and compared them with eight deep learning models. The experimental results show that the proposed MAE-Net can enhance the conversion quality and the accuracy of image classification results.
PMID:38624267 | DOI:10.1142/S0129065724500321
Knowledge-driven deep learning for fast MR imaging: Undersampled MR image reconstruction from supervised to un-supervised learning
Magn Reson Med. 2024 Apr 16. doi: 10.1002/mrm.30105. Online ahead of print.
ABSTRACT
Deep learning (DL) has emerged as a leading approach in accelerating MRI. It employs deep neural networks to extract knowledge from available datasets and then applies the trained networks to reconstruct accurate images from limited measurements. Unlike natural image restoration problems, MRI involves physics-based imaging processes, unique data properties, and diverse imaging tasks. This domain knowledge needs to be integrated with data-driven approaches. Our review will introduce the significant challenges faced by such knowledge-driven DL approaches in the context of fast MRI along with several notable solutions, which include learning neural networks and addressing different imaging application scenarios. The traits and trends of these techniques have also been given which have shifted from supervised learning to semi-supervised learning, and finally, to unsupervised learning methods. In addition, MR vendors' choices of DL reconstruction have been provided along with some discussions on open questions and future directions, which are critical for the reliable imaging systems.
PMID:38624162 | DOI:10.1002/mrm.30105
Adverse Event Signal Detection Using Patients' Concerns in Pharmaceutical Care Records: Evaluation of Deep Learning Models
J Med Internet Res. 2024 Apr 16;26:e55794. doi: 10.2196/55794.
ABSTRACT
BACKGROUND: Early detection of adverse events and their management are crucial to improving anticancer treatment outcomes, and listening to patients' subjective opinions (patients' voices) can make a major contribution to improving safety management. Recent progress in deep learning technologies has enabled various new approaches for the evaluation of safety-related events based on patient-generated text data, but few studies have focused on the improvement of real-time safety monitoring for individual patients. In addition, no study has yet been performed to validate deep learning models for screening patients' narratives for clinically important adverse event signals that require medical intervention. In our previous work, novel deep learning models have been developed to detect adverse event signals for hand-foot syndrome or adverse events limiting patients' daily lives from the authored narratives of patients with cancer, aiming ultimately to use them as safety monitoring support tools for individual patients.
OBJECTIVE: This study was designed to evaluate whether our deep learning models can screen clinically important adverse event signals that require intervention by health care professionals. The applicability of our deep learning models to data on patients' concerns at pharmacies was also assessed.
METHODS: Pharmaceutical care records at community pharmacies were used for the evaluation of our deep learning models. The records followed the SOAP format, consisting of subjective (S), objective (O), assessment (A), and plan (P) columns. Because of the unique combination of patients' concerns in the S column and the professional records of the pharmacists, this was considered a suitable data for the present purpose. Our deep learning models were applied to the S records of patients with cancer, and the extracted adverse event signals were assessed in relation to medical actions and prescribed drugs.
RESULTS: From 30,784 S records of 2479 patients with at least 1 prescription of anticancer drugs, our deep learning models extracted true adverse event signals with more than 80% accuracy for both hand-foot syndrome (n=152, 91%) and adverse events limiting patients' daily lives (n=157, 80.1%). The deep learning models were also able to screen adverse event signals that require medical intervention by health care providers. The extracted adverse event signals could reflect the side effects of anticancer drugs used by the patients based on analysis of prescribed anticancer drugs. "Pain or numbness" (n=57, 36.3%), "fever" (n=46, 29.3%), and "nausea" (n=40, 25.5%) were common symptoms out of the true adverse event signals identified by the model for adverse events limiting patients' daily lives.
CONCLUSIONS: Our deep learning models were able to screen clinically important adverse event signals that require intervention for symptoms. It was also confirmed that these deep learning models could be applied to patients' subjective information recorded in pharmaceutical care records accumulated during pharmacists' daily work.
PMID:38625718 | DOI:10.2196/55794
Stability of radiomic features from positron emission tomography images: a phantom study comparing advanced reconstruction algorithms and ordered subset expectation maximization
Phys Eng Sci Med. 2024 Apr 16. doi: 10.1007/s13246-024-01416-x. Online ahead of print.
ABSTRACT
In this study, we compared the repeatability and reproducibility of radiomic features obtained from positron emission tomography (PET) images according to the reconstruction algorithm used-advanced reconstruction algorithms, such as HYPER iterative (IT), HYPER deep learning reconstruction (DLR), and HYPER deep progressive reconstruction (DPR), or traditional Ordered Subset Expectation Maximization (OSEM)-to understand the potential variations and implications of using advanced reconstruction techniques in PET-based radiomics. We used a heterogeneous phantom with acrylic spherical beads (4- or 8-mm diameter) filled with 18F. PET images were acquired and reconstructed using OSEM, IT, DLR, and DPR. Original and wavelet radiomic features were calculated using SlicerRadiomics. Radiomic feature repeatability was assessed using the Coefficient of Variance (COV) and intraclass correlation coefficient (ICC), and inter-acquisition time reproducibility was assessed using the concordance correlation coefficient (CCC). For the 4- and 8-mm diameter beads phantom, the proportion of radiomic features with a COV < 10% was equivocal or higher for the advanced reconstruction algorithm than for OSEM. ICC indicated that advanced methods generally outperformed OSEM in repeatability, except for the original features of the 8-mm beads phantom. In the inter-acquisition time reproducibility analysis, the combinations of 3 and 5 min exhibited the highest reproducibility in both phantoms, with IT and DPR showing the highest proportion of radiomic features with CCC > 0.8. Advanced reconstruction methods provided enhanced stability of radiomic features compared with OSEM, suggesting their potential for optimal image reconstruction in PET-based radiomics, offering potential benefits in clinical diagnostics and prognostics.
PMID:38625624 | DOI:10.1007/s13246-024-01416-x
Automated analysis of pectoralis major thickness in pec-fly exercises: evolving from manual measurement to deep learning techniques
Vis Comput Ind Biomed Art. 2024 Apr 16;7(1):8. doi: 10.1186/s42492-024-00159-6.
ABSTRACT
This study addresses a limitation of prior research on pectoralis major (PMaj) thickness changes during the pectoralis fly exercise using a wearable ultrasound imaging setup. Although previous studies used manual measurement and subjective evaluation, it is important to acknowledge the subsequent limitations of automating widespread applications. We then employed a deep learning model for image segmentation and automated measurement to solve the problem and study the additional quantitative supplementary information that could be provided. Our results revealed increased PMaj thickness changes in the coronal plane within the probe detection region when real-time ultrasound imaging (RUSI) visual biofeedback was incorporated, regardless of load intensity (50% or 80% of one-repetition maximum). Additionally, participants showed uniform thickness changes in the PMaj in response to enhanced RUSI biofeedback. Notably, the differences in PMaj thickness changes between load intensities were reduced by RUSI biofeedback, suggesting altered muscle activation strategies. We identified the optimal measurement location for the maximal PMaj thickness close to the rib end and emphasized the lightweight applicability of our model for fitness training and muscle assessment. Further studies can refine load intensities, investigate diverse parameters, and employ different network models to enhance accuracy. This study contributes to our understanding of the effects of muscle physiology and exercise training.
PMID:38625580 | DOI:10.1186/s42492-024-00159-6
Exploring the potential of machine learning in gynecological care: a review
Arch Gynecol Obstet. 2024 Apr 16. doi: 10.1007/s00404-024-07479-1. Online ahead of print.
ABSTRACT
Gynecological health remains a critical aspect of women's overall well-being, with profound implications for maternal and reproductive outcomes. This comprehensive review synthesizes the current state of knowledge on four pivotal aspects of gynecological health: preterm birth, breast cancer and cervical cancer and infertility treatment. Machine learning (ML) has emerged as a transformative technology with the potential to revolutionize gynecology and women's healthcare. The subsets of AI, namely, machine learning (ML) and deep learning (DL) methods, have aided in detecting complex patterns from huge datasets and using such patterns in making predictions. This paper investigates how machine learning (ML) algorithms are employed in the field of gynecology to tackle crucial issues pertaining to women's health. This paper also investigates the integration of ultrasound technology with artificial intelligence (AI) during the initial, intermediate, and final stages of pregnancy. Additionally, it delves into the diverse applications of AI throughout each trimester.This review paper provides an overview of machine learning (ML) models, introduces natural language processing (NLP) concepts, including ChatGPT, and discusses the clinical applications of artificial intelligence (AI) in gynecology. Additionally, the paper outlines the challenges in utilizing machine learning within the field of gynecology.
PMID:38625543 | DOI:10.1007/s00404-024-07479-1
SadNet: a novel multimodal fusion network for protein-ligand binding affinity prediction
Phys Chem Chem Phys. 2024 Apr 16. doi: 10.1039/d3cp05664c. Online ahead of print.
ABSTRACT
Protein-ligand binding affinity prediction plays an important role in the field of drug discovery. Existing deep learning-based approaches have significantly improved the efficiency of protein-ligand binding affinity prediction through their excellent inductive bias capability. However, these methods only focus on fragmented three-dimensional data, which truncates the integrity of pocket data, leading to the neglect of potential long-range interactions. In this paper, we propose a dual-stream framework, with amino acid sequence assisting the atomic data fusion for graph neural network (termed SadNet), to fuse both 3D atomic data and sequence data for more accurate prediction results. In detail, SadNet consists of a pocket module and a sequence module. The sequence module expands the "receptive field" of the pocket module through a mid-term virtual node fusion. To better integrate sequence-level information from the sequence module and 3D structural information from the pocket module, we incorporate structural information for each amino acid within the sequence module. Besides, to better understand the intrinsic relationship between sequences and 3D atomic information, our SadNet utilizes information stacking from both the early stage and later stage. Experimental results on publicly available benchmark datasets demonstrate the superiority of the proposed dual-stream approach over the state-of-the-art alternatives. The code of this work is available online at https://github.com/wardhong/SadNet.
PMID:38625412 | DOI:10.1039/d3cp05664c
A survey on recent trends in deep learning for nucleus segmentation from histopathology images
Evol Syst (Berl). 2023 Mar 6:1-46. doi: 10.1007/s12530-023-09491-3. Online ahead of print.
ABSTRACT
Nucleus segmentation is an imperative step in the qualitative study of imaging datasets, considered as an intricate task in histopathology image analysis. Segmenting a nucleus is an important part of diagnosing, staging, and grading cancer, but overlapping regions make it hard to separate and tell apart independent nuclei. Deep Learning is swiftly paving its way in the arena of nucleus segmentation, attracting quite a few researchers with its numerous published research articles indicating its efficacy in the field. This paper presents a systematic survey on nucleus segmentation using deep learning in the last five years (2017-2021), highlighting various segmentation models (U-Net, SCPP-Net, Sharp U-Net, and LiverNet) and exploring their similarities, strengths, datasets utilized, and unfolding research areas.
PMID:38625364 | PMC:PMC9987406 | DOI:10.1007/s12530-023-09491-3
An efficient real-time stock prediction exploiting incremental learning and deep learning
Evol Syst (Berl). 2022 Dec 21:1-19. doi: 10.1007/s12530-022-09481-x. Online ahead of print.
ABSTRACT
Intraday trading is popular among traders due to its ability to leverage price fluctuations in a short timeframe. For traders, real-time price predictions for the next few minutes can be beneficial for making strategies. Real-time prediction is challenging due to the stock market's non-stationary, complex, noisy, chaotic, dynamic, volatile, and non-parametric nature. Machine learning models are considered effective for stock forecasting, yet, their hyperparameters need tuning with the latest market data to incorporate the market's complexities. Usually, models are trained and tested in batches, which smooths the correction process and speeds up the learning. When making intraday stock predictions, the models should forecast for each instance in contrast to the whole batch and learn simultaneously to ensure high accuracy. In this paper, we propose a strategy based on two different learning approaches: incremental learning and Offline-Online learning, to forecast the stock price using the real-time stream of the live market. In incremental learning, the model is updated continuously upon receiving the stock's next instance from the live-stream, while in Offline-Online learning, the model is retrained after each trading session to make sure it incorporates the latest data complexities. These methods were applied to univariate time-series (established from historical stock price) and multivariate time-series (considering historical stock price as well as technical indicators). Extensive experiments were performed on the eight most liquid stocks listed on the American NASDAQ and Indian NSE stock exchanges, respectively. The Offline-Online models outperformed incremental models in terms of low forecasting error.
PMID:38625328 | PMC:PMC9769488 | DOI:10.1007/s12530-022-09481-x
PDRF-Net: a progressive dense residual fusion network for COVID-19 lung CT image segmentation
Evol Syst (Berl). 2023 Feb 17:1-17. doi: 10.1007/s12530-023-09489-x. Online ahead of print.
ABSTRACT
The lungs of patients with COVID-19 exhibit distinctive lesion features in chest CT images. Fast and accurate segmentation of lesion sites from CT images of patients' lungs is significant for the diagnosis and monitoring of COVID-19 patients. To this end, we propose a progressive dense residual fusion network named PDRF-Net for COVID-19 lung CT segmentation. Dense skip connections are introduced to capture multi-level contextual information and compensate for the feature loss problem in network delivery. The efficient aggregated residual module is designed for the encoding-decoding structure, which combines a visual transformer and the residual block to enable the network to extract richer and minute-detail features from CT images. Furthermore, we introduce a bilateral channel pixel weighted module to progressively fuse the feature maps obtained from multiple branches. The proposed PDRF-Net obtains good segmentation results on two COVID-19 datasets. Its segmentation performance is superior to baseline by 11.6% and 11.1%, and outperforming other comparative mainstream methods. Thus, PDRF-Net serves as an easy-to-train, high-performance deep learning model that can realize effective segmentation of the COVID-19 lung CT images.
PMID:38625320 | PMC:PMC9936947 | DOI:10.1007/s12530-023-09489-x
AMTLDC: a new adversarial multi-source transfer learning framework to diagnosis of COVID-19
Evol Syst (Berl). 2023 Jan 12:1-15. doi: 10.1007/s12530-023-09484-2. Online ahead of print.
ABSTRACT
In recent years, deep learning techniques have been widely used to diagnose diseases. However, in some tasks, such as the diagnosis of COVID-19 disease, due to insufficient data, the model is not properly trained and as a result, the generalizability of the model decreases. For example, if the model is trained on a CT scan dataset and tested on another CT scan dataset, it predicts near-random results. To address this, data from several different sources can be combined using transfer learning, taking into account the intrinsic and natural differences in existing datasets obtained with different medical imaging tools and approaches. In this paper, to improve the transfer learning technique and better generalizability between multiple data sources, we propose a multi-source adversarial transfer learning model, namely AMTLDC. In AMTLDC, representations are learned that are similar among the sources. In other words, extracted representations are general and not dependent on the particular dataset domain. We apply the AMTLDC to predict Covid-19 from medical images using a convolutional neural network. We show that accuracy can be improved using the AMTLDC framework, and surpass the results of current successful transfer learning approaches. In particular, we show that the AMTLDC works well when using different dataset domains, or when there is insufficient data.
PMID:38625255 | PMC:PMC9838404 | DOI:10.1007/s12530-023-09484-2
Detection of anomalies in cycling behavior with convolutional neural network and deep learning
Eur Transp Res Rev. 2023;15(1):9. doi: 10.1186/s12544-023-00583-4. Epub 2023 Mar 23.
ABSTRACT
BACKGROUND: Cycling has always been considered a sustainable and healthy mode of transport. With the increasing concerns of greenhouse gases and pollution, policy makers are intended to support cycling as commuter mode of transport. Moreover, during Covid-19 period, cycling was further appreciated by citizens as an individual opportunity of mobility. Unfortunately, bicyclist safety has become a challenge with growing number of bicyclists in the 21st century. When compared to the traditional road safety network screening, availability of suitable data for bicycle based crashes is more difficult. In such framework, new technologies based smart cities may require new opportunities of data collection and analysis.
METHODS: This research presents bicycle data requirements and treatment to get suitable information by using GPS device. Mainly, this paper proposed a deep learning-based approach "BeST-DAD" to detect anomalies and spot dangerous points on map for bicyclist to avoid a critical safety event (CSE). BeST-DAD follows Convolutional Neural Network and Autoencoder (AE) for anomaly detection. Proposed model optimization is carried out by testing different data features and BeST-DAD parameter settings, while another comparison performance is carried out between BeST-DAD and Principal Component Analysis (PCA).
RESULT: BeST-DAD over perform than traditional PCA statistical approaches for anomaly detection by achieving 77% of the F-score. When the trained model is tested with data from different users, 100% recall is recorded for individual user's trained models.
CONCLUSION: The research results support the notion that proper GPS trajectory data and deep learning classification can be applied to identify anomalies in cycling behavior.
PMID:38625141 | PMC:PMC10033296 | DOI:10.1186/s12544-023-00583-4
Ultrasound-based deep learning radiomics model for differentiating benign, borderline, and malignant ovarian tumours: a multi-class classification exploratory study
BMC Med Imaging. 2024 Apr 15;24(1):89. doi: 10.1186/s12880-024-01251-2.
ABSTRACT
BACKGROUND: Accurate preoperative identification of ovarian tumour subtypes is imperative for patients as it enables physicians to custom-tailor precise and individualized management strategies. So, we have developed an ultrasound (US)-based multiclass prediction algorithm for differentiating between benign, borderline, and malignant ovarian tumours.
METHODS: We randomised data from 849 patients with ovarian tumours into training and testing sets in a ratio of 8:2. The regions of interest on the US images were segmented and handcrafted radiomics features were extracted and screened. We applied the one-versus-rest method in multiclass classification. We inputted the best features into machine learning (ML) models and constructed a radiomic signature (Rad_Sig). US images of the maximum trimmed ovarian tumour sections were inputted into a pre-trained convolutional neural network (CNN) model. After internal enhancement and complex algorithms, each sample's predicted probability, known as the deep transfer learning signature (DTL_Sig), was generated. Clinical baseline data were analysed. Statistically significant clinical parameters and US semantic features in the training set were used to construct clinical signatures (Clinic_Sig). The prediction results of Rad_Sig, DTL_Sig, and Clinic_Sig for each sample were fused as new feature sets, to build the combined model, namely, the deep learning radiomic signature (DLR_Sig). We used the receiver operating characteristic (ROC) curve and the area under the ROC curve (AUC) to estimate the performance of the multiclass classification model.
RESULTS: The training set included 440 benign, 44 borderline, and 196 malignant ovarian tumours. The testing set included 109 benign, 11 borderline, and 49 malignant ovarian tumours. DLR_Sig three-class prediction model had the best overall and class-specific classification performance, with micro- and macro-average AUC of 0.90 and 0.84, respectively, on the testing set. Categories of identification AUC were 0.84, 0.85, and 0.83 for benign, borderline, and malignant ovarian tumours, respectively. In the confusion matrix, the classifier models of Clinic_Sig and Rad_Sig could not recognise borderline ovarian tumours. However, the proportions of borderline and malignant ovarian tumours identified by DLR_Sig were the highest at 54.55% and 63.27%, respectively.
CONCLUSIONS: The three-class prediction model of US-based DLR_Sig can discriminate between benign, borderline, and malignant ovarian tumours. Therefore, it may guide clinicians in determining the differential management of patients with ovarian tumours.
PMID:38622546 | DOI:10.1186/s12880-024-01251-2
An Automated Multi-scale Feature Fusion Network for Spine Fracture Segmentation Using Computed Tomography Images
J Imaging Inform Med. 2024 Apr 15. doi: 10.1007/s10278-024-01091-0. Online ahead of print.
ABSTRACT
Spine fractures represent a critical health concern with far-reaching implications for patient care and clinical decision-making. Accurate segmentation of spine fractures from medical images is a crucial task due to its location, shape, type, and severity. Addressing these challenges often requires the use of advanced machine learning and deep learning techniques. In this research, a novel multi-scale feature fusion deep learning model is proposed for the automated spine fracture segmentation using Computed Tomography (CT) to these challenges. The proposed model consists of six modules; Feature Fusion Module (FFM), Squeeze and Excitation (SEM), Atrous Spatial Pyramid Pooling (ASPP), Residual Convolution Block Attention Module (RCBAM), Residual Border Refinement Attention Block (RBRAB), and Local Position Residual Attention Block (LPRAB). These modules are used to apply multi-scale feature fusion, spatial feature extraction, channel-wise feature improvement, segmentation border results border refinement, and positional focus on the region of interest. After that, a decoder network is used to predict the fractured spine. The experimental results show that the proposed approach achieves better accuracy results in solving the above challenges and also performs well compared to the existing segmentation methods.
PMID:38622384 | DOI:10.1007/s10278-024-01091-0