Semantic Web

Semantic Web, Reusable Learning Objects, Personal Learning Networks in Health: Key Pieces for Digital Health Literacy.

Fri, 2017-07-07 07:58
Related Articles

Semantic Web, Reusable Learning Objects, Personal Learning Networks in Health: Key Pieces for Digital Health Literacy.

Stud Health Technol Inform. 2017;238:219-222

Authors: Konstantinidis ST, Wharrad H, Windle R, Bamidis PD

Abstract
The knowledge existing in the World Wide Web is exponentially expanding, while continuous advancements in health sciences contribute to the creation of new knowledge. There are a lot of efforts trying to identify how the social connectivity can endorse patients' empowerment, while other studies look at the identification and the quality of online materials. However, emphasis has not been put on the big picture of connecting the existing resources with the patients "new habits" of learning through their own Personal Learning Networks. In this paper we propose a framework for empowering patients' digital health literacy adjusted to patients' currents needs by utilizing the contemporary way of learning through Personal Learning Networks, existing high quality learning resources and semantics technologies for interconnecting knowledge pieces. The framework based on the concept of knowledge maps for health as defined in this paper. Health Digital Literacy needs definitely further enhancement and the use of the proposed concept might lead to useful tools which enable use of understandable health trusted resources tailored to each person needs.

PMID: 28679928 [PubMed - in process]

Categories: Literature Watch

Issues Associated With the Use of Semantic Web Technology in Knowledge Acquisition for Clinical Decision Support Systems: Systematic Review of the Literature.

Fri, 2017-07-07 07:58
Related Articles

Issues Associated With the Use of Semantic Web Technology in Knowledge Acquisition for Clinical Decision Support Systems: Systematic Review of the Literature.

JMIR Med Inform. 2017 Jul 05;5(3):e18

Authors: Zolhavarieh S, Parry D, Bai Q

Abstract
BACKGROUND: Knowledge-based clinical decision support system (KB-CDSS) can be used to help practitioners make diagnostic decisions. KB-CDSS may use clinical knowledge obtained from a wide variety of sources to make decisions. However, knowledge acquisition is one of the well-known bottlenecks in KB-CDSSs, partly because of the enormous growth in health-related knowledge available and the difficulty in assessing the quality of this knowledge as well as identifying the "best" knowledge to use. This bottleneck not only means that lower-quality knowledge is being used, but also that KB-CDSSs are difficult to develop for areas where expert knowledge may be limited or unavailable. Recent methods have been developed by utilizing Semantic Web (SW) technologies in order to automatically discover relevant knowledge from knowledge sources.
OBJECTIVE: The two main objectives of this study were to (1) identify and categorize knowledge acquisition issues that have been addressed through using SW technologies and (2) highlight the role of SW for acquiring knowledge used in the KB-CDSS.
METHODS: We conducted a systematic review of the recent work related to knowledge acquisition MeM for clinical decision support systems published in scientific journals. In this regard, we used the keyword search technique to extract relevant papers.
RESULTS: The retrieved papers were categorized based on two main issues: (1) format and data heterogeneity and (2) lack of semantic analysis. Most existing approaches will be discussed under these categories. A total of 27 papers were reviewed in this study.
CONCLUSIONS: The potential for using SW technology in KB-CDSS has only been considered to a minor extent so far despite its promise. This review identifies some questions and issues regarding use of SW technology for extracting relevant knowledge for a KB-CDSS.

PMID: 28679487 [PubMed - in process]

Categories: Literature Watch

CodeMapper: semiautomatic coding of case definitions. A contribution from the ADVANCE project.

Thu, 2017-06-29 06:53

CodeMapper: semiautomatic coding of case definitions. A contribution from the ADVANCE project.

Pharmacoepidemiol Drug Saf. 2017 Jun 28;:

Authors: Becker BFH, Avillach P, Romio S, van Mulligen EM, Weibel D, Sturkenboom MCJM, Kors JA, ADVANCE consortium

Abstract
BACKGROUND: Assessment of drug and vaccine effects by combining information from different healthcare databases in the European Union requires extensive efforts in the harmonization of codes as different vocabularies are being used across countries. In this paper, we present a web application called CodeMapper, which assists in the mapping of case definitions to codes from different vocabularies, while keeping a transparent record of the complete mapping process.
METHODS: CodeMapper builds upon coding vocabularies contained in the Metathesaurus of the Unified Medical Language System. The mapping approach consists of three phases. First, medical concepts are automatically identified in a free-text case definition. Second, the user revises the set of medical concepts by adding or removing concepts, or expanding them to related concepts that are more general or more specific. Finally, the selected concepts are projected to codes from the targeted coding vocabularies. We evaluated the application by comparing codes that were automatically generated from case definitions by applying CodeMapper's concept identification and successive concept expansion, with reference codes that were manually created in a previous epidemiological study.
RESULTS: Automated concept identification alone had a sensitivity of 0.246 and positive predictive value (PPV) of 0.420 for reproducing the reference codes. Three successive steps of concept expansion increased sensitivity to 0.953 and PPV to 0.616.
CONCLUSIONS: Automatic concept identification in the case definition alone was insufficient to reproduce the reference codes, but CodeMapper's operations for concept expansion provide an effective, efficient, and transparent way for reproducing the reference codes.

PMID: 28657162 [PubMed - as supplied by publisher]

Categories: Literature Watch

Knowledge Based Topic Model for Unsupervised Object Discovery and Localization.

Tue, 2017-06-27 08:57

Knowledge Based Topic Model for Unsupervised Object Discovery and Localization.

IEEE Trans Image Process. 2017 Jun 22;:

Authors: Niu Z, Hua G, Wang L, Gao X

Abstract
Unsupervised object discovery and localization is to discover some dominant object classes and localize all of object instances from a given image collection without any supervision. Previous work has attempted to tackle this problem with vanilla topic models such as Latent Dirichlet Allocation (LDA). However, in those methods no prior knowledge for the given image collection is exploited to facilitate object discovery. On the other hand, the topic models used in those methods suffer from the topic coherence issue-some inferred topics do not have clear meaning, which limits the final performance of object discovery. In this paper, prior knowledge in terms of the so-called Must-Links is exploited from Web images on the Internet. Furthermore, a novel knowledge-based topic model, called Latent Dirichlet Allocation with Mixture of Dirichlet Trees (LDA-MDT), is proposed to incorporate the Must-Links into topic modeling for object discovery. In particular, to better deal with the polysemy phenomenon of visual words, the Must-Link is re-defined as that one Must-Link only constrains one or some topic(s) instead of all topics, which leads to significantly improved topic coherence. Moreover, the Must-Links are built and grouped with respect to specific object classes, thus the Must-Links in our approach are semantic-specific, which allows to more efficiently exploit discriminative prior knowledge from Web images. Extensive experiments validated the efficiency of our proposed approach on several datasets. It is shown that our method significantly improves topic coherence and outperforms the unsupervised methods for object discovery and localization. In addition, compared to discriminative methods, the naturally existing object classes in the given image collection can be subtly discovered, which makes our approach well suited for realistic applications of unsupervised object discovery.

PMID: 28650813 [PubMed - as supplied by publisher]

Categories: Literature Watch

Robust Web Image Annotation via Exploring Multi-facet and Structural Knowledge.

Sat, 2017-06-24 07:32
Related Articles

Robust Web Image Annotation via Exploring Multi-facet and Structural Knowledge.

IEEE Trans Image Process. 2017 Jun 19;:

Authors: Hu M, Yang Y, Shen F, Zhang L, Shen HT, Li X

Abstract
Driven by the rapid development of Internet and digital technologies, we have witnessed the explosive growth of Web images in recent years. Seeing that labels can reflect the semantic contents of the images, automatic image annotation, which can further facilitate the procedure of image semantic indexing, retrieval and other image management tasks, has become one of the most crucial research directions in multimedia. Most of the existing annotation methods heavily rely on well-labeled training data (expensive to collect) and/or single view of visual features (insufficient representative power). In this paper, inspired by the promising advance of feature engineering (e.g., CNN feature and SIFT feature) and inexhaustible image data (associated with noisy and incomplete labels) on the Web, we propose an effective and robust scheme, termed Robust Multi-view Semi-supervised Learning (RMSL), for facilitating image annotation task. Specifically, we exploit both labeled images and unlabeled images to uncover the intrinsic data structural information. Meanwhile, to comprehensively describe an individual datum, we take advantage of the correlated and complemental information derived from multiple facets of image data (i.e. multiple views or features). We devise a robust pair-wise constraint on outcomes of different views to achieve annotation consistency. Furthermore, we integrate a robust classifier learning component via ℓ2,p loss, which can provide effective noise identification power during the learning process. Finally, we devise an efficient iterative algorithm to solve the optimization problem in RMSL. We conduct comprehensive experiments on three different datasets, and the results illustrate that our proposed approach is promising for automatic image annotation.

PMID: 28641261 [PubMed - as supplied by publisher]

Categories: Literature Watch

NCBO Ontology Recommender 2.0: an enhanced approach for biomedical ontology recommendation.

Fri, 2017-06-09 08:53
Related Articles

NCBO Ontology Recommender 2.0: an enhanced approach for biomedical ontology recommendation.

J Biomed Semantics. 2017 Jun 07;8(1):21

Authors: Martínez-Romero M, Jonquet C, O'Connor MJ, Graybeal J, Pazos A, Musen MA

Abstract
BACKGROUND: Ontologies and controlled terminologies have become increasingly important in biomedical research. Researchers use ontologies to annotate their data with ontology terms, enabling better data integration and interoperability across disparate datasets. However, the number, variety and complexity of current biomedical ontologies make it cumbersome for researchers to determine which ones to reuse for their specific needs. To overcome this problem, in 2010 the National Center for Biomedical Ontology (NCBO) released the Ontology Recommender, which is a service that receives a biomedical text corpus or a list of keywords and suggests ontologies appropriate for referencing the indicated terms.
METHODS: We developed a new version of the NCBO Ontology Recommender. Called Ontology Recommender 2.0, it uses a novel recommendation approach that evaluates the relevance of an ontology to biomedical text data according to four different criteria: (1) the extent to which the ontology covers the input data; (2) the acceptance of the ontology in the biomedical community; (3) the level of detail of the ontology classes that cover the input data; and (4) the specialization of the ontology to the domain of the input data.
RESULTS: Our evaluation shows that the enhanced recommender provides higher quality suggestions than the original approach, providing better coverage of the input data, more detailed information about their concepts, increased specialization for the domain of the input data, and greater acceptance and use in the community. In addition, it provides users with more explanatory information, along with suggestions of not only individual ontologies but also groups of ontologies to use together. It also can be customized to fit the needs of different ontology recommendation scenarios.
CONCLUSIONS: Ontology Recommender 2.0 suggests relevant ontologies for annotating biomedical text data. It combines the strengths of its predecessor with a range of adjustments and new features that improve its reliability and usefulness. Ontology Recommender 2.0 recommends over 500 biomedical ontologies from the NCBO BioPortal platform, where it is openly available (both via the user interface at http://bioportal.bioontology.org/recommender , and via a Web service API).

PMID: 28592275 [PubMed - in process]

Categories: Literature Watch

Building a semantic web-based metadata repository for facilitating detailed clinical modeling in cancer genome studies.

Wed, 2017-06-07 07:42

Building a semantic web-based metadata repository for facilitating detailed clinical modeling in cancer genome studies.

J Biomed Semantics. 2017 Jun 05;8(1):19

Authors: Sharma DK, Solbrig HR, Tao C, Weng C, Chute CG, Jiang G

Abstract
BACKGROUND: Detailed Clinical Models (DCMs) have been regarded as the basis for retaining computable meaning when data are exchanged between heterogeneous computer systems. To better support clinical cancer data capturing and reporting, there is an emerging need to develop informatics solutions for standards-based clinical models in cancer study domains. The objective of the study is to develop and evaluate a cancer genome study metadata management system that serves as a key infrastructure in supporting clinical information modeling in cancer genome study domains.
METHODS: We leveraged a Semantic Web-based metadata repository enhanced with both ISO11179 metadata standard and Clinical Information Modeling Initiative (CIMI) Reference Model. We used the common data elements (CDEs) defined in The Cancer Genome Atlas (TCGA) data dictionary, and extracted the metadata of the CDEs using the NCI Cancer Data Standards Repository (caDSR) CDE dataset rendered in the Resource Description Framework (RDF). The ITEM/ITEM_GROUP pattern defined in the latest CIMI Reference Model is used to represent reusable model elements (mini-Archetypes).
RESULTS: We produced a metadata repository with 38 clinical cancer genome study domains, comprising a rich collection of mini-Archetype pattern instances. We performed a case study of the domain "clinical pharmaceutical" in the TCGA data dictionary and demonstrated enriched data elements in the metadata repository are very useful in support of building detailed clinical models.
CONCLUSION: Our informatics approach leveraging Semantic Web technologies provides an effective way to build a CIMI-compliant metadata repository that would facilitate the detailed clinical modeling to support use cases beyond TCGA in clinical cancer study domains.

PMID: 28583204 [PubMed - in process]

Categories: Literature Watch

Increased in synthetic cannabinoids-related harms: Results from a longitudinal web-based content analysis.

Mon, 2017-06-05 06:42

Increased in synthetic cannabinoids-related harms: Results from a longitudinal web-based content analysis.

Int J Drug Policy. 2017 Jun 01;44:121-129

Authors: Lamy FR, Daniulaityte R, Nahhas RW, Barratt MJ, Smith AG, Sheth A, Martins SS, Boyer EW, Carlson RG

Abstract
BACKGROUND: Synthetic Cannabinoid Receptor Agonists (SCRA), also known as "K2" or "Spice," have drawn considerable attention due to their potential of abuse and harmful consequences. More research is needed to understand user experiences of SCRA-related effects. We use semi-automated information processing techniques through eDrugTrends platform to examine SCRA-related effects and their variations through a longitudinal content analysis of web-forum data.
METHOD: English language posts from three drug-focused web-forums were extracted and analyzed between January 1st 2008 and September 30th 2015. Search terms are based on the Drug Use Ontology (DAO) created for this study (189 SCRA-related and 501 effect-related terms). EDrugTrends NLP-based text processing tools were used to extract posts mentioning SCRA and their effects. Generalized linear regression was used to fit restricted cubic spline functions of time to test whether the proportion of drug-related posts that mention SCRA (and no other drug) and the proportion of these "SCRA-only" posts that mention SCRA effects have changed over time, with an adjustment for multiple testing.
RESULTS: 19,052 SCRA-related posts (Bluelight (n=2782), Forum A (n=3882), and Forum B (n=12,388)) posted by 2543 international users were extracted. The most frequently mentioned effects were "getting high" (44.0%), "hallucinations" (10.8%), and "anxiety" (10.2%). The frequency of SCRA-only posts declined steadily over the study period. The proportions of SCRA-only posts mentioning positive effects (e.g., "High" and "Euphoria") steadily decreased, while the proportions of SCRA-only posts mentioning negative effects (e.g., "Anxiety," 'Nausea," "Overdose") increased over the same period.
CONCLUSION: This study's findings indicate that the proportion of negative effects mentioned in web forum posts and linked to SCRA has increased over time, suggesting that recent generations of SCRA generate more harms. This is also one of the first studies to conduct automated content analysis of web forum data related to illicit drug use.

PMID: 28578250 [PubMed - as supplied by publisher]

Categories: Literature Watch

"gnparser": a powerful parser for scientific names based on Parsing Expression Grammar.

Sun, 2017-05-28 08:47
Related Articles

"gnparser": a powerful parser for scientific names based on Parsing Expression Grammar.

BMC Bioinformatics. 2017 May 26;18(1):279

Authors: Mozzherin DY, Myltsev AA, Patterson DJ

Abstract
BACKGROUND: Scientific names in biology act as universal links. They allow us to cross-reference information about organisms globally. However variations in spelling of scientific names greatly diminish their ability to interconnect data. Such variations may include abbreviations, annotations, misspellings, etc. Authorship is a part of a scientific name and may also differ significantly. To match all possible variations of a name we need to divide them into their elements and classify each element according to its role. We refer to this as 'parsing' the name. Parsing categorizes name's elements into those that are stable and those that are prone to change. Names are matched first by combining them according to their stable elements. Matches are then refined by examining their varying elements. This two stage process dramatically improves the number and quality of matches. It is especially useful for the automatic data exchange within the context of "Big Data" in biology.
RESULTS: We introduce Global Names Parser (gnparser). It is a Java tool written in Scala language (a language for Java Virtual Machine) to parse scientific names. It is based on a Parsing Expression Grammar. The parser can be applied to scientific names of any complexity. It assigns a semantic meaning (such as genus name, species epithet, rank, year of publication, authorship, annotations, etc.) to all elements of a name. It is able to work with nested structures as in the names of hybrids. gnparser performs with ≈99% accuracy and processes 30 million name-strings/hour per CPU thread. The gnparser library is compatible with Scala, Java, R, Jython, and JRuby. The parser can be used as a command line application, as a socket server, a web-app or as a RESTful HTTP-service. It is released under an Open source MIT license.
CONCLUSIONS: Global Names Parser (gnparser) is a fast, high precision tool for biodiversity informaticians and biologists working with large numbers of scientific names. It can replace expensive and error-prone manual parsing and standardization of scientific names in many situations, and can quickly enhance the interoperability of distributed biological information.

PMID: 28549446 [PubMed - in process]

Categories: Literature Watch

Semantic Technologies and Bio-Ontologies.

Fri, 2017-05-26 07:47
Related Articles

Semantic Technologies and Bio-Ontologies.

Methods Mol Biol. 2017;1617:83-91

Authors: Gutierrez F

Abstract
As information available through data repositories constantly grows, the need for automated mechanisms for linking, querying, and sharing data has become a relevant factor both in research and industry. This situation is more evident in research fields such as the life sciences, where new experiments by different research groups are constantly generating new information regarding a wide variety of related study objects. However, current methods for representing information and knowledge are not suited for machine processing. The Semantic Technologies are a set of standards and protocols that intend to provide methods for representing and handling data that encourages reusability of information and is machine-readable. In this chapter, we will provide a brief introduction to Semantic Technologies, and how these protocols and standards have been incorporated into the life sciences to facilitate dissemination and access to information.

PMID: 28540678 [PubMed - in process]

Categories: Literature Watch

ODMSummary: A Tool for Automatic Structured Comparison of Multiple Medical Forms Based on Semantic Annotation with the Unified Medical Language System.

Tue, 2017-05-23 06:17
Related Articles

ODMSummary: A Tool for Automatic Structured Comparison of Multiple Medical Forms Based on Semantic Annotation with the Unified Medical Language System.

PLoS One. 2016;11(10):e0164569

Authors: Storck M, Krumm R, Dugas M

Abstract
INTRODUCTION: Medical documentation is applied in various settings including patient care and clinical research. Since procedures of medical documentation are heterogeneous and developed further, secondary use of medical data is complicated. Development of medical forms, merging of data from different sources and meta-analyses of different data sets are currently a predominantly manual process and therefore difficult and cumbersome. Available applications to automate these processes are limited. In particular, tools to compare multiple documentation forms are missing. The objective of this work is to design, implement and evaluate the new system ODMSummary for comparison of multiple forms with a high number of semantically annotated data elements and a high level of usability.
METHODS: System requirements are the capability to summarize and compare a set of forms, enable to estimate the documentation effort, track changes in different versions of forms and find comparable items in different forms. Forms are provided in Operational Data Model format with semantic annotations from the Unified Medical Language System. 12 medical experts were invited to participate in a 3-phase evaluation of the tool regarding usability.
RESULTS: ODMSummary (available at https://odmtoolbox.uni-muenster.de/summary/summary.html) provides a structured overview of multiple forms and their documentation fields. This comparison enables medical experts to assess multiple forms or whole datasets for secondary use. System usability was optimized based on expert feedback.
DISCUSSION: The evaluation demonstrates that feedback from domain experts is needed to identify usability issues. In conclusion, this work shows that automatic comparison of multiple forms is feasible and the results are usable for medical experts.

PMID: 27736972 [PubMed - indexed for MEDLINE]

Categories: Literature Watch

A Web Based Tool to Enhance Monitoring and Retention in Care for Tuberculosis Affected Patients.

Wed, 2017-05-10 08:17
Related Articles

A Web Based Tool to Enhance Monitoring and Retention in Care for Tuberculosis Affected Patients.

Stud Health Technol Inform. 2017;237:204-208

Authors: Giannini B, Riccardi N, Di Biagio A, Cenderello G, Giacomini M

Abstract
Tuberculosis (TB) is responsible for a global epidemic. TB treatment requires long-term therapy usually with multiple drugs, which have potential side effects and interactions that may influence patients' adherence to treatment. The TB Ge network is a multi-centric web based platform that collects clinical information of TB affected patients to increase their support and retention in care. The system stores the list of all tuberculosis episodes for each patient with the related data, starting from the first visit including follow-ups clinical evaluations, laboratory tests, imaging and therapies. Data can be manually input through the web interface or can be automatically imported from hospitals Laboratory Information Systems without human intervention. Automatic data import enhances data reuse and prevents errors introduction and time wasting. The network is an implementation of the Healthcare Services Specification Project (HSSP), as the Retrieve, Locate, and Update Service (RLUS) is used to manage patients' data. Clinical data are shared through standard HL7 Clinical Document Architecture (CDA) documents. Semantic interoperability is granted by the adoption of LOINC and ATC codes.

PMID: 28479569 [PubMed - in process]

Categories: Literature Watch

Neuro-symbolic representation learning on biological knowledge graphs.

Fri, 2017-04-28 08:37
Related Articles

Neuro-symbolic representation learning on biological knowledge graphs.

Bioinformatics. 2017 Apr 25;:

Authors: Alshahrani M, Khan MA, Maddouri O, Kinjo AR, Queralt-Rosinach N, Hoehndorf R

Abstract
Motivation: Biological data and knowledge bases increasingly rely on Semantic Web technologies and the use of knowledge graphs for data integration, retrieval and federated queries. In the past years, feature learning methods that are applicable to graph-structured data are becoming available, but have not yet widely been applied and evaluated on structured biological knowledge.
Results: We develop a novel method for feature learning on biological knowledge graphs. Our method combines symbolic methods, in particular knowledge representation using symbolic logic and automated reasoning, with neural networks to generate embeddings of nodes that encode for related information within knowledge graphs. Through the use of symbolic logic, these embeddings contain both explicit and implicit information. We apply these embeddings to the prediction of edges in the knowledge graph representing problems of function prediction, finding candidate genes of diseases, protein-protein interactions, or drug target relations, and demonstrate performance that matches and sometimes outperforms traditional approaches based on manually crafted features. Our method can be applied to any biological knowledge graph, and will thereby open up the increasing amount of SemanticWeb based knowledge bases in biology to use in machine learning and data analytics.
Availability and Implementation: https://github.com/bio-ontology-research-group/walking-rdf-and-owl.
Contact: robert.hoehndorf@kaust.edu.sa.
Supplementary information: Supplementary data are available at Bioinformatics online.

PMID: 28449114 [PubMed - as supplied by publisher]

Categories: Literature Watch

A Digital Framework to Support Providers and Patients in Diabetes Related Behavior Modification.

Fri, 2017-04-21 08:22
Related Articles

A Digital Framework to Support Providers and Patients in Diabetes Related Behavior Modification.

Stud Health Technol Inform. 2017;235:589-593

Authors: Abidi S, Vallis M, Piccinini-Vallis H, Imran SA, Abidi SSR

Abstract
We present Diabetes Web-Centric Information and Support Environment (D-WISE) that features: (a) Decision support tool to assist family physicians to administer Behavior Modification (BM) strategies to patients; and (b) Patient BM application that offers BM strategies and motivational interventions to engage patients. We take a knowledge management approach, using semantic web technologies, to model the social cognition theory constructs, Canadian diabetes guidelines and BM protocols used locally, in terms of a BM ontology that drives the BM decision support to physicians and BM strategy adherence monitoring and messaging to patients. We present the qualitative analysis of D-WISE usability by both physicians and patients.

PMID: 28423861 [PubMed - in process]

Categories: Literature Watch

An Approach for the Support of Semantic Workflows in Electronic Health Records.

Fri, 2017-04-21 08:22
Related Articles

An Approach for the Support of Semantic Workflows in Electronic Health Records.

Stud Health Technol Inform. 2017;235:501-505

Authors: Schweitzer M, Hoerbst A

Abstract
With the unprecedented increase of healthcare data, technologies need to be both, highly efficient for the meaningful utilization of accessible data and flexible to adapt to future challenges and individual preferences. The OntoHealth system makes use of semantic technologies to enable flexible and individual interaction with Electronic Health Records (EHR) for physicians. This is achieved by the execution of formally modelled clinical workflows and the composition of Semantic Web Services (SWS). Several seamless components provide a service-oriented structure defined by individual designed EHR-workflows. This work gives an overview of the planned architecture and its main components. The architecture constitutes the basis for the prototype implementation of all components. With its highly dynamic structure based on SWS, the architecture will be able to cope with both, the individual users' needs as well as the quick evolving healthcare domain.

PMID: 28423843 [PubMed - in process]

Categories: Literature Watch

Discovering Central Practitioners in a Medical Discussion Forum Using Semantic Web Analytics.

Fri, 2017-04-21 08:22
Related Articles

Discovering Central Practitioners in a Medical Discussion Forum Using Semantic Web Analytics.

Stud Health Technol Inform. 2017;235:486-490

Authors: Rajabi E, Abidi SSR

Abstract
The aim of this paper is to investigate semantic web based methods to enrich and transform a medical discussion forum in order to perform semantics-driven social network analysis. We use the centrality measures as well as semantic similarity metrics to identify the most influential practitioners within a discussion forum. The centrality results of our approach are in line with centrality measures produced by traditional SNA methods, thus validating the applicability of semantic web based methods for SNA, particularly for analyzing social networks for specialized discussion forums.

PMID: 28423840 [PubMed - in process]

Categories: Literature Watch

A Semantic Framework for Logical Cross-Validation, Evaluation and Impact Analyses of Population Health Interventions.

Fri, 2017-04-21 08:22
Related Articles

A Semantic Framework for Logical Cross-Validation, Evaluation and Impact Analyses of Population Health Interventions.

Stud Health Technol Inform. 2017;235:481-485

Authors: Shaban-Nejad A, Okhmatovskaia A, Shin EK, Davis RL, Franklin BE, Buckeridge DL

Abstract
Most chronic diseases are a result of a complex web of causative and correlated factors. As a result, effective public health or clinical interventions that intend to generate a sustainable change in these diseases most often use a combination of strategies or programs. To optimize comparative effectiveness evaluations and select the most efficient intervention(s), stakeholders (i.e. public health institutions, policy-makers and advocacy groups, practitioners, insurers, clinicians, and researchers) need access to reliable assessment methods. Building on the theory of Evidence-Based Public Health (EBPH) we introduce a knowledge-based framework for evaluating the consistency and effectiveness of public health programs, interventions, and policies. We use a semantic inference model that assists decision-makers in finding inconsistencies, identifying selection and information biases, and with identifying confounding and hidden dependencies in different public health programs and interventions. The use of formal ontologies for automatic evaluation and assessment of public health programs improves program transparency to stakeholders and decision makers, which in turn increases buy-in and acceptance of methods, connects multiple evaluation activities, and strengthens cost analysis.

PMID: 28423839 [PubMed - in process]

Categories: Literature Watch

Combining Archetypes, Ontologies and Formalization Enables Automated Computation of Quality Indicators.

Fri, 2017-04-21 08:22
Related Articles

Combining Archetypes, Ontologies and Formalization Enables Automated Computation of Quality Indicators.

Stud Health Technol Inform. 2017;235:416-420

Authors: Legaz-García MDC, Dentler K, Fernández-Breis JT, Cornet R

Abstract
ArchMS is a framework that represents clinical information and knowledge using ontologies in OWL, which facilitates semantic interoperability and thereby the exploitation and secondary use of clinical data. However, it does not yet support the automated assessment of quality of care. CLIF is a stepwise method to formalize quality indicators. The method has been implemented in the CLIF tool which supports its users in generating computable queries based on a patient data model which can be based on archetypes. To enable the automated computation of quality indicators using ontologies and archetypes, we tested whether ArchMS and the CLIF tool can be integrated. We successfully automated the process of generating SPARQL queries from quality indicators that have been formalized with CLIF and integrated them into ArchMS. Hence, ontologies and archetypes can be combined for the execution of formalized quality indicators.

PMID: 28423826 [PubMed - in process]

Categories: Literature Watch

Linked Data Applications Through Ontology Based Data Access in Clinical Research.

Fri, 2017-04-21 08:22
Related Articles

Linked Data Applications Through Ontology Based Data Access in Clinical Research.

Stud Health Technol Inform. 2017;235:131-135

Authors: Kock-Schoppenhauer AK, Kamann C, Ulrich H, Duhm-Harbeck P, Ingenerf J

Abstract
Clinical care and research data are widely dispersed in isolated systems based on heterogeneous data models. Biomedicine predominantly makes use of connected datasets based on the Semantic Web paradigm. Initiatives like Bio2RDF created Resource Description Framework (RDF) versions of Omics resources, enabling sophisticated Linked Data applications. In contrast, electronic healthcare records (EHR) data are generated and processed in diverse clinical subsystems within hospital information systems (HIS). Usually, each of them utilizes a relational database system with a different proprietary schema. Semantic integration and access to the data is hardly possible. This paper describes ways of using Ontology Based Data Access (OBDA) for bridging the semantic gap between existing raw data and user-oriented views supported by ontology-based queries. Based on mappings between entities of data schemas and ontologies data can be made available as materialized or virtualized RDF triples ready for querying and processing. Our experiments based on CentraXX for biobank and study management demonstrate the advantages of abstracting away from low level details and semantic mediation. Furthermore, it becomes clear that using a professional platform for Linked Data applications is recommended due to the inherent complexity, the inconvenience to confront end users with SPARQL, and scalability and performance issues.

PMID: 28423769 [PubMed - in process]

Categories: Literature Watch

Improving data workflow systems with cloud services and use of open data for bioinformatics research.

Wed, 2017-04-19 07:17

Improving data workflow systems with cloud services and use of open data for bioinformatics research.

Brief Bioinform. 2017 Apr 16;:

Authors: Karim MR, Michel A, Zappa A, Baranov P, Sahay R, Rebholz-Schuhmann D

Abstract
Data workflow systems (DWFSs) enable bioinformatics researchers to combine components for data access and data analytics, and to share the final data analytics approach with their collaborators. Increasingly, such systems have to cope with large-scale data, such as full genomes (about 200 GB each), public fact repositories (about 100 TB of data) and 3D imaging data at even larger scales. As moving the data becomes cumbersome, the DWFS needs to embed its processes into a cloud infrastructure, where the data are already hosted. As the standardized public data play an increasingly important role, the DWFS needs to comply with Semantic Web technologies. This advancement to DWFS would reduce overhead costs and accelerate the progress in bioinformatics research based on large-scale data and public resources, as researchers would require less specialized IT knowledge for the implementation. Furthermore, the high data growth rates in bioinformatics research drive the demand for parallel and distributed computing, which then imposes a need for scalability and high-throughput capabilities onto the DWFS. As a result, requirements for data sharing and access to public knowledge bases suggest that compliance of the DWFS with Semantic Web standards is necessary. In this article, we will analyze the existing DWFS with regard to their capabilities toward public open data use as well as large-scale computational and human interface requirements. We untangle the parameters for selecting a preferable solution for bioinformatics research with particular consideration to using cloud services and Semantic Web technologies. Our analysis leads to research guidelines and recommendations toward the development of future DWFS for the bioinformatics research community.

PMID: 28419324 [PubMed - as supplied by publisher]

Categories: Literature Watch

Pages