Coordinator: Prof. Dimitris Plexousakis
During last years we have witnessed a tremendous increase in the amount of semantic data that is available on the Web in almost every field of human activity (e.g. Wikipedia, U.S. Census, CIA World Factbook, open government sites in the US and the UK, news and entertainment sources, e-science ontologies and data) have been created and published online. Semantic data management refers to a range of techniques that can be employed for storing, querying, manipulating and integrating data based on its meaning. Current ISL activities in this area focus on three main topics: provenance, large scale semantic integration and the management of the evolution of semantically-enriched datasets (change detection, evolution management, inconsistency handling).
Provenance means the origin of data, and refers from where and how the piece of data was obtained and it is critical in applications such as trust assessment, access control, data cleaning and data/information quality. ISL activities in the domain of provenance for semantic web data target the following research activities:
Since the traditional data integration approaches cannot scale easily to large number of sources, ISL also focuses on methods for Large Scale Semantic Integration including:
Datasets found on the web exhibit a highly volatile and dynamic behaviour, as the information they encode is constantly subject to change, due to new data, erroneous inclusion of information, the discovery of inconsistencies etc. In this setting, several problems arise, including:
In support of national Precision Medicine Networks, the laboratory designed a full-fledged ICT ecosystem that supports the integration of clinical, -omics and imaging data. It comprises of a suite of tools for managing clinical, genetic, molecular, biochemical, and epidemiological parameters related to the development and evolution of diseases that have a genetic cause.