Warning
This page is located in archive. Go to the latest version of this course pages. Go the latest version of this page.

Checkpoint 3 (max 15 pts)


Deadline 8. 1. 2022

Goal

Last checkpoint combines and integrates data through the integration of models. As a prove of interconnected datasets, you will formulate non-trivial SPARQL query answering the scientific question.

Deliverable

  • create a mapping between domain ontology of conceptual model and ontologies of datasets, using OWL, SPARQL and/or SHACL and upload it to the repository,
  • prove the interconnection of data by the sample integration of various knowledge among datasets with high number of mutually interlinked resource from all the datasets, delivered as RDF file(s) or as content of your GraphDB repository,
  • design non-trivial SPARQL queries over the datasets you create, showing the integration capabilities of the integrated datasets by answering the question given in the Checkpoint 1 (include how long queries run, results and their interpretation),
  • and include them into the description (1-2 page extension of the report from Checkpoint 2). In the description also sum up the design decisions you made, the pros/cons of the ontology, description and evaluation of SPARQL queries and conclusion you make out of the semestral work.

In the last tutorial, everyone takes 5 minutes presentation showing the results.

Details

Integration of annotated and well-described data is done by the interconnection of the ontologies describing them. By the beginning of the checkpoint, you have separated ontologies describing datasets and ono or more ontologies describing the domain. Now it is time to interconnect them, using OWL statements and rules, SHACL shapes and/or SPARQL queries.

Mapping may be done in a standalone RDF file importing all ontologies. To interconnect concepts you may use simple relations such as subClassOf, subPropertyOf, sameAs (think twice or more times before using it, remember its symmetry), but you may also need to set up rules or even create new classes and properties (e.g. class inferring all instances of some classes with specific attribute value). You may create the mapping even on the dataset ontology level, but do not get lost.

As the ontologies are interconnected, prove the interconnection of data by returning a big amount of resources with knowledge from the various datasets. Do that either by consolidation of data in the triple store, or by the SPARQL query. Export the output into RDF serialization and deliver (upload to the repository).

Now formulate non-trivial SPARQL queryt or a combination of non-trivial queries over an OWL/RDF representation of the integrated datasets to answer the given question. Remember, do not query the data itself, but the concepts they represent. Practically meaning, in the query, there shall be no specific identifiers of the specific data, but you shall ask for the instances of a specific class, eventually having connections to other classes or specific attribute values.

Presentation of the answer is also rated (by maximum 5 points). Come with a creative way of presentation, e.g. web application allowing user to set the parameters and run query, showing results in a map, or nice visualization of data. Remember, customer does not care how difficult it is to process the data, but how it looks like in the final presentation.

courses/b4m36osw/cp3.txt · Last modified: 2022/09/12 10:42 by medmicha