Skip to content
andimou edited this page Jan 30, 2017 · 5 revisions

Task 3

… of the Semantic Publishing Challenge 2017.

Motivation

Information about workshops, their proceedings and conferences, as well as their papers and their authors published in CEUR-WS.org might exist in other datasets, too. Our goal is to interlink such entities with the same entities as they appear in other datasets.

Persons acting e.g as authors of a publication or editors of a workshop, and their affiliations might already appear on other datasets of LOD, e.g. DBLP. Similarly, events, as conferences and workshops, might also appear on the aforementioned datasets or at COLINDA and their venue at DBpedia. All those entities should be identified and interlinked.

This way, the knowledge regarding CEUR-WS proceedings is extended beyond the dataset's boundaries. Participants are required to identify the entities that appear in CEUR-WS dataset in other datasets too and interlink the CEUR-WS.org linked dataset with relevant datasets already existing at the Linked Open Data cloud. Moreover, as triples are generated from different sources and due to different activities, tracking provenance information becomes increasingly important.

Training Datasets

The input datasets consists of a set of datasets, taken from the Linked Open Data cloud and interlinked with the CEUR-WS dataset.

CEUR-WS

Metadata about CEUR-WS.org Workshop Proceedings

homepage: http://ceur-ws.org/

data dump: CEUR-Task-1, CEUR-Task-2

Triple Pattern Fragments: http://data.linkeddatafragments.org/ceur-ws

COLINDA

Metadata about events announced at Eventseer and WikiCfP

homepage: http://www.colinda.org/

data dump: https://github.com/ceurws/lod/blob/master/data/colinda.nt

endpoint: http://data.colinda.org/endpoint.html

Triple Pattern Fragments: http://data.linkeddatafragments.org/colinda

DBLP

Metadata about Computer Science publications in the DBLP collection

homepage: http://dblp.l3s.de/dblp++.php

data dump: http://dblp.l3s.de/dblp.rdf.gz

endpoint: http://dblp.l3s.de/d2r/sparql

Triple Pattern Fragments: http://data.linkeddatafragments.org/dblp

Scholarly data

Scholarlydata dataset is a refactoring of the Semantic Web Dog Food (SWDF)

homepage: http://www.scholarlydata.org/

data dump: http://www.scholarlydata.org/dumps/workshops/simple/ AND http://www.scholarlydata.org/dumps/conferences/simple/

endpoint: http://www.scholarlydata.org/sparql/

Triple Pattern Fragments: http://data.linkeddatafragments.org/scholarlydata

Springer LD

Metadata about conference proceedings published at Springer

homepage: http://lod.springer.com

data dump: http://lod.springer.com/data/dump/lodspringerdump-20170101060027728.zip

endpoint: http://lod.springer.com/sparql

Triple Pattern Fragments: http://data.linkeddatafragments.org/springer

Queries

Participants are required to produce a dataset for answering the following queries:

  • Q3.1: Same person - Multiple URIs
  • Q3.2: Same conference - Multiple URIs
  • Q3.3: Same cited paper - Multiple URIs
  • Q3.4: Same people - Different URIs in CEUR-WS subsets
  • Q3.5: Same workshops in the CEUR-WS and Colinda datasets
  • Q3.6: Same workshops in the CEUR-WS and DBLP datasets
  • Q3.7: Same people in the CEUR-WS and DBLP datasets
  • Q3.8: Cited papers in CEUR dataset presented at conferences of Springer dataset

More details are available at detailed rules. The same page describes to rules which the output of each query should be compliant to.

These queries have to be translated in SPARQL according to the challenge's general rules.