News (08/11/2019): Ground truths are now open. All results available online. Challenge winners announced.

Semantic Web Challenge on Tabular Data to Knowledge Graph Matching

Tabular data in the form of CSV files is the common input format in a data analytics pipeline. However a lack of understanding of the semantic structure and meaning of the content may hinder the data analytics process. Thus gaining this semantic understanding will be very valuable for data integration, data cleaning, data mining, machine learning and knowledge discovery tasks. For example, understanding what the data is can help assess what sorts of transformation are appropriate on the data.

Tables on the Web may also be the source of highly valuable data. The addition of semantic information to Web tables may enhance a wide range of applications, such as web search, question answering, and knowledge base (KB) construction.

Tabular data to Knowledge Graph (KG) matching is the process of assigning semantic tags from Knowledge Graphs (e.g., Wikidata or DBpedia) to the elements of the table. This task however is often difficult in practice due to metadata (e.g., table and column names) being missing, incomplete or ambiguous.

This challenge aims at benchmarking systems dealing with the tabular data to KG matching problem, so as to facilitate their comparison on the same basis and the reproducibility of the results.

The 2019 edition of this challenge will be collocated with the 18th International Semantic Web Conference and the 14th International Workshop on Ontology Matching.

Results and Challenge Prizes

Results of all four rounds available here. Summary of SemTab 2019 results.

Prizes sponsored by SIRIUS and IBM Research:

Datasets and Evaluator

The challenge datasets and ground truths are now open:

You can cite the dataset as:
Oktie Hassanzadeh, Vasilis Efthymiou, Jiaoyan Chen, Ernesto Jimenez-Ruiz, and Kavitha Srinivas. (2019). SemTab2019: Semantic Web Challenge on Tabular Data to Knowledge Graph Matching - 2019 Data Sets (Version 2019) [Data set]. Zenodo.

The codes of the AICrowd evaluator are also available here.

ISWC Challenge Presentations

The results of the challenge will be presented on October 30 (11:40-12:40). See full ISWC program here. Four participating teams will also present their systems.

Presentations during the Ontology Matching workshop on October 26:

System papers

Papers published in the Vol-2553 of CEUR Workshop Proceedings.

Challenge Tasks

The challenge includes the following tasks organised into several evaluation rounds:

The challenge will be run with the support of the AICrowd platform.

NEW: please register your system details here.

NEW: we have created a discussion group for the challenge.

Support for ontology alignment and link discovery

Ontology alignment and link discovery systems are welcome to participate. Please follow the instructions for the CEA task.

Challenge Prizes

There will be prizes sponsored by SIRIUS and IBM Research for the best systems and the best student systems in the challenge.

The prize winners will be announced during the ISWC conference (on October 30, 2019).

We will take into account all evaluation rounds specially the one running till the conference dates, the covered tasks and the novelty of the applied techniques (we encourage the submission of a system paper).

Important Dates

Guidelines for System Papers

We encourage participants to submit a system paper. The paper should be no more than 8 pages long and formatted using the LNCS Style. System papers will be revised by 1-2 challenge organisers. Please use this form for the submission (requires a google account and a valid email).

To ensure easy comparability among the participants we suggest the following outline:

  1. Presentation of the system
    1. State, purpose, general statement
    2. Specific techniques used
    3. Adaptations made for the evaluation
    4. Link to the system and parameters file
  2. Results
  3. General comments (if relevant)
    1. Comments on the results (strength and weaknesses)
    2. Discussions on the way to improve the proposed system
    3. Comments on the challenge procedure
    4. Comments on the challenge test cases
    5. Comments on the challenge measures
    6. Proposal of new datasets, tasks or measures
  4. Conclusions
  5. References


Challenge chairs

This track is organised by Kavitha Srinivas (IBM Research), Ernesto Jimenez-Ruiz (City, University of London; Alan Turing Institute; University of Oslo), Oktie Hassanzadeh (IBM Research), Jiaoyan Chen (University of Oxford) and Vasilis Efthymiou (IBM Research). If you have any problems working with the datasets or any suggestions related to this challenge, do not hesitate to contact us.

Challenge committee members


The challenge is currently supported by the AIDA project, the SIRIUS Centre for Research-driven Innovation, and IBM Research.