-
Notifications
You must be signed in to change notification settings - Fork 2
SemPub17_Task1
… of the Semantic Publishing Challenge 2017.
Participants are required to extract information from the tables of the papers (in PDF). Extracting content from tables is a difficult task, which has been tackled by different researchers in the past. Our focus is on tables in scientific papers and solutions for re-publishing structured data as LOD. Tables will be collected from CEUR-WS.org publications and participants will be required to identify their structure and content. The task then will require PDF mining and data processing techniques.
Datasets can be downloaded here:
- http://ceur-ws.org/Vol-782/GoerlitzAndStaab_COLD2011.pdf
- http://ceur-ws.org/Vol-905/MontoyaEtAl_COLD2012.pdf
- http://ceur-ws.org/Vol-912/paper7.pdf
- http://ceur-ws.org/Vol-996/papers/ldow2013-paper-07.pdf
- http://ceur-ws.org/Vol-1456/paper5.pdf
- http://ceur-ws.org/Vol-1457/SSWS2015_paper4.pdf
- http://ceur-ws.org/Vol-1597/PROFILES2016_paper4.pdf
- http://ceur-ws.org/Vol-1700/paper-01.pdf
- http://ceur-ws.org/Vol-1193/paper_65.pdf
- http://ceur-ws.org/Vol-846/paper_20.pdf
- http://ceur-ws.org/Vol-1193/paper_62.pdf
- http://ceur-ws.org/Vol-1193/paper_50.pdf
- http://ceur-ws.org/Vol-1350/paper-35.pdf
- http://ceur-ws.org/Vol-1193/paper_6.pdf
- http://ceur-ws.org/Vol-1766/oaei16_paper13.pdf
- http://ceur-ws.org/Vol-666/paper2.pdf
- http://ceur-ws.org/Vol-1350/paper-37.pdf
- http://ceur-ws.org/Vol-1265/owled2014_submission_1.pdf
- http://ceur-ws.org/Vol-858/ore2012_paper3.pdf
- http://ceur-ws.org/Vol-1193/paper_58.pdf
- http://ceur-ws.org/Vol-1387/paper_1.pdf
- http://ceur-ws.org/Vol-1193/paper_52.pdf
- http://ceur-ws.org/Vol-1766/oaei16_paper0.pdf
- http://ceur-ws.org/Vol-1545/oaei15_paper0.pdf
- http://ceur-ws.org/Vol-1317/oaei14_paper0.pdf
- http://ceur-ws.org/Vol-1111/oaei13_paper0.pdf
- http://ceur-ws.org/Vol-946/oaei12_paper0.pdf
- http://ceur-ws.org/Vol-814/oaei11_paper0.pdf
- http://ceur-ws.org/Vol-689/oaei10_paper0.pdf
- http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-CabrioEt2013.pdf
- http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-HeEt2013.pdf
- http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-CabrioEt2013.pdf
- http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-UngerEt2014.pdf
- http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-HamonEt2014.pdf
- http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-ParkEt2014.pdf
- http://ceur-ws.org/Vol-1391/173-CR.pdf
- http://ceur-ws.org/Vol-1391/164-CR.pdf
- http://ceur-ws.org/Vol-1391/164-CR.pdf
- http://ceur-ws.org/Vol-1391/173-CR.pdf
List of URLs for one-time download:
http://ceur-ws.org/Vol-782/GoerlitzAndStaab_COLD2011.pdf
http://ceur-ws.org/Vol-905/MontoyaEtAl_COLD2012.pdf
http://ceur-ws.org/Vol-912/paper7.pdf
http://ceur-ws.org/Vol-996/papers/ldow2013-paper-07.pdf
http://ceur-ws.org/Vol-1456/paper5.pdf
http://ceur-ws.org/Vol-1457/SSWS2015_paper4.pdf
http://ceur-ws.org/Vol-1597/PROFILES2016_paper4.pdf
http://ceur-ws.org/Vol-1700/paper-01.pdf
http://ceur-ws.org/Vol-1193/paper_65.pdf
http://ceur-ws.org/Vol-846/paper_20.pdf
http://ceur-ws.org/Vol-1193/paper_62.pdf
http://ceur-ws.org/Vol-1193/paper_50.pdf
http://ceur-ws.org/Vol-1350/paper-35.pdf
http://ceur-ws.org/Vol-1193/paper_6.pdf
http://ceur-ws.org/Vol-1766/oaei16_paper13.pdf
http://ceur-ws.org/Vol-666/paper2.pdf
http://ceur-ws.org/Vol-1350/paper-37.pdf
http://ceur-ws.org/Vol-1265/owled2014_submission_1.pdf
http://ceur-ws.org/Vol-858/ore2012_paper3.pdf
http://ceur-ws.org/Vol-1193/paper_58.pdf
http://ceur-ws.org/Vol-1387/paper_1.pdf
http://ceur-ws.org/Vol-1193/paper_52.pdf
http://ceur-ws.org/Vol-1766/oaei16_paper0.pdf
http://ceur-ws.org/Vol-1545/oaei15_paper0.pdf
http://ceur-ws.org/Vol-1317/oaei14_paper0.pdf
http://ceur-ws.org/Vol-1111/oaei13_paper0.pdf
http://ceur-ws.org/Vol-946/oaei12_paper0.pdf
http://ceur-ws.org/Vol-814/oaei11_paper0.pdf
http://ceur-ws.org/Vol-689/oaei10_paper0.pdf
http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-CabrioEt2013.pdf
http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-HeEt2013.pdf
http://ceur-ws.org/Vol-1179/CLEF2013wn-QALD3-CabrioEt2013.pdf
http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-UngerEt2014.pdf
http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-HamonEt2014.pdf
http://ceur-ws.org/Vol-1180/CLEF2014wn-QA-ParkEt2014.pdf
http://ceur-ws.org/Vol-1391/173-CR.pdf
http://ceur-ws.org/Vol-1391/164-CR.pdf
http://ceur-ws.org/Vol-1391/164-CR.pdf
http://ceur-ws.org/Vol-1391/173-CR.pdf
The following ZIP file contains the expected output of all queries on all papers in the training dataset: sempub17-TD1.zip
The archive contains the full list of queries (in QUERIES-LIST.csv
) and the output of each of them in a separate .csv file.
For each query there is an entry in QUERIES-LIST.csv
indicating the identifier of the query and the natural language description. The output of that query is contained in the corresponding .csv file, as shown below:
QueryID | Natural language description | CSV output file |
---|---|---|
Q1.1 | For each dataset D of FedBench in paper P, find the number of subjects X: X=subjects, P=http://ceur-ws.org/Vol-782/GoerlitzAndStaab_COLD2011.pdf | Q1.1.csv |
The final evaluation will be performed by using SemPubEvaluator on a set of papers to be disclosed a few days before the deadline.
Participants are required to produce a dataset for answering the following queries.
- Q1.1 (Subjects of a dataset in a paper): For each dataset D of FedBench in paper P, find the number of subjects X.
- Q1.2 (Test cases): Which are the test cases T for OAEI in year Y?
- Q1.3 (Common test cases of two editions): Which test cases T did all editions have in common from year Y1 to Y2?
- Q1.4 (Persistant participants): Which participants P participated in all editions from Y1 to Y2?
- Q1.5 (Repeated test cases): Which participants P have addressed test case T for every edition from Y1 to Y2?
- Q1.6 (Information of datasets): How many X did test dataset D have in year Y?
- Q1.7 (Information of systems for datasets): What was the X of system S in year Y for test dataset D?
- Q1.8 (Systems best performance): Which system S in year Y had the best X ever?
- Q1.9 (All tools ever mentioned in one edition of QALD): FList the names of all systems that have ever been used in an experiment in an edition of QALD.
- Q1.10 (All tools ever mentioned in one edition of QALD): For Multilingual QA in the QALD challenge, what system in what edition had the best precision, and what system in what edition had the worst recall?
These queries have to be translated in SPARQL according to the challenge's general rules and have to produce an output according to the detailed rules.