Difference between revisions of "Task1"
Line 4: | Line 4: | ||
|id=1 | |id=1 | ||
|title=Extraction and Assessment of Workshop Proceedings Information | |title=Extraction and Assessment of Workshop Proceedings Information | ||
+ | |objective=Common questions related to the quality of a scientific workshop or conference | ||
+ | include whether a researcher should submit a paper to it or accept an invitation | ||
+ | to its program committee, whether a publisher should publish its proceedings, | ||
+ | or whether a company should sponsor it [2]. Moreover, knowing the quality of | ||
+ | an event helps to assess the quality of the papers accepted there. In the 2014 | ||
+ | Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work- | ||
+ | shop proceedings volumes RDF that would enable the computation of certain | ||
+ | indicators for the workshops’ quality [10]. The second objective of this effort was | ||
+ | to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400 | ||
+ | at the time of this writing – as linked data. As discussed above in Section 2, | ||
+ | we reused the 2014 queries, with two exceptions. As only one of the three 2014 | ||
+ | submissions had addressed the two Task 1 queries that required metadata ex- | ||
+ | traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused | ||
+ | on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by | ||
+ | similar queries that only relied on information available from HTML sources. | ||
}} | }} | ||
=Freitext= | =Freitext= |
Revision as of 11:41, 19 March 2023
Task
Task | |
---|---|
id | 1 |
title | Extraction and Assessment of Workshop Proceedings Information |
objective | Common questions related to the quality of a scientific workshop or conference
include whether a researcher should submit a paper to it or accept an invitation to its program committee, whether a publisher should publish its proceedings, or whether a company should sponsor it [2]. Moreover, knowing the quality of an event helps to assess the quality of the papers accepted there. In the 2014 Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work- shop proceedings volumes RDF that would enable the computation of certain indicators for the workshops’ quality [10]. The second objective of this effort was to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400 at the time of this writing – as linked data. As discussed above in Section 2, we reused the 2014 queries, with two exceptions. As only one of the three 2014 submissions had addressed the two Task 1 queries that required metadata ex- traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by similar queries that only relied on information available from HTML sources. |
since |