Difference between revisions of "Task1"

From BITPlan cr Wiki
Jump to navigation Jump to search
Line 4: Line 4:
 
|id=1
 
|id=1
 
|title=Extraction and Assessment of Workshop Proceedings Information
 
|title=Extraction and Assessment of Workshop Proceedings Information
|objective=Common questions related to the quality of a scientific workshop or conference
+
|objective=Common questions related to the quality of a scientific workshop or conference include whether a researcher should submit a paper to it or accept an invitation
include whether a researcher should submit a paper to it or accept an invitation
 
 
to its program committee, whether a publisher should publish its proceedings,
 
to its program committee, whether a publisher should publish its proceedings,
 
or whether a company should sponsor it [2]. Moreover, knowing the quality of
 
or whether a company should sponsor it [2]. Moreover, knowing the quality of

Revision as of 11:41, 19 March 2023

Task

Task
id  1
title  Extraction and Assessment of Workshop Proceedings Information
objective  Common questions related to the quality of a scientific workshop or conference include whether a researcher should submit a paper to it or accept an invitation

to its program committee, whether a publisher should publish its proceedings, or whether a company should sponsor it [2]. Moreover, knowing the quality of an event helps to assess the quality of the papers accepted there. In the 2014 Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work- shop proceedings volumes RDF that would enable the computation of certain indicators for the workshops’ quality [10]. The second objective of this effort was to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400 at the time of this writing – as linked data. As discussed above in Section 2, we reused the 2014 queries, with two exceptions. As only one of the three 2014 submissions had addressed the two Task 1 queries that required metadata ex- traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by similar queries that only relied on information available from HTML sources.

since  

Freitext