Difference between revisions of "Task1"

From BITPlan cr Wiki
Jump to navigation Jump to search
Line 4: Line 4:
 
|id=1
 
|id=1
 
|title=Extraction and Assessment of Workshop Proceedings Information
 
|title=Extraction and Assessment of Workshop Proceedings Information
 +
|objective=Common questions related to the quality of a scientific workshop or conference
 +
include whether a researcher should submit a paper to it or accept an invitation
 +
to its program committee, whether a publisher should publish its proceedings,
 +
or whether a company should sponsor it [2]. Moreover, knowing the quality of
 +
an event helps to assess the quality of the papers accepted there. In the 2014
 +
Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work-
 +
shop proceedings volumes RDF that would enable the computation of certain
 +
indicators for the workshops’ quality [10]. The second objective of this effort was
 +
to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400
 +
at the time of this writing – as linked data. As discussed above in Section 2,
 +
we reused the 2014 queries, with two exceptions. As only one of the three 2014
 +
submissions had addressed the two Task 1 queries that required metadata ex-
 +
traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused
 +
on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by
 +
similar queries that only relied on information available from HTML sources.
 
}}
 
}}
 
=Freitext=
 
=Freitext=

Revision as of 11:41, 19 March 2023

Task

Task
id  1
title  Extraction and Assessment of Workshop Proceedings Information
objective  Common questions related to the quality of a scientific workshop or conference

include whether a researcher should submit a paper to it or accept an invitation to its program committee, whether a publisher should publish its proceedings, or whether a company should sponsor it [2]. Moreover, knowing the quality of an event helps to assess the quality of the papers accepted there. In the 2014 Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work- shop proceedings volumes RDF that would enable the computation of certain indicators for the workshops’ quality [10]. The second objective of this effort was to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400 at the time of this writing – as linked data. As discussed above in Section 2, we reused the 2014 queries, with two exceptions. As only one of the three 2014 submissions had addressed the two Task 1 queries that required metadata ex- traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by similar queries that only relied on information available from HTML sources.

since  

Freitext