Difference between revisions of "Task1"

From BITPlan cr Wiki
Jump to navigation Jump to search
Line 18: Line 18:
 
on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by
 
on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by
 
similar queries that only relied on information available from HTML sources.
 
similar queries that only relied on information available from HTML sources.
 +
|since=2015-08-25
 
}}
 
}}
 
=Freitext=
 
=Freitext=

Revision as of 06:43, 21 March 2023

Task

Task
id  1
title  Extraction and Assessment of Workshop Proceedings Information
objective  Common questions related to the quality of a scientific workshop or conference include whether a researcher should submit a paper to it or accept an invitation

to its program committee, whether a publisher should publish its proceedings, or whether a company should sponsor it [2]. Moreover, knowing the quality of an event helps to assess the quality of the papers accepted there. In the 2014 Challenge, we had designed Task 1 to extract from selected CEUR-WS.org work- shop proceedings volumes RDF that would enable the computation of certain indicators for the workshops’ quality [10]. The second objective of this effort was to bootstrap the publication of all CEUR-WS.org workshops – more than 1,400 at the time of this writing – as linked data. As discussed above in Section 2, we reused the 2014 queries, with two exceptions. As only one of the three 2014 submissions had addressed the two Task 1 queries that required metadata ex- traction from the PDF full text of the papers (cf. [7]), and as Task 2 focused on full-text extraction anyway, we replaced these queries (Q1.19 and Q1.20) by similar queries that only relied on information available from HTML sources.

since  2015-08-25

Freitext