@griesbau

Experimental evaluation and hypertexts

. Proceedings of the First Workshop on Personalised Multilingual Hypertext Retrieval, page 8--9. New York, NY, USA, ACM, (2011)
DOI: 10.1145/2047403.2047406

Abstract

The information retrieval field has a strong and long tradition, that dates back to the sixties of the past century, in the experimental evaluation of <i>Information Retrieval (IR)</i> systems in order to assess their performances in a scientifically sound and comparable way. In this context, large-scale international evaluation campaigns, such as <i>Text REtrieval Conference (TREC)</i>, <i>Cross-Language Evaluation Forum (CLEF)</i>, and <i>NII-NACSIS Test Collection for IR Systems (NTCIR)</i>, have been the vehicle and the conduit for the advancement of state-of-the-art techniques and for the development of innovative information systems through common evaluation procedures, regular and systematic evaluation cycles, comparison and benchmarking of the adopted approaches and solutions, spreading and exchange of knowledge and know-how.</p> <p>Hypertexts play an important role in the information retrieval field, especially since the growth of the Web has given raise to an unprecedented need for effective information access techniques that take into consideration the multilinguality, multimodality, and hypertextual nature of the relevant information resources. This posed novel challenges for experimental evaluation which has to devise techniques for coping with experimental collections able to mimic the Web scale and for designing evaluation tasks that were representative of user needs on the Web.</p> <p>This talk will discuss open issues concerning how experimental evaluation and hypertext could be better benefit each other. On the one hand, it is time for experimental evaluation to explicitly take into consideration the hypertextual nature of the resources when assessing performances based on retrieved items and not only considering systems as black-boxes that internally exploit the exiting hypertext. On the other hand, experimental evaluation produces huge amount of scientific data that would be better understood and interpreted if they were enriched with links to each other, to other resources, and to user-generated content, such as annotations explaining them.

Description

Experimental evaluation and hypertexts

Links and resources

Tags

community

  • @griesbau
  • @dblp
@griesbau's tags highlighted