Evaluating non-deterministic retrieval systems


Autoria(s): Jayasinghe,GK; Webber,W; Sanderson,M; Dharmasena,LS; Culpepper,JS
Data(s)

01/01/2014

Resumo

The use of sampling, randomized algorithms, or training based on the unpredictable inputs of users in Information Retrieval often leads to non-deterministic outputs. Evaluating the effectiveness of systems incorporating these methods can be challenging since each run may produce different effectiveness scores. Current IR evaluation techniques do not address this problem. Using the context of distributed information retrieval as a case study for our investigation, we propose a solution based on multivariate linear modeling. We show that the approach provides a consistent and reliable method to compare the effectiveness of non-deterministic IR algorithms, and explain how statistics can safely be used to show that two IR algorithms have equivalent effectiveness. Copyright 2014 ACM.

Identificador

http://hdl.handle.net/10536/DRO/DU:30070462

Idioma(s)

eng

Publicador

ACM

Relação

http://dro.deakin.edu.au/eserv/DU:30070462/dharmasena-evaluatingnon-2014.pdf

http://www.dx.doi.org/10.1145/2600428.2609472

Direitos

2014, Association for Computing Machinery

Palavras-Chave #Effectiveness evaluation #Experimental design #Experimentation #Information retrieval #Measurement #Statistical analysis
Tipo

Conference Paper