You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For all systems, we report results from parameter tuning to optimize average precision (AP) at rank 1000 on the newswire collections, WT10g, and Gov2, and NDCG@20 for the ClueWeb collections.
There was no separation of training and test data, so these results should be interpreted as oracle settings.
If you're going to refer to these effectiveness results, please be aware of what you're comparing!
Additional note: The values produced by these scripts are slightly different than those reported in the article.
The reason for these differences stems from the fact that Anserini evolved throughout the peer review process; the values reported in the article were those generated when the manuscript was submitted.
By the time the article was published, the implementation of Anserini has progressed.
As Anserini continues to improve we will update these scripts, which will lead to further divergences between the published values.
Unfortunately, this is an unavoidable aspect of empirical research on software artifacts.
Update (12/18/2018):
Regression effectiveness values changed at commit e71df7a with upgrade to Lucene 7.6.
Update (6/12/2019):
With commit 75e36f9, which upgrades Anserini to Lucene 8.0, we are no longer maintaining the reproducibility of these experiments.
That is, running these commands will produce results different from the numbers reported here.
The most recent version in which these results are reproduce is the v0.5.1 release (6/11/2019).
Parameter Tuning
Invoke the tuning script on various collections as follows, on tuna: