What is it about?
A nonstandard evaluation of the Waterloo spam rankings for the ClueWeb datasets. A standard evaluation measures the effectiveness improvement gained by the spam filtering. This one uses qrels (manually judged document-query pairs) as the ground truth. Binary classification (spam vs. non-spam) accuracy is measured/reported and the results are discussed from a different perspective.
Photo by Hannes Johnson on Unsplash
Why is it important?
It is argued that spam scores (ClueWeb09) are actually document-quality metrics (e.g. PageRank). Eliminating 70% of the corpus means working with the highest quality (30%) subset of the full dataset.
Read the Original
This page is a summary of: An intrinsic evaluation of the Waterloo spam rankings of the ClueWeb09 and ClueWeb12 datasets, Journal of Information Science, August 2019, SAGE Publications, DOI: 10.1177/0165551519866551.
You can read the full text:
The following have contributed to this page