What is it about?
When you browse or click on items online, recommendation systems use that behavior to suggest things you might like. But not every click reflects genuine interest — sometimes people click by accident, or are tricked by misleading thumbnails. This "noise" confuses the system and leads to worse recommendations. Most existing methods try to clean up this noise by looking at each individual click, but they miss a key insight: some users are naturally more random in their behavior, and some items are inherently more likely to attract meaningless clicks. Our method, called EARD, takes a different approach. Instead of only examining each click in isolation, it evaluates the overall reliability of each user and each item based on how they behave across all their interactions. It then combines this "reputation score" with information about individual clicks to decide how much to trust each piece of data during training. The method is lightweight, works with any recommendation model, and only needs two simple settings to tune. Experiments on three large real-world datasets show it significantly outperforms existing approaches — improving recommendation accuracy by up to 27% — while adding almost no extra computational cost.
Featured Image
Photo by Elin Melaas on Unsplash
Why is it important?
This work addresses a fundamental bottleneck in real-world recommendation systems: noisy implicit feedback degrades the quality of suggestions that billions of users receive daily on platforms like e-commerce sites, streaming services, and social media. Unlike previous solutions that are either too computationally expensive for industrial use or require extensive manual tuning, EARD introduces a practical framework that is efficient, easy to deploy, and model-agnostic. Its key innovation — modeling noise at the entity level (users and items) rather than just the interaction level — opens a new direction for building more robust recommender systems. The method's minimal overhead and strong empirical results across diverse settings make it immediately applicable to large-scale production environments where both accuracy and efficiency matter.
Perspectives
This project started from a simple but often overlooked observation: not all users and items are equally trustworthy. While previous denoising research focused on individual interactions, we noticed that some users consistently produce noisy signals and some items systematically attract unreliable clicks. Once we confirmed this pattern empirically across multiple datasets, the design of EARD followed naturally. What excites us most is the framework's simplicity — it requires only two hyperparameters and integrates seamlessly into existing training pipelines. We believe this kind of lightweight, principled approach is what the recommendation community needs to bridge the gap between academic research and industrial deployment. We hope EARD inspires further work on entity-aware modeling and encourages the field to prioritize practical scalability alongside performance gains.
Ze Liu
University of Science and Technology of China
Read the Original
This page is a summary of: From Entity Reliability to Clean Feedback: An Entity-Aware Denoising Framework Beyond Interaction-Level Signals, April 2026, ACM (Association for Computing Machinery),
DOI: 10.1145/3774904.3792128.
You can read the full text:
Resources
EARD Source Code
Official implementation of the EARD framework in Python. Includes training scripts, evaluation code, and configurations for all experiments on Yelp, Amazon-Book, and ML-1M datasets.
EARD Paper — ACM Digital Library (Open Access)
Full paper published at The ACM Web Conference 2026 (WWW '26), licensed under Creative Commons Attribution 4.0 International (CC BY 4.0).
Contributors
The following have contributed to this page







