Quality Metrics in Recommender Systems: Do We Calculate Metrics Consistently?

06/26/2022
by   Yan-Martin Tamm, et al.
0

Offline evaluation is a popular approach to determine the best algorithm in terms of the chosen quality metric. However, if the chosen metric calculates something unexpected, this miscommunication can lead to poor decisions and wrong conclusions. In this paper, we thoroughly investigate quality metrics used for recommender systems evaluation. We look at the practical aspect of implementations found in modern RecSys libraries and at the theoretical aspect of definitions in academic papers. We find that Precision is the only metric universally understood among papers and libraries, while other metrics may have different interpretations. Metrics implemented in different libraries sometimes have the same name but measure different things, which leads to different results given the same input. When defining metrics in an academic paper, authors sometimes omit explicit formulations or give references that do not contain explanations either. In 47 metric is defined because the definition is not clear or absent. These findings highlight yet another difficulty in recommender system evaluation and call for a more detailed description of evaluation protocols.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/14/2023

Bridging Offline-Online Evaluation with a Time-dependent and Popularity Bias-free Offline Metric for Recommenders

The evaluation of recommendation systems is a complex task. The offline ...
research
09/12/2023

Distributionally-Informed Recommender System Evaluation

Current practice for evaluating recommender systems typically focuses on...
research
09/18/2022

Offline Evaluation of Reward-Optimizing Recommender Systems: The Case of Simulation

Both in academic and industry-based research, online evaluation methods ...
research
04/15/2023

More Is Less: When Do Recommenders Underperform for Data-rich Users?

Users of recommender systems tend to differ in their level of interactio...
research
08/28/2017

It's Time to Consider "Time" when Evaluating Recommender-System Algorithms [Proposal]

In this position paper, we question the current practice of calculating ...
research
10/11/2018

A Distributed and Accountable Approach to Offline Recommender Systems Evaluation

Different software tools have been developed with the purpose of perform...
research
02/20/2020

The Problem with Metrics is a Fundamental Problem for AI

Optimizing a given metric is a central aspect of most current AI approac...

Please sign up or login with your details

Forgot password? Click here to reset