\ RecSys 2023 - Additional Information and Results

RecSys 2023 - Singapore
Everyone’s a Winner! On Hyperparameter Tuning of Recommendation Models
Additional Information: Source Code, Optimized Hyper-Parameters and Additional Result Tables


Faisal Shehzad
faisal.shehzad@aau.at
Dietmar Jannach
dietmar.jannach@aau.at

The performance of a recommender system algorithm in terms of common offline accuracy measures often strongly depends on the chosen hyperparameters. Therefore, when comparing algorithms in offline experiments, we can obtain reliable insights regarding the effectiveness of a newly proposed algorithm only if we compare it to a number of state-of-the-art baselines that are carefully tuned for each of the considered datasets. While this fundamental principle of any area of applied machine learning is undisputed, we find that the tuning process for the baselines in the current literature is barely documented in much of today's published research. Ultimately, in case the baselines are actually not carefully tuned, progress may remain unclear. In this paper, we showcase how every method in such an unsound comparison can be reported to be outperforming the state-of-the-art. Finally, we iterate appropriate research practices to avoid unreliable algorithm comparisons in the future.

Source Code and Datasets

The full source code of the framework and datasets can be found here:
https://github.com/Faisalse/RecSys2023_hyperparameter_tuning

aaaaa
ccccc
ddddd
iiiiii
fffff
ggggg
hhhhh
iiiiii
aaaaa
bbbbb
ccccc
ddddd