This document summarizes a presentation on additive smoothing for relevance-based language modelling of recommender systems. It discusses using pseudo-relevance feedback and relevance models for collaborative filtering recommendations. Specifically, it examines how different collection-based smoothing techniques like Dirichlet priors, Jelinek-Mercer, and absolute discounting can demote the desired IDF effect, which promotes less popular items. The document proposes using additive smoothing, which does not demote the IDF effect. Experiments on movie recommendation datasets show additive smoothing achieves better accuracy, diversity, and novelty than other smoothing methods.
Additive Smoothing for Relevance-Based Language Modelling of Recommender Systems [CERI '16 Slides]
1. CERI 2016, GRANADA, SPAIN
ADDITIVE SMOOTHING FOR RELEVANCE-BASED
LANGUAGE MODELLING OF RECOMMENDER SYSTEMS
Daniel Valcarce, Javier Parapar, Álvaro Barreiro
@dvalcarce @jparapar @AlvaroBarreiroG
Information Retrieval Lab
@IRLab_UDC
University of A Coruña
Spain
2. Outline
1. Recommender Systems
2. Pseudo-Relevance Feedback
3. Relevance-Based Language Modelling of Recommender
Systems
4. IDF Effect and Additive Smoothing
5. Experiments
6. Conclusions and Future Directions
1/26
4. Recommender Systems
Recommender systems generate personalised suggestions for
items that may be of interest to the users.
Top-N Recommendation: create a ranking of the N most
relevant items for each user.
Collaborative filtering: exploit only user-item interactions
(ratings, clicks, etc.).
3/26
6. Pseudo-Relevance Feedback (I)
In Information Retrieval, Pseudo-Relevance Feedback (PRF) is
an automatic query expansion method.
The goal is to expand the original query with new terms to
improve the quality of the search results.
These new terms are extracted automatically from a first
retrieval using the original query.
5/26
17. Relevance-Based Language Models (RM)
Relevance-Based Language Models or Relevance Models (RM)
are a state-of-the-art PRF technique (Lavrenko & Croft, SIGIR
2001).
Two models: RM1 and RM2.
RM1 works better than RM2 in retrieval.
9/26
18. Relevance-Based Language Models (RM)
Relevance-Based Language Models or Relevance Models (RM)
are a state-of-the-art PRF technique (Lavrenko & Croft, SIGIR
2001).
Two models: RM1 and RM2.
RM1 works better than RM2 in retrieval.
Relevance Models have been recently adapted to collaborative
filtering (Parapar et al., IPM 2013).
For recommendation, RM2 is the preferred method.
9/26
19. Relevance Models for Collaborative Filtering
RM2 : p(i|Ru) ∝ p(i)
j∈Iu v∈Vu
p(i|v) p(v)
p(i)
p(j|v)
Iu is the set of items rated by the user u.
Vu is neighbourhood of the user u. This is computed using
a clustering algorithm.
p(i) and p(v) are the item and user priors.
p(i|u) is computed smoothing the maximum likelihood
estimate with the probability in the collection.
10/26
20. Relevance Models for Collaborative Filtering
RM2 : p(i|Ru) ∝ p(i)
j∈Iu v∈Vu
p(i|v) p(v)
p(i)
p(j|v)
Iu is the set of items rated by the user u.
Vu is neighbourhood of the user u. This is computed using
a clustering algorithm.
p(i) and p(v) are the item and user priors.
p(i|u) is computed smoothing the maximum likelihood
estimate with the probability in the collection.
10/26
21. Relevance Models for Collaborative Filtering
RM2 : p(i|Ru) ∝ p(i)
j∈Iu v∈Vu
p(i|v) p(v)
p(i)
p(j|v)
Iu is the set of items rated by the user u.
Vu is neighbourhood of the user u. This is computed
using a clustering algorithm.
p(i) and p(v) are the item and user priors.
p(i|u) is computed smoothing the maximum likelihood
estimate with the probability in the collection.
10/26
22. Relevance Models for Collaborative Filtering
RM2 : p(i|Ru) ∝ p(i)
j∈Iu v∈Vu
p(i|v) p(v)
p(i)
p(j|v)
Iu is the set of items rated by the user u.
Vu is neighbourhood of the user u. This is computed using
a clustering algorithm.
p(i) and p(v) are the item and user priors.
p(i|u) is computed smoothing the maximum likelihood
estimate with the probability in the collection.
10/26
23. Relevance Models for Collaborative Filtering
RM2 : p(i|Ru) ∝ p(i)
j∈Iu v∈Vu
p(i|v) p(v)
p(i)
p(j|v)
Iu is the set of items rated by the user u.
Vu is neighbourhood of the user u. This is computed using
a clustering algorithm.
p(i) and p(v) are the item and user priors.
p(i|u) is computed smoothing the maximum likelihood
estimate with the probability in the collection.
10/26
25. Collection-based Smoothing Techniques (II)
Absolute Discounting, Jelinek-Mercer and Dirichlet Priors have
been studied in the context of:
Text Retrieval (Zhai & Lafferty, ACM TOIS 2004)
Collaborative Filtering (Valcarce et al., ECIR 2015)
12/26
26. Collection-based Smoothing Techniques (II)
Absolute Discounting, Jelinek-Mercer and Dirichlet Priors have
been studied in the context of:
Text Retrieval (Zhai & Lafferty, ACM TOIS 2004)
◦ Absolute Discounting performs very poorly.
◦ Dirichlet Priors is the most popular approach.
◦ Jelinek-Mercer is a bit better for long queries.
Collaborative Filtering (Valcarce et al., ECIR 2015)
◦ Absolute Discounting is the best smoothing method.
12/26
27. Collection-based Smoothing Techniques (II)
Absolute Discounting, Jelinek-Mercer and Dirichlet Priors have
been studied in the context of:
Text Retrieval (Zhai & Lafferty, ACM TOIS 2004)
◦ Absolute Discounting performs very poorly.
◦ Dirichlet Priors is the most popular approach.
◦ Jelinek-Mercer is a bit better for long queries.
Collaborative Filtering (Valcarce et al., ECIR 2015)
◦ Absolute Discounting is the best smoothing method.
Can we do better?
12/26
29. Axiomatic Analysis of the IDF Effect in IR
A recent work performed an axiomatic analysis of several PRF
methods (Hazimeh & Zhai, ICTIR 2015).
They found out that RM1 with Dirichlet Priors and
Jelinek-Mercer smoothing methods demote the IDF effect.
The IDF effect is a desirable property that, intuitively,
promotes documents with very specific terms.
14/26
30. Axiomatic Analysis of the IDF Effect in IR
A recent work performed an axiomatic analysis of several PRF
methods (Hazimeh & Zhai, ICTIR 2015).
They found out that RM1 with Dirichlet Priors and
Jelinek-Mercer smoothing methods demote the IDF effect.
The IDF effect is a desirable property that, intuitively,
promotes documents with very specific terms.
Can we use this result in recommendation?
What is the IDF effect in recommendation? Is it a desirable
property?
They studied RM1, what about RM2?
14/26
31. The IDF Effect in Recommendation (I)
This retrieval idea is related to the novelty in recommendation.
Definition (IDF effect)
A recommender system supports the IDF effect if p(i1|Ru) >
p(i2|Ru) when
two items i1 and i2
have the same ratings r(v, i1) r(v, i2) for all v ∈ Vu
and different popularity p(i1|C) < p(i2|C)
In simply words, if we have the same feedback for two items,
we should recommend the least popular one.
15/26
32. The IDF Effect in Recommendation (II)
We performed an axiomatic analysis of RM21 using the
following smoothing methods:
Dirichlet Priors
Jelinek-Mercer
Absolute Discounting
1Math proofs in the paper!
16/26
33. The IDF Effect in Recommendation (II)
We performed an axiomatic analysis of RM21 using the
following smoothing methods:
Dirichlet Priors
Jelinek-Mercer
Absolute Discounting
1Math proofs in the paper!
16/26
34. The IDF Effect in Recommendation (II)
We performed an axiomatic analysis of RM21 using the
following smoothing methods:
Dirichlet Priors
Jelinek-Mercer
Absolute Discounting
Additive Smoothing
pγ(i|u)
r(u, i) + γ
j∈Iu
r(u, j) + γ|I|
1Math proofs in the paper!
16/26
42. Conclusions
The IDF effect from IR is related to the novelty of the
recommendations.
The use of collection-based smoothing methods with RM2
demotes the IDF effect.
Additive smoothing is a simple method that does not demote
(nor promote) the IDF effect.
Additive smoothing provides better accuracy, diversity and
novelty figures than collection-based smoothing methods.
24/26
43. Future work
Envision new ways of enhancing the IDF effect in RM2:
Design smoothing methods that actively promote the IDF
effect.
Use non-uniform prior estimates.
Study axiomatically other IR properties that can be useful in
recommendation.
25/26