Upgrade to Pro — share decks privately, control downloads, hide ads and more …

Recommendations under sparsity

Recommendations under sparsity

In this talk, I look at the advantages and disadvantages of collaborative filtering and content-based recommenders when interaction data is sparse, and describe a hybrid approach implemented in the LightFM package.

Maciej Kula

October 06, 2015
Tweet

More Decks by Maciej Kula

Other Decks in Programming

Transcript

  1. Represent a user-item interaction matrix as a product of two

    lower-dimensional matrices to fill in the missing entries. Matrix Factorisation user = (-0.3, 2.1, 0.5) product = (5.2, 0.3, -0.5)
  2. 0.91 ROC AUC (probability that a randomly chosen item the

    user likes will be ranked higher than an item the user dislikes)
  3. Besides, we want to get people to answer question that

    have no answers yet! Not enough information to estimate representations for each user and question. 14
  4. In particular, each question is described by a set of

    tags. We can use question metadata. 15
  5. Predict probability that a user will answer a question as

    a function of its tags. Use them to build logistic regression models for each user. 16
  6. Better, but: 1. There is no transfer of information between

    users 2. We are not capturing tag similarity 3. Item representations remain high-dimensional 0.66 ROC AUC 17
  7. Instead of finding embeddings for items, let’s find embeddings for

    item features Then add them together to represent items.
  8. 26 Multiple loss functions • Logistic loss for explicit binary

    feedback • BPR • WARP • k-th order statistic WARP loss
  9. Two learning rate schedules: • adagrad • adadelta Trained with

    asynchronous stochastic gradient descent. 27