Recommending for new users is surprisingly difficult

By on

For years MovieLens has required users to enter 15 ratings before they are allowed to get personalized recommendations. This design makes sense: how can we make recommendations for a user we know nothing about? That said, we don’t know if this provides users with the best experience. Why should users have to enter fifteen ratings, why not ten or five? What would happen if we let users into the system without any ratings? To answer these questions we need to understand how our algorithms behave for users with very few ratings.

To understand how algorithms behave for users just joining the system, we looked at historic MovieLens ratings. We trained three popular recommender algorithms: ItemItem, UserUser, and SVD on this rating data. While training, we limited some users to have only a small number of ratings. We used the ratings that were not given to the algorithm to measure several things:

  • How accurate are the predictions? Can the algorithm accurately predict the user’s future ratings?
  • How good are the recommendations? Does the algorithm suggest movies for the user that the user would like?
  • What type of recommendations does the algorithm generate? Is there a good diversity of movies? Are the movies popular, or more obscure?

(more…)

Evaluating an External Recommender written in Python using LensKit

By on

In this article, we show how to use LensKit to evaluate a recommender written in Python.  We wrote this article to help people who want to use LensKit’s built-in evaluation capabilities and comparison algorithms, but don’t want to implement their own algorithms in Java.  Evaluating an external recommender — whether in R, Python, or MatLab, involves three primary steps:

  • Writing the recommender. We will need a simple recommender written in language other than Java (Python in this case) that can take test data to build up a simple model and generate recommendations for a given list of test users.
  • Setting up a shim class. We will need to write a small class that teaches LensKit how to use our external algorithm.
  • Setting up LensKit evaluation. Finally we show how we setup an experiment using the shim class in a LensKit eval script to evaluate the external recommender.

Note, that the data we will use to test this recommender is a MovieLens rating dataset. The data consists of movie ratings with each row being <userId,itemId,rating>. You can read more about the dataset here. (more…)

College guidance platform PossibilityU forms research partnership with GroupLens

By on

White House, Department of Education, and Chronicle of Higher Education highlight our college recommendation software

PossibilityU logo

Two organizations have joined forces to answer important questions related to how search, discovery and recommendation online are affecting student college choices. PossibilityU, an innovative college guidance platform, and GroupLens, a research lab dedicated to recommender systems and online communities in the Department of Computer Science and Engineering at the University of Minnesota, Twin Cities, announced a research partnership to explore how college recommendations online can effect a broad range of student outcomes.

For the past 3 years, PossibilityU’s unique college guidance program has been matching high school students across the country with colleges that match their talents, aspirations and family budgets. The program features guided inquiry, data visualizations and personalized, Netflix-like recommendations—all designed to give students confidence in their choices. PossibilityU also features a blended-learning video curriculum created to help students and parents become better consumers of higher education.

(more…)

Taking Recommendations Improves Consumption Diversity — A Surprise Result from Exploring the Filter Bubble and MovieLens

By on

This post describes work being presented at WWW 2014, by Tien Nguyen

Those of you following recommender systems have almost certainly heard the debate about filter bubbles.  This concept, perhaps best articulated by Eli Pariser, argues that recommenders have the potential to trap users into increasingly similar content, isolating them from the diversity of content that makes people rich learners.

We decided to test this concept empirically, using longitudinal data from MovieLens.  Specifically, we wanted to answer two questions:

  • Do recommended movies get narrower as users continue to rate movies?

  • Do users consume narrower movies — and if so, is this a consequence of taking recommendations?

What we found surprised us.

(more…)