Simultaneous Preference and Metric Learning from Paired Comparisons
Austin Xu, Mark Davenport
Spotlight presentation: Orals & Spotlights Track 05: Clustering/Ranking
on 2020-12-08T07:10:00-08:00 - 2020-12-08T07:20:00-08:00
on 2020-12-08T07:10:00-08:00 - 2020-12-08T07:20:00-08:00
Toggle Abstract Paper (in Proceedings / .pdf)
Abstract: A popular model of preference in the context of recommendation systems is the so-called ideal point model. In this model, a user is represented as a vector u together with a collection of items x_1 ... x_N in a common low-dimensional space. The vector u represents the user's "ideal point," or the ideal combination of features that represents a hypothesized most preferred item. The underlying assumption in this model is that a smaller distance between u and an item x_j indicates a stronger preference for x_j. In the vast majority of the existing work on learning ideal point models, the underlying distance has been assumed to be Euclidean. However, this eliminates any possibility of interactions between features and a user's underlying preferences. In this paper, we consider the problem of learning an ideal point representation of a user's preferences when the distance metric is an unknown Mahalanobis metric. Specifically, we present a novel approach to estimate the user's ideal point u and the Mahalanobis metric from paired comparisons of the form "item x_i is preferred to item x_j.'' This can be viewed as a special case of a more general metric learning problem where the location of some points are unknown a priori. We conduct extensive experiments on synthetic and real-world datasets to exhibit the effectiveness of our algorithm.