Multi-view visual Bayesian personalized ranking for restaurant recommendation
Abstract
In recent recommendation systems, the image information of items is often used in conjunction with deep convolution network to directly learn the visual features of items. However, the existing approaches usually use only one image to represent an item. These approaches are inadequate for an item with multi-view related images. For a restaurant, it has visual information of food, drink, environment, and so on. Each view of an item can be represented by multiple images. In this paper, we propose a new factorization model that combines multi-view visual information with the implicit feedback data for restaurant prediction and ranking. The visual features (visual information) of images are extracted by using a deep convolution network and are integrated into a collaborative filtering framework. In order to conduct personalized recommendation better, the multi-view visual features are fused through user related weights. User related weights reflect the personalized visual preference for restaurants and the weights are different and independent between users. We applied this model to make personalized recommendations for users on two real-world restaurant review datasets. Experimental results show that our model with multi-view visual information achieves better performance than models without or with only single-view visual information.
This work was supported in part by the National Natural Science Foundation of China (Grant No. 61602313, 61702084).
BibTeX
@article{Zhang-2020-126830,author = {Xiaoyan Zhang and Haihua Luo and Bowei Chen and Guibing Guo},
title = {Multi-view visual Bayesian personalized ranking for restaurant recommendation},
journal = {Applied Intelligence},
year = {2020},
month = {September},
volume = {50},
number = {9},
pages = {2901 - 2915},
}