Unified Visual Preference Learning for User Intent Understanding

被引:1
|
作者
Wen, Yihua [1 ,3 ]
Chen, Si [2 ]
Tian, Yu [1 ,3 ]
Guan, Wanxian [2 ]
Wang, Pengjie [2 ]
Deng, Hongbo [2 ]
Xu, Jian [2 ]
Zheng, Bo [2 ]
Li, Zihao [2 ]
Zou, Lixin [1 ]
Li, Chenliang [1 ]
机构
[1] Wuhan Univ, Minist Educ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Wuhan, Peoples R China
[2] Alibaba Grp China, Hangzhou, Peoples R China
[3] Alibaba Grp, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Visual Preference Modeling; Disentangled Representation Learning; Search and Recommendation; SEARCH;
D O I
10.1145/3616855.3635858
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the world of E-Commerce, the core task is to understand the personalized preference from various kinds of heterogeneous information, such as textual reviews, item images and historical behaviors. In current systems, these heterogeneous information are mainly exploited to generate better item or user representations. For example, in scenario of visual search, the importance of modeling query image has been widely acknowledged. But, these existing solutions focus on improving the representation quality of the query image, overlooking the personalized visual preference of the user. Note that the visual features affect the user's decision significantly, e.g., a user could be more likely to click the items with her preferred design. Hence, it is fruitful to exploit the visual preference to deliver better capacity for personalization. To this end, we propose a simple yet effective target-aware visual preference learning framework (named Tavern) for both item recommendation and search. The proposed Tavern works as an individual and generic model that can be smoothly plugged into different downstream systems. Specifically, for visual preference learning, we utilize the image of the target item to derive the visual preference signals for each historical clicked item. This procedure is modeled as a form of representation disentanglement, where the visual preference signals are extracted by taking off the noisy information irrelevant to visual preference from the shared visual information between the target and historical items. During this process, a novel selective orthogonality disentanglement is proposed to avoid the significant information loss. Then, a GRU network is utilized to aggregate these signals to form the final visual preference representation. Extensive experiments over three large-scale real-world datasets covering visual search, product search and recommendation well demonstrate the superiority of our proposed Tavern against existing technical alternatives. Further ablation study also confirms the validity of each design choice
引用
收藏
页码:816 / 825
页数:10
相关论文
共 50 条
  • [31] User and Item Preference Learning for Hybrid Recommendation Systems
    Prakash, Kaavya
    Asad, Fayiz
    Urolagin, Siddhaling
    SMART TECHNOLOGIES AND INNOVATION FOR A SUSTAINABLE FUTURE, 2019, : 447 - 455
  • [32] Various aspects of user preference learning and recommender systems
    Eckhardt, Alan
    DATESO 2009 - DATABASES, TEXTS, SPECIFICATIONS, OBJECTS: PROCEEDINGS OF THE 9TH ANNUAL INTERNATIONAL WORKSHOP, 2009, 471 : 56 - 67
  • [33] A weighted adaptation method on learning user preference profile
    Zhang, Zhiyuan
    Liu, Yun
    Xu, Guandong
    Chen, Haiqiang
    KNOWLEDGE-BASED SYSTEMS, 2016, 112 : 114 - 126
  • [34] Clustering of Conversational Bandits for User Preference Learning and Elicitation
    Wu, Junda
    Zhao, Canzhe
    Yu, Tong
    Li, Jingyang
    Li, Shuai
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2129 - 2139
  • [35] Learning Styles and Graphical User Interface: Is there any Preference?
    Dedic, Velimir
    Markovic, Suzana
    Kuleto, Valentin
    NEW EDUCATIONAL REVIEW, 2012, 28 (02): : 73 - 83
  • [36] Unified Transformer Multi-Task Learning for Intent Classification With Entity Recognition
    Benayas Alamos, Alberto Jose
    Hashempou, Reyhaneh
    Rumble, Damian
    Jameel, Shoaib
    De Amorim, Renato Cordeiro
    IEEE ACCESS, 2021, 9 : 147306 - 147314
  • [37] Understanding Temporal Intent of User Query Based on Time-Based Query Classification
    Ren, Pengjie
    Chen, Zhumin
    Song, Xiaomeng
    Li, Bin
    Yang, Haopeng
    Ma, Jun
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2013, 2013, 400 : 334 - 345
  • [38] A user preference modelling method for the assessment of visual complexity in building facade
    Hussein, Dheyaa
    SMART AND SUSTAINABLE BUILT ENVIRONMENT, 2020, 9 (04) : 483 - 501
  • [39] Deep learning for visual understanding: A review
    Guo, Yanming
    Liu, Yu
    Oerlemans, Ard
    Lao, Songyang
    Wu, Song
    Lew, Michael S.
    NEUROCOMPUTING, 2016, 187 : 27 - 48
  • [40] DeepStyle: Learning User Preferences for Visual Recommendation
    Liu, Qiang
    Wu, Shu
    Wang, Liang
    SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 841 - 844