Visible to the public Matching User Photos to Online Products with Robust Deep Features

TitleMatching User Photos to Online Products with Robust Deep Features
Publication TypeConference Paper
Year of Publication2016
AuthorsWang, Xi, Sun, Zhenfeng, Zhang, Wenqiang, Zhou, Yu, Jiang, Yu-Gang
Conference NameProceedings of the 2016 ACM on International Conference on Multimedia Retrieval
PublisherACM
Conference LocationNew York, NY, USA
ISBN Number978-1-4503-4359-6
KeywordsDeep Learning, image retrieval, pubcrawl, pubcrawl170201, science of security, visual similarity
Abstract

This paper focuses on a practically very important problem of matching a real-world product photo to exactly the same item(s) in online shopping sites. The task is extremely challenging because the user photos (i.e., the queries in this scenario) are often captured in uncontrolled environments, while the product images in online shops are mostly taken by professionals with clean backgrounds and perfect lighting conditions. To tackle the problem, we study deep network architectures and training schemes, with the goal of learning a robust deep feature representation that is able to bridge the domain gap between the user photos and the online product images. Our contributions are two-fold. First, we propose an alternative of the popular contrastive loss used in siamese deep networks, namely robust contrastive loss, where we "relax" the penalty on positive pairs to alleviate over-fitting. Second, a multi-task fine-tuning approach is introduced to learn a better feature representation, which not only incorporates knowledge from the provided training photo pairs, but also explores additional information from the large ImageNet dataset to regularize the fine-tuning procedure. Experiments on two challenging real-world datasets demonstrate that both the robust contrastive loss and the multi-task fine-tuning approach are effective, leading to very promising results with a time cost suitable for real-time retrieval.

URLhttps://dl.acm.org/doi/10.1145/2911996.2912002
DOI10.1145/2911996.2912002
Citation Keywang_matching_2016