Realistic Instance-level Product Retrieval (Product1M) Track on LID CVPR 2021 Challenge

Organized by zhanxlin - Current server time: March 10, 2025, 4:01 p.m. UTC

First phase

Final
Feb. 18, 2021, midnight UTC

End

Competition Ends
June 7, 2021, 11 p.m. UTC

Overview

With the the diversification of online customer's demand in E-commerce, Online merchandise has increasingly diversified categories and a large proportion of them are exhibited as a product portfolio where multiple instances of different products exist in one image. Previous retrieval methods focus on the relatively simple case, i.e., image-level retrieval for single-product images and the instance-level nature of retrieval is unexplored. To bridge this gap and advance the related research, we introduce a new task, instance-level product retrieval. Specifically, Given an image containing multiple product instances and a user-provided description, this task aims to retrieve the correct single product image in the gallery.

Workshop link: https://l2id.github.io/

Data Description

To facilitate the study of product retrieval research, we collect real-world product photos in e-commerce website to establish the first standard and comprehensive benchmark Product1M for instance-level product retrieval. Product1M is split into the train, val, test, and gallery set. The train set contains 1132830 samples including both the single-product and multi-product samples. There are only multi-product samples in the val and test set, which contain 2673 and 6547 samples respectively. The gallery set has 40033 samples for 458 categories. The samples in the gallery, val and test set are annotated with class labels for the purpose of evaluation,i,e., they are not involved in the training process, and the samples in the train set are not annotated.

You can download the dataset at Product1M(Google Drive) or Product1M(Baidu Drive -- sie3).

Dataset Examples

 

Organizers

Xiaodan Liang (Sun Yat-sen University)

Yunchao Wei (University of Technology Sydney)

Xunlin Zhan (Sun Yat-sen University)

Xiao Dong (Sun Yat-sen University)

YangXin Wu (Sun Yat-sen University)

Gengwei Zhang (University of Technology Sydney)

Minlong Lu (Alibaba Group)

Yichi Zhang (Alibaba Group)

Evaluation Metrics

For this task, we adopt three metrics for the Product1M evaluation, i.e., Precision(Prec@N), mean Average Precision(mAP@N) and mean Average Recall (mAR@N).

  • Precision(%) for information retrieval, used in the Image Retrieval
  • Mean Average Precision(%) for information retrieval, used in the Google Landmark
  • Mean Average Recall(%) takes the category distribution into account, i.e., the inclusion of instance ratio is informative for evaluating both the correctness and diversity of a retrieval algorithm and guarantees that some trivial results are not overestimated. mAR is computed as follow

More detail about mAR can be seen in Google doc.

Submission Format

The results should be written into a .txt file named retrieval_results.txt and then archived into a ZIP file (retrieval_results.zip). The example text file is available in retrieval_results.txt.

Specifically, each line in the text file contains the query id followed by a ranked retrieval id list, which is limited to 100 ids since we only evaluate at most `N=100` for all three metric.
Examples in the text file: `id_q,id_0,id_1,id_2,...,id_100`

After uploading your results, please wait for 40 to 50 minutes and refrash your page to see the scores.

Terms and Conditions

General Rules

  • Each entry is required to be associated to a team and its affiliation.
  • Using multiple accounts to increase the number of submissions is strictly prohibited.
  • Results should follow the correct format and must be uploaded to the evaluation server through the CodaLab competition site. Detailed information about how results will be evaluated is represented on the evaluation page.
  • The best entry of each team will be public in the leaderboard at all time.
  • The organizer reserves the absolute right to disqualify entries which is incomplete or illegible, late entries or entries that violate the rules.

Datasets

The datasets are released for academic research only and it is free to researchers from educational or research institutions for non-commercial purposes. When downloading the dataset you agree not to reproduce, duplicate, copy, sell, trade, resell or exploit for any commercial purposes, any portion of the images and any portion of derived data.

 

LICENSE

 

PRODUCT-1M DATASET LICENSE

COMMITMENT LETTER

Alibaba Group is the only owner of all intellectual property rights (including copyright) of the PRODUCT-1M DATASET (“Dataset” or “Data”). Alibaba Group reserves the right to terminate Licensee’s access to the Dataset at any time. This PRODUCT-1M DATASET LICENSE COMMITMENT LETTER (“Letter”) explains the rules before and after I/we (“Licensee”) download and use the PRODUCT-1M DATASET. By downloading or using the Dataset, as a Licensee I/we understand, acknowledge, and hereby agree to all the terms in this Letter.

I. APPLICATION REQUIREMENTS

The permission for application is only open to researchers or faculties of universities or research institutes who have successfully signed up for the Weakly-supervised Product Retrieval Competition (“Competition”). Alibaba Group reserves the right to distribute the license.

II. NO COMMERCIAL USE

Licenses free of charge are limited to non-commercial research use only. Licensee is only granted a limited, non-exclusive, non-assignable, and non-transferable license to use the Dataset, which cannot apply for any commercial use.

III. NO DISTRIBUTION

This license is not a sale of any or all of the owner’s rights. Licensee is not allowed to sublicense or distribute the Dataset in whole or in part to any third party. Licensee guarantees that the Dataset may only be used by himself, and Licensee cannot rent, lease, lend, sub-license, or transfer the Dataset or any rights under this Letter to anyone or any third party else.

IV. RESTRICTED USE IN RESEARCH

Licensee guarantees that the Dataset can only be used for essential purposes related to the Competition during the competition, such as analysis, experimentation, and display of competition results. Besides, it is prohibited for any business entity to obtain or use the Dataset.

V.LEGAL LIABILITY

Licensee shall indemnify, defend and hold harmless Alibaba Group, their directors, employees and representatives, from and against any and all claims arising out of Licensee’s use of Dataset. And Licensee is fully aware that Licensee shall make compensation for all the “Losses” suffered by Alibaba Group arising out of Licensee’s breach of any term of this Letter. The term “Losses” shall include, but not limited to: Any refunds, liquidated damages or compensation that Alibaba Group paid to third parties; Penalties; Legal fees and other fees and expenses incurred by Alibaba Group to eliminate, mitigate and/or otherwise manage the impact on Alibaba Group arising out of or related to Licensee’s act or failure to act.

 

 

 

Concate Us

For more information, please concate us at zhanxlin@mail2.sysu.edu.cn or dx.icandoit@gmail.com.

Final

Start: Feb. 18, 2021, midnight

Description: We evaluate mAR,mAP,Precision @10,@50,@100. Since system bug cannot be fixed, the order shown below is 100, 50, 10, respective.ly. The first column(mAR@100) is the most important indicator.

Competition Ends

June 7, 2021, 11 p.m.

You must be logged in to participate in competitions.

Sign In