The multimodal product summarization task aims at generating a condensed textual summary for a given product. The input contains a detailed product description, a product knowledge base, and a product image.
The dataset for this task is collected from JD.COM, a mainstream Chinese e-commerce platform. Each sample is a (product textual description, product knowledge base, product image, product summary) tuple. The dataset contains three product categories, including Cases&Bags, Home Appliances, and Clothing.
Cases&Bags | Home Appliances | Clothing | |
---|---|---|---|
Train | 50,000 | 100,000 | 200,000 |
Valid | 10,000 | 10,000 | 10,000 |
Test | 10,000 | 10,000 | 10,000 |
Input token/sample | 319.0 | 336.6 | 294.8 |
Product attribute/sample | 14.8 | 7.8 | 7.3 |
Product image/sample | 1 | 1 | 1 |
K-PLUG: Knowledge-injected Pre-trained Language Model for Natural Language Understanding and Generation in E-Commerce. (Findings of EMNLP 2021).
Note: submission of the baselines are NOT acceptable.
The results need to be submitted to the leaderboard of the challenge. The evaluation consists of two stages, an automatic evaluation, and a human evaluation. For automatic evaluation, we adopt the metric of ROUGE. We select the top-5 teams regarding the average score of ROUGE-1, ROUGE-2, and ROUGE-L to advance to the second round of evaluation, i.e., the human evaluation. For the human evaluation, we evaluate faithfulness, readability, non-redundancy, and importance, for 100 random sampled summaries each category. The final ranking is determined by the average score of the human evaluation.
Metrics | Scores |
Faithfulness | 0=unfaithful to the input 2=faithful to the input |
Readability | 0=hard to understand 1=partially hard to understand 2=easy to understand |
Non-redundancy | 0=full of redundant information 1=partially redundant information 2=no redundant information |
Importance | 0=no useful information 1=partially useful information 2=totally useful information |
If you are interested in our challenge, please fill out the application form and email lihaoran24 at jd.com (Please email us with your organization's email and note that you participate in the challenge). The dataset will be sent to you.
The top 3 participating teams will be certificated by NLPCC and CCF-NLP, as well as awarded cash rewards.
The first prize (*1): ¥3000
The second prize (*1): ¥2000
The third prize (*1): ¥1000
Announcement of shared tasks and call for participation: 2022/3/15
Registration open: 2022/3/15
Release of detailed task guidelines & training data: 2022/3/15
First submission of results on the blind test data: 2022/5/1
Registration deadline: 2022/5/5
Participants’ results submission deadline: 2022/6/3
Evaluation results release and call for system reports and conference paper: 2022/6/10
Conference paper submission deadline (only for shared tasks): 2022/6/20
Conference paper accept/reject notification: 2022/7/4
Camera-ready paper submission deadline: 2022/7/18
[1] Li et. al., Aspect-Aware Multimodal Summarization for Chinese E-Commerce Products. AAAI 2020.
[2] Yuan et. al., On the Faithfulness for E-commerce Product Summarization. COLING 2020.
[3] Xu et. al., Self-Attention Guided Copy Mechanism for Abstractive Summarization. ACL 2020.
[4] Xu et. al., K-PLUG: Knowledge-injected Pre-trained Language Model for Natural Language Understanding and Generation in E-Commerce. Findings of ACL: EMNLP 2021.