Skip to content

Latest commit

 

History

History
154 lines (111 loc) · 10.2 KB

README.md

File metadata and controls

154 lines (111 loc) · 10.2 KB

Atlas: A Dataset and Benchmark for E-commerce Clothing Product Categorization

Here you can get the dataset, source code pre-trained models of our paper runnable on GPU and CPU. You can clone and run this project in your CPU/GPU. We have dockerized our environment so that you can easily reproduce the results reported in the paper. For a broader overview, read more in our paper and check out our presentation at "The 2020 SIGIR Workshop On eCommerce" here.

What does this project do?

  • Performs automatic taxonomy prediction of Clothing images
  • Provides a dataset of 183,996 clothing images from 52 categories along with image description and pre-defined taxonomy

Download dataset Download Model

⭐ Star us on GitHub — it helps!

Table of contents

Getting started

Setting up the project

pip install -r models/product_categorization/requirements.txt

Predicting using Product Categorization model

Follow these steps to predict the category path for a product using pur pretrained product categoristion model:

  1. Download the pre-trained model and the word map file by filling up this form.
  2. Switch to models/product_categorization directory
  3. From the command line, point to the image, model, word map (and optionally, the beam size) as follows:
python caption_cbs.py --img='../../dataset/atlas_test/203_large.jpeg' --model='path/to/BEST_checkpoint_atlas_1_cap_per_img_1_min_word_freq.pth.tar' --word_map='path/to/WORDMAP_atlas_1_cap_per_img_1_min_word_freq.json' --karpathy_json='path/to/atlas_dataset.json' --beam_size=5

Our pre-trained model predicts the category path and displays an output image that shows which part of the image has been focussed by our model to predict the category level.

Similarly you can predict for other images using this command, by changing the path in --img parameter to point to your image location.

Few more sample predictions can be found in this section

Docker Instructions

Pull Docker image

To pull and build the image, use:

docker pull vumaasha2/altas-docker

Run Docker Container and mount local directory

Place all the required data (atlas_dataset.json, wordmap file and pre trained model) in a local directory. To start a container and mount local dorectory in it, replace </path/to/data> with the local directory path in the command shown below. The following line starts a container, mounts the folder in a folder named data in the container and names the container atlas, so you can easily refer to this later

docker run -it -v </path/to/data>:/data --name atlas -d vumaasha2/altas-docker

Access container terminal and execute commands

To open the terminal inside the running container and execute the commands:

docker exec -ti atlas bash

In the terminal, run this command to generate predictions for train, test and validation splits in atlas_dataset.json and get the classification metrics

python /Atlas/models/product_categorization/generate_metrics.py

To predict the category for an sample image and see the attention image:

python /Atlas/models/product_categorization/caption_cbs.py --img='/Atlas/dataset/atlas_test/203_large.jpeg' --model='/data/BEST_checkpoint_atlas_1_cap_per_img_1_min_word_freq.pth.tar' --word_map='/data/WORDMAP_atlas_1_cap_per_img_1_min_word_freq.json' --karpathy_json='/data/atlas_dataset.json' --beam_size=5

This will predict the category path and store the attention image in your local directory which you mounted into this container. You can see the image in the local directory path. To predict for other images, change the path in --img parameter to point to your image location

Concepts Overview

We use attention based neural network Encoder-Decoder model to generate the sequences in the taxonomy.

  • Encoder - The Encoder is a 101 layered Residual Network(ResNet) trained on the ImageNet classification that converts the input image into a fixed size vector.

  • Decoder - The Decoder is a combination of Long Short-Term Memory(LSTM) along with Attention Network. This is the part of the model that predicts sequences for the taxonomy. It combines the output from the encoder and attention weights to predict category paths as sequences for the taxonomy.

  • Attention - Attention Network learns which part of the image has to be focused to predict the next level in the category path while performing the sequence classification task.

  • Constrained Beam Search - Constrained Beam Search restricts the model from generating category paths that are not predefined in our taxonomy. It limits the sequences chosen by the Decoder in order to generate category paths within the taxonomy.

Implementation

Taxonomy Generation

We gathered taxonomy for the clothing top level category from popular Indian e-commerce fashion sites. We analyzed popular products, niche, and premium clothing products across these stores and developed our taxonomy with 52 category paths. The list of 52 category paths and additional details can be found here

Data Collection

For all categories in taxonomy tree, we collected product data and its images from popular Indian E-commerce stores. Web scraping tools like Scrapy and Selenium were used to extract the product title, breadcrumb, image and price of each product. Check out this section to know more about our data collection strategy for Atlas dataset.

The dataset, Atlas, we used for training our model is a high-quality product taxonomy dataset focusing on clothing products. It contains 183,996 images under 52 clothing categories.

We provide a JSON file atlas_dataset.json which has data and URL of the images for 183,996 products. A sample record from the JSON is shown below

{'filename': 'euro-fashion-men-s-cotton-brief-pack-of-3-c9f86351-product.jpeg', 
'title': "euro fashion\n men's cotton brief (pack of  3 )", 
'sentences': [{'tokens': ['Men', 'Inner Wear', 'Underwear']}], 
'image_url': 'https://images.voonik.com/01993582/euro-fashion-men-s-cotton-brief-pack-of-3-c9f86351-product.jpg?1522053196', 'split': 'train'}

Run dataset/create_dataset.py which crawls the images from all the image_url in JSON and creates our Atlas dataset

python dataset/create_dataset.py -m atlas 

Data Cleaning

After collecting data, we found that many product listing also included a zoomed in images that display intrinsic details such as the texture of the fabric, brand labels, button, and pocket styles. These zoomed in images would drastically affect the quality of the dataset. We automated the process of filtering out the noisy images with the help of a simple 3 layer CNN based classification model

More details about the architecture of CNN Model and how we used it to clean our dataset can be found here.

Note: Our Atlas dataset generated in the above section is already cleaned. No need to apply this Zoomed Vs Normal model on the dataset.

Model Training and Prediction

We approach the product categorization problem as a sequence prediction problem by leveraging the dependency between each level in the category path. We use attention based neural network Encoder-Decoder architecture to generate sequences.

Encoder is a 101 layered Residual Network(ResNet) trained on the ImageNet classification task which converts the input image to a fixed size vector. Decoder is a combination of Long Short-Term Memory(LSTM) along with attention network which combines the encoder output and attention weights to predict category paths as sequences.

We also extend our model by introducing constrained beam search on top of it to restrict the model from generating category paths that are not predefined in our taxonomy.

More details and procedures on model training can be found here

FAQs

Authors

See also the list of contributors who participated in this project.

Acknowledgements

We thank Krishna Sangeeth, Sriram Ramachandrasekaran, Anirudh Venkataraman, Manoj Mahalingam, Rajesh Muppalla, and Sridhar Venkatesh for their help and support.

Reference

If you use this code as part of any published research, please acknowledge the following paper (it encourages researchers who publish their code!)

License

This project is licensed under the MIT License - see the LICENSE.md file for details