/Hybrid-Style-Siamese-Network

Hybrid Style Siamese Network: Incorporating style loss in complementary apparels retrieval (CVFAD - CVPR2020)

Primary LanguagePythonMIT LicenseMIT

Hybrid Style Siamese Network

This is the code for the paper:

Hybrid Style Siamese Network: Incorporating style loss in complementary apparels retrieval
Mayukh Bhattacharyya, Sayan Nag
Computer Vision for Fashion, Art and Design, CVPR 2020

[Paper][Video Presentation]

Hybrid Style Siamese Network incorporates style loss into triplet loss, in order to aid in complementary images retrieval. In the paper, it had been used for the application of complementary apparels retrieval.

The performance in regards to the retrieval of complementary items in terms of MAP scores.

Model Seed 1 Seed 2 Seed 3 Mean
Siamese Network 0.1226 0.1323 0.1263 0.1271
Hybrid Style Siamese Network 0.1251 0.1343 0.1329 0.1308

Dependencies:

  1. pytorch (both torch and torchvision)
  2. numpy
  3. opencv
  4. scikit-learn

In order to reproduce the experiments in the paper:

python hssn.py -s seed_value -l learning rate -e epochs -p patience -b batch_size -f cv_fold_id --hybrid

--hybrid is the flag for running the hybrid style siamese network. Omitting it will run the same experiment with the normal siamese network. Explanation of other parameters are there in hssn.py