Pytorch code for Unsupervised Embedding Learning via Invariant and Spreading Instance Feature in CVPR 2019. (arXiv).
The goal of this work is to learn a feature extraction DNN, such that features of the same instance under different data augmentations should be invariant, while features of different image instances should be separated.
We propose to optimize the embedding directly on the real-time instance features with softmax function.
The proposed feature achieves much faster learning speed and better accuracy.
The learned embedding performs well on both seen and unseen testing categories.
Our code extends the pytorch implementation of NCE paper in Github. Please refer to the official repo for details of data preparation and hardware configurations.
-
Prerequisites: Python2.7, pytorch=0.4, Numpy, TensorboardX, Scikit_learn
-
Clone this repo:
git clone https://github.com/mangye16/Unsupervised_Embedding_Learning
The training and testing set share the same categories. We use kNN classifier to evaluate the performance follow the settings in NCE paper.
python demo_seen.py --dataset cifar --batch-t 0.1 --lr 0.03 --low-dim 128 --batch-size 128 --gpu 0,1,2,3
-
--dataset
: "cifar": cifar-10 dataset, "stl": stl-10 dataset. -
--batch-t
: softmax temperature parameter (0.05-0.1). -
--lr
: learning rate (initialization: 0.03). -
--low-dim
: the feature embedding dimension (default: 128). -
--batch-size
: the training batch size (Generally, larger batch size results in better performance) -
--gpu
: gpu used for training
The training and testing set DO NOT share any common categories. We use recall at K accuracy to evaluate the performance following existing deep metric learning papers.
Prepare the data. Download the datasets first. Then run codes in ./pre_process/
to download and pre-process each dataset
You may use
python pre_process_bird.py
Alternatively, you can use the code to directly download and pre-process the datasets.
Remember to change the dataset path to your own path.
python demo_unseen.py --dataset cub200 --arch inception_v1_ml --lr 0.001 --low-dim 128 --batch-size 64 --gpu 0
-
--dataset
: "cub200": CUB200-2011 dataset or, "car196": Cars196 dataset, "ebay": Stanford Online Product dataset. -
--arch
: backbone network structure (defualt: inception_v1_ml). -
--lr
: learning rate (initialization: 0.01). -
--gpu
: gpu used for training (only support single GPU training for inception_v1 backbone)
Please kindly cite this paper in your publications if it helps your research:
@inproceedings{cvpr19unsupervised,
title={Unsupervised Embedding Learning via Invariant and Spreading Instance Feature},
author={Ye, Mang and Zhang, Xu and Yuen, Pong C. and Chang, Shih-Fu},
booktitle={IEEE International Conference on Computer Vision and Pattern Recognition (CVPR)},
year={2019},
}
Contact: mangye16@gmail.com