This is a quick implementation for Deep Hypersphere Embedding for Face Recognition(CVPR 2017).This paper proposed the angular softmax loss that enables convolutional neural networks(CNNs) to learn angularly discriminative features. The main content I replicated contains:
1. mathematical comparison among original softmax, modified softmax and angular softmax;
2. show the accuracy and loss comparison of different softmax in the experiment;
3. 2D and 3D visualization of embeddings learned with different softmax loss on MNIST dataset;
many current CNNS can viewed as convolution feature learning guided by softmax loss on top. however, softmax is easy to to optimize but does not explicitly encourage large margin between different classes.
on this situation, the author proposed a new loss function that always encourages an angular decision margin between different classes.
softmax loss
softmax
formula
test acc(MNIST)
original softmax
0.9775
modified softmax
0.9847
angular softmax
0.9896
A toy example on MNIST dataset, CNN features can be visualized by setting the output dimension as 2 or 3, as shown in following figures.