Finger Vein Recognition Based on Multi-Receptive Field Bilinear Convolutional Neural Network

Kaixuan Wang, Guanghua Chen, Hongjia Chu

My Comments and Inspiration

Comments: The paper is in a bad organziation. The author post his idea step by step. He first proposed usage of dilation convolution to achieve “mulit scale receptive”, and then he thinks that the bilinear network is expensive on computation. Thus he then propose a light-weight CNN module by introducing depthwise separable convolution. Finally, a very complexity attention machenism is proposed to strengthen the relationship between space and channel.

提出的轻量模型,一点也不轻量。主要减少了Conv的参数,实际上很大的参数量都是来源于双线性池化后的全连接层,根据文章所示,光是最后一层FC的参数量就应该为 128*128*nClass

No loss function mentioned.

No inspiration.

Contributions and Important Conclusions

N/A

Motivation

  1. The author thinks the vein images are similar, and the second-order feature is a good choice to make feature more dicriminative.
  2. Computational cost is expensive
  3. Most

Methods

Step by step, the latter one aims to improve the former one.

  1. Backbone, Multi-Receptive Field Bilinear Convolutional Neural Network (MRFBCNN)

    |600

    where 蓝色的圆× means Bilinear Pooling Operation, the details can be found here

    In author’s description, the branch A and C, B and D are the same. Thus, in practical opreation, the author only used A and B branch to obtain two features and . The output of bilinear pooling are sent to FC to get final feature.

  1. Make MRFBCNN be lightweight.

    The author construct the lightweight module,consisting of depthwise sparable convolutional layer (DSConv1), like this

    In addition, he also replace the last convolutional layer with kernel.

    The architecture of L-MRFBCNN is To understand above figure, you should read it with the figure in step 1.

  2. Attention Mechanism: Dimensional Interactive Attention Mechanism (DIAM)

    In order to strengthen the relationship between space and channel, the author propose DIAM.

    Easy to follow the picture below.

    The output is

    means element-wise multiplication means matrix multiplication means matrix add.

Experiments

Database

SDU and USM, 1:9 = test:training

So ridiculous and make no sense....

Details

  • GTX 1660

This work maybe run on his own laptop.

  • LR: 0.0001
  • epoch: 200
  • bs: 24 (It should be limited by its memory.)
  • Optimizer: Adam, no more information about Adam configuration.

Some Descriptions

Footnotes

  1. F. Mamalet and C. Garcia. “Simplifying convnets for fast learning,” in Proc. Int. Conf. Artif. Neural Netw., 2020, pp. 58–65.