Student Network Learning via Evolutionary Knowledge Distillation.


Kangkai ZhangChunhui ZhangShikun LiDan ZengShiming Ge


Paper [IEEE TCSVT]   




Abstract

Knowledge distillation provides an effective way to transfer knowledge via teacher-student learning, where most existing distillation approaches apply a fixed pre-trained model as teacher to supervise the learning of student network. This manner usually brings in a big capability gap between teacher and student networks during learning. Recent researches have observed thata small teacher-student capability gap can facilitate knowledge transfer. Inspired by that, we propose an evolutionary knowledge distillation approach to improve the transfer effectiveness of teacher knowledge. Instead of a fixed pre-trained teacher, an evolutionary teacher is learned online and consistently transfers intermediate knowledge to supervise student network learning on-the-fly. To enhance intermediate knowledge representation and mimicking, several simple guided modules are introduced between corresponding teacher-student blocks. In this way, the student can simultaneously obtain rich internal knowledge and capture its growth process, leading to effective student network learning. Extensive experiments clearly demonstrate the effectiveness of our approach as well as good adaptability in the low-resolution and few-sample scenarios.


BibTex

            
@article{zhang2021ekd,
    author={Kangkai Zhang, Chunhui Zhang, Shikun Li, Dan Zeng, and Shiming Ge},
    title={Student Network Learning via Evolutionary Knowledge Distillation}, 
    journal={IEEE Transactions on Circuits and Systems for Video Technology},
    year={2021},
    volume={},
    number={},
    pages={1-13}
}

        




Efficient Low-Resolution Face Recognition via Bridge Distillation


Shiming GeShengwei ZhaoChenyu LiYu ZhangJia Li


Paper [IEEE TIP]   




Abstract

Face recognition in the wild is now advancing towards light-weight models, fast inference speed and resolution-adapted capability. In this paper, we propose a bridge distillation approach to turn a complex face model pretrained on private high-resolution faces into a light-weight one for low-resolution face recognition. In our approach, such a cross-dataset resolution-adapted knowledge transfer problem is solved via two-step distillation. In the first step, we conduct cross-dataset distillation to transfer the prior knowledge from private high-resolution faces to public high-resolution faces and generate compact and discriminative features. In the second step, the resolution-adapted distillation is conducted to further transfer the prior knowledge to synthetic low-resolution faces via multi-task learning. By learning low-resolution face representations and mimicking the adapted high-resolution knowledge, a light-weight student model can be constructed with high efficiency and promising accuracy in recognizing low-resolution faces. Experimental results show that the student model performs impressively in recognizing low-resolution faces with only 0.21M parameters and 0.057MB memory. Meanwhile, its speed reaches up to 14,705, 934 and 763 faces per second on GPU, CPU and mobile phone, respectively.


BibTex

            
@article{ge2020bd,
    author={Shiming  Ge and Shengwei Zhao and Chenyu Li and Yu Zhang and Jia Li},
    title={Efficient Low-Resolution Face Recognition via Bridge Distillation}, 
    journal={IEEE Transactions on Image Processing},
    year={2020},
    volume={29},
    number={},
    pages={6898-6908}
}

        




Look One and More: Distilling Hybrid Order Relational Knowledge for Cross-Resolution Image Recognition


Shiming GeKangkai ZhangHaolin LiuYingying HuaShengwei ZhaoXin JinHao Wen


Paper [AAAI 2020]   




Abstract

In spite of great success in many image recognition tasks achieved by recent deep models, directly applying them to recognize low-resolution images may suffer from low accuracy due to the missing of informative details during resolution degradation. However, these images are still recognizable for subjects who are familiar with the corresponding high-resolution ones. Inspired by that, we propose a teacher-student learning approach to facilitate low-resolution image recognition via hybrid order relational knowledge distillation. The approach refers to three streams: the teacher stream is pretrained to recognize high-resolution images in high accuracy, the student stream is learned to identify low-resolution images by mimicking the teacher's behaviors, and the extra assistant stream is introduced as bridge to help knowledge transfer across the teacher to the student. To extract sufficient knowledge for reducing the loss in accuracy, the learning of student is supervised with multiple losses, which preserves the similarities in various order relational structures. In this way, the capability of recovering missing details of familiar low-resolution images can be effectively enhanced, leading to a better knowledge transfer. Extensive experiments on metric learning, low-resolution image classification and low-resolution face recognition tasks show the effectiveness of our approach, while taking reduced models.


BibTex

            
@inproceedings{AAAI20/Ge/HORKD,
    author    = {Shiming Ge and Kangkai Zhang and Haolin Liu and Yingying Hua and Shengwei Zhao and Xin Jin and Hao Wen},
    title     = {Look One and More: Distilling Hybrid Order Relational Knowledge for Cross-Resolution Image Recognition},
    booktitle = {The Thirty-Fourth {AAAI} Conference on Artificial Intelligence, {AAAI}
                    2020, The Thirty-Second Innovative Applications of Artificial Intelligence
                    Conference, {IAAI} 2020, The Tenth {AAAI} Symposium on Educational
                    Advances in Artificial Intelligence, {EAAI} 2020, New York, NY, USA,
                    February 7-12, 2020},
    pages     = {10845--10852},
    publisher = {{AAAI} Press},
    year      = {2020},
}

        




Low-Resolution Face Recognition in the Wild via Selective Knowledge Distillation
Shiming GeShengwei ZhaoChenyu LiJia Li


Paper [IEEE TIP]   




Abstract

Typically, the deployment of face recognition models in the wild needs to identify low-resolution faces with extremely low computational cost. To address this problem, a feasible solution is compressing a complex face model to achieve higher speed and lower memory at the cost of minimal performance drop. Inspired by that, this paper proposes a learning approach to recognize low-resolution faces via selective knowledge distillation. In this approach, a two-stream convolutional neural network (CNN) is first initialized to recognize high-resolution faces and resolution-degraded faces with a teacher stream and a student stream, respectively. The teacher stream is represented by a complex CNN for high-accuracy recognition, and the student stream is represented by a much simpler CNN for low-complexity recognition. To avoid significant performance drop at the student stream, we then selectively distil the most informative facial features from the teacher stream by solving a sparse graph optimization problem, which are then used to regularize the fine-tuning process of the student stream. In this way, the student stream is actually trained by simultaneously handling two tasks with limited computational resources: approximating the most informative facial cues via feature regression, and recovering the missing facial cues via low-resolution face classification. Experimental results show that the student stream performs impressively in recognizing low-resolution faces and costs only 0.15-MB memory and runs at 418 faces per second on CPU and 9433 faces per second on GPU.


BibTex

            
@article{TIP19/Ge/SKD,
    author    = {Shiming Ge and Shengwei Zhao and Chenyu Li and Jia Li},
    title     = {Low-Resolution Face Recognition in the Wild via Selective Knowledge Distillation},
    journal   = {{IEEE} Trans. Image Process.},
    volume    = {28},
    number    = {4},
    pages     = {2051--2062},
    year      = {2019},
}

        


Low-Resolution Face Recognition in the Wild with Mixed-Domain Distillation


Shengwei ZhaoXindi GaoShikun LiShiming Ge


Paper [BigMM 2019]   




Abstract

Low-resolution face recognition in the wild still is an open problem. In this paper, we propose to address this problem via a novel learning approach called Mixed-Domain Distillation (MDD). The approach applies a teacher-student framework to mix and distill knowledge from four different domain datasets, including private high-resolution, public high-resolution, public low-resolution web and target lowresolution wild face datasets. In this way, high-resolution knowledge from the well-trained complex teacher model is first adapted to public high-resolution faces and then transferred to a simply student model. The student model is designed to identify low-resolution faces, and could perform face recognition in the wild effectively and efficiently. Experimental results show that our proposed model outperforms several existing models for low-resolution face recognition in the wild.


BibTex

            
@inproceedings{BigMM19/Zhao/MDD,
    author    = {Shengwei Zhao and Xindi Gao and Shikun Li and Shiming Ge},
    title     = {Low-Resolution Face Recognition in the Wild with Mixed-Domain Distillation},
    booktitle = {Fifth {IEEE} International Conference on Multimedia Big Data, BigMM
                    2019, Singapore, September 11-13, 2019},
    pages     = {148--154},
    publisher = {{IEEE}},
    year      = {2019},
}        

        


Fewer-Shots and Lower-Resolutions: Towards Ultrafast Face Recognition in the Wild


Shiming GeShengwei ZhaoXindi GaoJia Li


Paper [ACM Multimedia]   




Abstract

Is it possible to train an effective face recognition model with fewer shots that works efficiently on low-resolution faces in the wild? To answer this question, this paper proposes a few-shot knowledge distillation approach to learn an ultrafast face recognizer via two steps. In the first step, we initialize a simple yet effective face recognition model on synthetic low-resolution faces by distilling knowledge from an existing complex model. By removing the redundancies in both face images and the model structure, the initial model can provide an ultrafast speed with impressive recognition accuracy. To further adapt this model into the wild scenarios with fewer faces per person, the second step refines the model via few-shot learning by incorporating a relation module that compares low-resolution query faces with faces in the support set. In this manner, the performance of the model can be further enhanced with only fewer low-resolution faces in the wild. Experimental results show that the proposed approach performs favorably against state-of-the-arts in recognizing low-resolution faces with an extremely low memory of 30KB and runs at an ultrafast speed of 1,460 faces per second on CPU or 21,598 faces per second on GPU.


BibTex

            
@inproceedings{MM19/Ge/FSLR,
    author    = {Shiming Ge and Shengwei Zhao and Xindi Gao and Jia Li},
    title     = {Fewer-Shots and Lower-Resolutions: Towards Ultrafast Face Recognition in the Wild},
    booktitle = {Proceedings of the 27th {ACM} International Conference on Multimedia,
                    {MM} 2019, Nice, France, October 21-25, 2019},
    pages     = {229--237},
    publisher = {{ACM}},
    year      = {2019},
}