Yizeng Han


Ph.D Candidate, advised by Prof. Gao Huang and Prof. Shiji Song.
Department of Automation, Tsinghua University.

Education

  • Ph.D, Tsinghua University, 2018 - present.
  • B.E., Tsinghua University, 2014 - 2018.

  • Research Experience

  • Intern, Georgia Institute of Technology, 06/2017 - 08/2017

  • Research Interest

    My research focuses on machine learning and computer vision, in particular deep learning, efficient inference and dynamic neural networks.

    Recent Publications & Preprints (Google Scholar)


    Dynamic Neural Networks: A Survey. [智源社区][机器之心-在线讲座][Bilibili]
    Yizeng Han*, Gao Huang*, Shiji Song, Le Yang, Honghui Wang, Yulin Wang.
    Arxiv Preprint.
    Dynamic neural network is an emerging research topic in deep learning. Compared to static models which have fixed computational graphs and parameters at the inference stage, dynamic networks can adapt their structures or parameters to different inputs, leading to notable advantages in terms of accuracy, computational efficiency, adaptiveness, etc. In this survey, we comprehensively review this rapidly developing area. The important research problems of dynamic networks, e.g., architecture design, decision making scheme, optimization technique and applications, are reviewed systematically. Finally, we discuss the open problems in this field together with interesting future research directions.
    Resolution Adaptive Networks for Efficient Inference. [code]
    Le Yang*, Yizeng Han*, Xi Chen*, Shiji Song, Jifeng Dai, Gao Huang.
    IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2020.
    In this paper, we focus on spatial redundancy of input samples and propose a novel Resolution Adaptive Network (RANet), which is inspired by the intuition that low-resolution representations are sufficient for classifying “easy” inputs containing large objects with prototypical features, while only some “hard” samples need spatially detailed information. Empirically, we demonstrate the effectiveness of the pro- posed RANet on the CIFAR-10, CIFAR-100 and ImageNet datasets in both the anytime prediction setting and the budgeted batch classification setting.
    Adaptive Focus for Efficient Video Recognition. [code]
    Yulin Wang, Zhaoxi Chen, Haojun Jiang, Shiji Song, Yizeng Han, and Gao Huang.
    IEEE/CVF International Conference on Computer Vision (ICCV Oral) 2021.
    In this paper, we explore the spatial redundancy in video recognition with the aim to improve the computational efficiency. Extensive experiments on five benchmark datasets, i.e., ActivityNet, FCVID, Mini-Kinetics, Something-Something V1&V2, demonstrate that our method is significantly more efficient than the competitive baselines.
    Towards Learning Spatially Discriminative Feature Representations.
    Chaofei Wang*, Jiayu Xiao*, Yizeng Han, Qisen Yang, Shiji Song, Gao Huang.
    IEEE/CVF International Conference on Computer Vision (ICCV) 2021.
    We propose a novel loss function, termed as CAM-loss, to constrain the embedded feature maps with the class activation maps (CAMs) which indicate the spatially discriminative regions of an image for particular categories. Experimental results show that CAM-loss is applicable to a variety of network structures and can be combined with mainstream regularization methods to improve the performance of image classification. The strong generalization ability of CAM-loss is validated in the transfer learning and few shot learning tasks. Based on CAM-loss, we also propose a novel CAAM-CAM matching knowledge distillation method. This method directly uses the CAM generated by the teacher network to supervise the CAAM generated by the student network, which effectively improves the accuracy and convergence rate of the student network.
    * Equal Contribution.

    Awards

    • Comprehensive Merit Scholarship, 2017, 2016 at Tsinghua University.
    • Academic Excellence Scholarship, 2015 at Tsinghua University.

    Contact

    • hanyz18 at mails dot tsinghua dot edu dot cn.
    • 616 Centre Main Building, Tsinghua University, Beijing 100084, China.