About Me

Jian Liang

梁坚 (Jian Liang, Tim)
Associate Professor
Center for Research on Intelligent Perception and Computing
Institute of Automation, Chinese Academy of Sciences

Room 1505, Intelligent Building, 95 Zhongguancun East Road
100190, Haidian District, Beijing, China

🎖Google Scholar
  liangjian92🌀gmail.com or jian.liang🌀nlpr.ia.ac.cn

Before joining CASIA in June 2021, I was a research fellow at the Vision and Learning Group, National University of Singapore, working with Dr. Jiashi Feng from June 2019 to April 2021. I obtained Ph.D. in Pattern Recognition and Intelligent Systems from CASIA in Jan 2019, under the supervision of Prof. Tieniu Tan and co-supervision of Prof. Ran He and Prof. Zhenan Sun, and received my bachelor degree in Automation from Xi'an Jiaotong University in June 2013.

My current research interests mainly focus on representation learning, knowledge transfer, trustworthy AI (including security, privacy, or robustness in AI), and their applications in various computer vision problems.

I am looking for a master's student who will start in 2024.
I am open to discussion or collaboration. Feel free to drop me an email if you're interested.

Recent Publications (Full list)

DINE: Domain Adaptation from Single and Multiple Black-box Predictors

Abstract: To ease the burden of labeling, unsupervised domain adaptation (UDA) aims to transfer knowledge in previous and related labeled datasets (sources) to a new unlabeled dataset (target). Despite impressive progress, prior methods always need to access the raw source data and develop data-dependent alignment approaches to recognize the target samples in a transductive learning manner, which may raise privacy concerns from source individuals. Several recent studies resort to an alternative solution by exploiting the well-trained white-box model from the source domain, yet, it may still leak the raw data through generative adversarial learning. This paper studies a practical and interesting setting for UDA, where only black-box source models (i.e., only network predictions are available) are provided during adaptation in the target domain. To solve this problem, we propose a new two-step knowledge adaptation framework called DIstill and fine-tuNE (DINE). Taking into consideration the target data structure, DINE first distills the knowledge from the source predictor to a customized target model, then fine-tunes the distilled model to further fit the target domain. Besides, neural networks are not required to be identical across domains in DINE, even allowing effective adaptation on a low-resource device. Empirical results on three UDA scenarios (i.e., single-source, multi-source, and partial-set) confirm that DINE achieves highly competitive performance compared to state-of-the-art data-dependent approaches.

Jian Liang, Dapeng Hu, Jiashi Feng, Ran He.
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022 (Oral)
[Paper] [Code] [Slides]

This paper studies a practical and interesting setting for UDA, where only black-box source models (i.e., only network predictions are available) are provided during adaptation in the target domain.

Mimicking the Oracle: An Initial Phase Decorrelation Approach for Class Incremental Learning
Abstract: Class Incremental Learning (CIL) aims at learning a multi-class classifier in a phase-by-phase manner, in which only data of a subset of the classes are provided at each phase. Previous works mainly focus on mitigating forgetting in phases after the initial one. However, we find that improving CIL at its initial phase is also a promising direction. Specifically, we experimentally show that directly encouraging CIL Learner at the initial phase to output similar representations as the model jointly trained on all classes can greatly boost the CIL performance. Motivated by this, we study the difference between a naïvely-trained initial-phase model and the oracle model. Specifically, since one major difference between these two models is the number of training classes, we investigate how such difference affects the model representations. We find that, with fewer training classes, the data representations of each class lie in a long and narrow region; with more training classes, the representations of each class scatter more uniformly. Inspired by this observation, we propose Class-wise Decorrelation (CwD) that effectively regularizes representations of each class to scatter more uniformly, thus mimicking the model jointly trained with all classes (i.e., the oracle model). Our CwD is simple to implement and easy to plug into existing methods. Extensive experiments on various benchmark datasets show that CwD consistently and significantly improves the performance of existing state-of-the-art methods by around 1\% to 3\%.
Yujun Shi, Kuangqi Zhou, Jian Liang, Zihang Jiang, Jiashi Feng, Philip Torr, Song Bai, Vincent Y. F. Tan.
IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022
[Paper] [Code]

We propose a Class-wise Decorrelation regularizer that enables CIL learner at initial phase to mimic representations produced by the oracle model (the model jointly trained on all classes) and thus boosting Class Incremental Learning.

META: Mimicking Embedding via oThers' Aggregation for Generalizable Person Re-identification.
Abstract: Domain generalizable (DG) person re-identification (ReID) aims to test across unseen domains without access to the target domain data at training time, which is a realistic but challenging problem. In contrast to methods assuming an identical model for different domains, Mixture of Experts (MoE) exploits multiple domain-specific networks for leveraging complementary information between domains, obtaining impressive results. However, prior MoE-based DG ReID methods suffer from a large model size with the increase of the number of source domains, and most of them overlook the exploitation of domain-invariant characteristics. To handle the two issues above, this paper presents a new approach called Mimic Embedding via adapTive Aggregation ($\mathsf{META}$) for DG person ReID. To avoid the large model size, experts in $\mathsf{META}$ do not adopt a branch network for each source domain but share all the parameters except for the batch normalization layers. Besides multiple experts, $\mathsf{META}$ leverages Instance Normalization (IN) and introduces it into a global branch to pursue invariant features across domains. Meanwhile, $\mathsf{META}$ considers the relevance of an unseen target sample and source domains via normalization statistics and develops an aggregation module to adaptively integrate multiple experts for mimicking unseen target domain. Benefiting from a proposed consistency loss and an episodic training algorithm, $\mathsf{META}$ is expected to mimic embedding for a truly unseen target domain. Extensive experiments verify that $\mathsf{META}$ surpasses state-of-the-art DG person ReID methods by a large margin.
Boqiang Xu, Jian Liang, Lingxiao He, Zhenan Sun.
European Conference on Computer Vision (ECCV), 2022.
[Paper] [Code]

We propose a new domain generalization method for person re-identification.

Are You Stealing My Model? Sample Correlation for Fingerprinting Deep Neural Networks.
Abstract: An off-the-shelf model as a commercial service could be stolen by model stealing attacks, posing great threats to the rights of the model owner. Model fingerprinting aims to verify whether a suspect model is stolen from the victim model, which gains more and more attention nowadays. Previous methods always leverage the transferable adversarial examples as the model fingerprint, which is sensitive to adversarial defense or transfer learning scenarios. To address this issue, we consider the pairwise relationship between samples instead and propose a novel yet simple model stealing detection method based on SAmple Correlation (SAC). Specifically, we present SAC-w that selects wrongly classified normal samples as model inputs and calculates the mean correlation among their model outputs. To reduce the training time, we further develop SAC-m that selects CutMix Augmented samples as model inputs, without the need for training the surrogate models or generating adversarial examples. Extensive results validate that SAC successfully defends against various model stealing attacks, even including adversarial training or transfer learning, and detects the stolen models with the best performance in terms of AUC across different datasets and model architectures.
Jiyang Guan, Jian Liang, Ran He.
Annual Conference on Neural Information Processing Systems (NeurIPS), 2022
[Paper] [Code]

We propose a simple yet effective method for fingerprinting deep models.