+ All Categories
Home > Documents > M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features....

M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features....

Date post: 24-Aug-2020
Category:
Upload: others
View: 4 times
Download: 0 times
Share this document with a friend
15
Published as a conference paper at ICLR 2020 MUTUAL MEAN -T EACHING : P SEUDO L ABEL R EFINERY FOR U NSUPERVISED D O - MAIN A DAPTATION ON P ERSON R E - IDENTIFICATION Yixiao Ge, Dapeng Chen & Hongsheng Li The Chinese University of Hong Kong {yxge@link,hsli@ee}.cuhk.edu.hk ABSTRACT Person re-identification (re-ID) aims at identifying the same persons’ images across different cameras. However, domain diversities between different datasets pose an evident challenge for adapting the re-ID model trained on one dataset to another one. State-of-the-art unsupervised domain adaptation methods for per- son re-ID transferred the learned knowledge from the source domain by opti- mizing with pseudo labels created by clustering algorithms on the target domain. Although they achieved state-of-the-art performances, the inevitable label noise caused by the clustering procedure was ignored. Such noisy pseudo labels sub- stantially hinders the model’s capability on further improving feature representa- tions on the target domain. In order to mitigate the effects of noisy pseudo labels, we propose to softly refine the pseudo labels in the target domain by proposing an unsupervised framework, Mutual Mean-Teaching (MMT), to learn better fea- tures from the target domain via off-line refined hard pseudo labels and on-line refined soft pseudo labels in an alternative training manner. In addition, the com- mon practice is to adopt both the classification loss and the triplet loss jointly for achieving optimal performances in person re-ID models. However, conven- tional triplet loss cannot work with softly refined labels. To solve this problem, a novel soft softmax-triplet loss is proposed to support learning with soft pseudo triplet labels for achieving the optimal domain adaptation performance. The pro- posed MMT framework achieves considerable improvements of 14.4%, 18.2%, 13.4% and 16.4% mAP on Market-to-Duke, Duke-to-Market, Market-to-MSMT and Duke-to-MSMT unsupervised domain adaptation tasks. 1 1 I NTRODUCTION Person re-identification (re-ID) aims at retrieving the same persons’ images from images captured by different cameras. In recent years, person re-ID datasets with increasing numbers of images were proposed to facilitate the research along this direction. All the datasets require time-consuming an- notations and are keys for re-ID performance improvements. However, even with such large-scale datasets, for person images from a new camera system, the person re-ID models trained on exist- ing datasets generally show evident performance drops because of the domain gaps. Unsupervised Domain Adaptation (UDA) is therefore proposed to adapt the model trained on the source image do- main (dataset) with identity labels to the target image domain (dataset) with no identity annotations. State-of-the-art UDA methods (Song et al., 2018; Zhang et al., 2019b; Yang et al., 2019) for person re-ID group unannotated images with clustering algorithms and train the network with clustering- generated pseudo labels. Although the pseudo label generation and feature learning with pseudo labels are conducted alternatively to refine the pseudo labels to some extent, the training of the neural network is still substantially hindered by the inevitable label noise. The noise derives from the limited transferability of source-domain features, the unknown number of target-domain identities, and the imperfect results of the clustering algorithm. The refinery of noisy pseudo labels has crucial influences to the final performance, but is mostly ignored by the clustering-based UDA methods. 1 Code is available at https://github.com/yxgeee/MMT. 1 arXiv:2001.01526v2 [cs.CV] 30 Jan 2020
Transcript
Page 1: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

MUTUAL MEAN-TEACHING:PSEUDO LABEL REFINERY FOR UNSUPERVISED DO-MAIN ADAPTATION ON PERSON RE-IDENTIFICATION

Yixiao Ge, Dapeng Chen & Hongsheng LiThe Chinese University of Hong Kong{yxge@link,hsli@ee}.cuhk.edu.hk

ABSTRACT

Person re-identification (re-ID) aims at identifying the same persons’ imagesacross different cameras. However, domain diversities between different datasetspose an evident challenge for adapting the re-ID model trained on one dataset toanother one. State-of-the-art unsupervised domain adaptation methods for per-son re-ID transferred the learned knowledge from the source domain by opti-mizing with pseudo labels created by clustering algorithms on the target domain.Although they achieved state-of-the-art performances, the inevitable label noisecaused by the clustering procedure was ignored. Such noisy pseudo labels sub-stantially hinders the model’s capability on further improving feature representa-tions on the target domain. In order to mitigate the effects of noisy pseudo labels,we propose to softly refine the pseudo labels in the target domain by proposingan unsupervised framework, Mutual Mean-Teaching (MMT), to learn better fea-tures from the target domain via off-line refined hard pseudo labels and on-linerefined soft pseudo labels in an alternative training manner. In addition, the com-mon practice is to adopt both the classification loss and the triplet loss jointlyfor achieving optimal performances in person re-ID models. However, conven-tional triplet loss cannot work with softly refined labels. To solve this problem,a novel soft softmax-triplet loss is proposed to support learning with soft pseudotriplet labels for achieving the optimal domain adaptation performance. The pro-posed MMT framework achieves considerable improvements of 14.4%, 18.2%,13.4% and 16.4% mAP on Market-to-Duke, Duke-to-Market, Market-to-MSMTand Duke-to-MSMT unsupervised domain adaptation tasks. 1

1 INTRODUCTION

Person re-identification (re-ID) aims at retrieving the same persons’ images from images capturedby different cameras. In recent years, person re-ID datasets with increasing numbers of images wereproposed to facilitate the research along this direction. All the datasets require time-consuming an-notations and are keys for re-ID performance improvements. However, even with such large-scaledatasets, for person images from a new camera system, the person re-ID models trained on exist-ing datasets generally show evident performance drops because of the domain gaps. UnsupervisedDomain Adaptation (UDA) is therefore proposed to adapt the model trained on the source image do-main (dataset) with identity labels to the target image domain (dataset) with no identity annotations.

State-of-the-art UDA methods (Song et al., 2018; Zhang et al., 2019b; Yang et al., 2019) for personre-ID group unannotated images with clustering algorithms and train the network with clustering-generated pseudo labels. Although the pseudo label generation and feature learning with pseudolabels are conducted alternatively to refine the pseudo labels to some extent, the training of theneural network is still substantially hindered by the inevitable label noise. The noise derives from thelimited transferability of source-domain features, the unknown number of target-domain identities,and the imperfect results of the clustering algorithm. The refinery of noisy pseudo labels has crucialinfluences to the final performance, but is mostly ignored by the clustering-based UDA methods.

1Code is available at https://github.com/yxgeee/MMT.

1

arX

iv:2

001.

0152

6v2

[cs

.CV

] 3

0 Ja

n 20

20

Page 2: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Net1

Net2

Noisy Hard Pseudo Labels by Clustering Robust Soft Pseudo Labels by MMT

Figure 1: Person image A1 and A2 belong to the same identity while B with similar appearance isfrom another person. However, clustering-generated pseudo labels in state-of-the-art UnsupervisedDomain Adaptation (UDA) methods contain much noise that hinders feature learning. We proposepseudo label refinery with on-line refined soft pseudo labels to effectively mitigate the influence ofnoisy pseudo labels and improve UDA performance on person re-ID.

To effectively address the problem of noisy pseudo labels in clustering-based UDA methods (Songet al., 2018; Zhang et al., 2019b; Yang et al., 2019) (Figure 1), we propose an unsupervised MutualMean-Teaching (MMT) framework to effectively perform pseudo label refinery by optimizing theneural networks under the joint supervisions of off-line refined hard pseudo labels and on-line refinedsoft pseudo labels. Specifically, our proposed MMT framework provides robust soft pseudo labelsin an on-line peer-teaching manner, which is inspired by the teacher-student approaches (Tarvainen& Valpola, 2017; Zhang et al., 2018b) to simultaneously train two same networks. The networksgradually capture target-domain data distributions and thus refine pseudo labels for better featurelearning. To avoid training error amplification, the temporally average model of each network isproposed to produce reliable soft labels for supervising the other network in a collaborative trainingstrategy. By training peer-networks with such on-line soft pseudo labels on the target domain, thelearned feature representations can be iteratively improved to provide more accurate soft pseudolabels, which, in turn, further improves the discriminativeness of learned feature representations.

The classification and triplet losses are commonly adopted together to achieve state-of-the-art per-formances in both fully-supervised (Luo et al., 2019) and unsupervised (Zhang et al., 2019b; Yanget al., 2019) person re-ID models. However, the conventional triplet loss (Hermans et al., 2017) can-not work with such refined soft labels. To enable using the triplet loss with soft pseudo labels in ourMMT framework, we propose a novel soft softmax-triplet loss so that the network can benefit fromsoftly refined triplet labels. The introduction of such soft softmax-triplet loss is also the key to thesuperior performance of our proposed framework. Note that the collaborative training strategy onthe two networks is only adopted in the training process. Only one network is kept in the inferencestage without requiring any additional computational or memory cost.

The contributions of this paper could be summarized as three-fold. (1) We propose to tackle thelabel noise problem in state-of-the-art clustering-based UDA methods for person re-ID, which ismostly ignored by existing methods but is shown to be crucial for achieving superior final per-formance. The proposed Mutual Mean-Teaching (MMT) framework is designed to provide morereliable soft labels. (2) Conventional triplet loss can only work with hard labels. To enable train-ing with soft triplet labels for mitigating the pseudo label noise, we propose the soft softmax-tripletloss to learn more discriminative person features. (3) The MMT framework shows exceptionallystrong performances on all UDA tasks of person re-ID. Compared with state-of-the-art methods,it leads to significant improvements of 14.4%, 18.2%, 13.4%, 16.4% mAP on Market-to-Duke,Duke-to-Market, Market-to-MSMT, Duke-to-MSMT re-ID tasks.

2 RELATED WORK

Unsupervised domain adaptation (UDA) for person re-ID. UDA methods have attracted muchattention because their capability of saving the cost of manual annotations. There are three maincategories of methods. The first category of clustering-based methods maintains state-of-the-art per-formance to date. (Fan et al., 2018) proposed to alternatively assign labels for unlabeled trainingsamples and optimize the network with the generated targets. (Lin et al., 2019) proposed a bottom-

2

Page 3: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

up clustering framework with a repelled loss. (Yang et al., 2019) introduced to assign hard pseudolabels for both global and local features. However, the training of the neural network was substan-tially hindered by the noise of the hard pseudo labels generated by clustering algorithms, which wasmostly ignored by existing methods. The second category of methods learns domain-invariant fea-tures from style-transferred source-domain images. SPGAN (Deng et al., 2018) and PTGAN (Weiet al., 2018) transformed source-domain images to match the image styles of the target domain whilemaintaining the original person identities. The style-transferred images and their identity labels werethen used to fine-tune the model. HHL (Zhong et al., 2018) learned camera-invariant features withcamera style transferred images. However, the retrieval performances of these methods deeply reliedon the image generation quality, and they did not explore the complex relations between differentsamples in the target domain. The third category of methods attempts on optimizing the neuralnetworks with soft labels for target-domain samples by computing the similarities with referenceimages or features. ENC (Zhong et al., 2019) assigned soft labels by saving averaged features withan exemplar memory module. MAR (Yu et al., 2019) conducted multiple soft-label learning bycomparing with a set of reference persons. However, the reference images and features might not berepresentative enough to generate accurate labels for achieving advanced performances.

Generic domain adaptation methods for close-set recognition. Generic domain adaptation meth-ods learn features that can minimize the differences between data distributions of source and targetdomains. Adversarial learning based methods (Zhang et al., 2018a; Tzeng et al., 2017; Ghifaryet al., 2016; Bousmalis et al., 2016; Tzeng et al., 2015) adopted a domain classifier to dispel the dis-criminative domain information from the learned features in order to reduce the domain gap. Therealso exist methods (Tzeng et al., 2014; Long et al., 2015; Yan et al., 2017; Saito et al., 2018; Ghi-fary et al., 2016) that minimize the Maximum Mean Discrepancy (MMD) loss between source- andtarget-domain distributions. However, these methods assume that the classes on different domainsare shared, which is not suitable for unsupervised domain adaptation on person re-ID.

Teacher-student models have been widely studied in semi-supervised learning methods and knowl-edge/model distillation methods. The key idea of teacher-student models is to create consistenttraining supervisions for labeled/unlabeled data via different models’ predictions. Temporal ensem-bling (Laine & Aila, 2016) maintained an exponential moving average prediction for each sampleas the supervisions of the unlabeled samples, while the mean-teacher model (Tarvainen & Valpola,2017) averaged model weights at different training iterations to create the supervisions for unlabeledsamples. Deep mutual learning (Zhang et al., 2018b) adopted a pool of student models instead ofthe teacher models by training them with supervisions from each other. However, existing methodswith teacher-student mechanisms are mostly designed for close-set recognition problems, whereboth labeled and unlabeled data share the same set of class labels and could not be directly utilizedon unsupervised domain adaptation tasks of person re-ID.

Generic methods for handling noisy labels can be classified into four categories. Loss correctionmethods (Patrini et al., 2017; Vahdat, 2017; Xiao et al., 2015) tried to model the noise transitionmatrix, however, such matrix is hard to estimate in real-world tasks, e.g. unsupervised person re-IDwith noisy pseudo labels obtained via clustering algorithm. (Veit et al., 2017; Lee et al., 2018; Liet al., 2017; Han et al., 2019) attempted to correct the noisy labels directly, while the clean set re-quired by such methods limits their generalization on real-world applications. Noise-robust methodsdesigned robust loss functions against label noises, for instance, Mean Absolute Error (MAE) loss(Ghosh et al., 2017), Generalized Cross Entropy (GCE) loss (Zhang & Sabuncu, 2018) and LabelSmoothing Regularization (LSR) (Szegedy et al., 2016). However, these methods did not studyhow to handle the triplet loss with noisy labels, which is crucial for learning discriminative featurerepresentations on person re-ID. The last kind of methods which focused on refining the trainingstrategies is mostly related to our method. Co-teaching (Han et al., 2018) trained two collabora-tive networks and conducted noisy label detection by selecting on-line clean data for each other,Co-mining (Wang et al., 2019) further extended this method on the face recognition task with are-weighting function for Arc-Softmax loss (Deng et al., 2019). However, the above methods arenot designed for the open-set person re-ID task and could not achieve state-of-the-art performancesunder the more challenge unsupervised settings.

3

Page 4: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

3 PROPOSED APPROACH

We propose a novel Mutual Mean-Teaching (MMT) framework for tackling the problem of noisypseudo labels in clustering-based Unsupervised Domain Adaptation (UDA) methods. The labelnoise has important impacts to the domain adaptation performance but was mostly ignored by thosemethods. Our key idea is to conduct pseudo label refinery in the target domain by optimizing theneural networks with off-line refined hard pseudo labels and on-line refined soft pseudo labels ina collaborative training manner. In addition, the conventional triplet loss cannot properly workwith soft labels. A novel soft softmax-triplet loss is therefore introduced to better utilize the softlyrefined pseudo labels. Both the soft classification loss and the soft softmax-triplet loss work jointlyto achieve optimal domain adaptation performances.

Formally, we denote the source domain data as Ds = {(xsi ,ysi )|Nsi=1}, where xsi and ysi denote

the i-th training sample and its associated person identity label, Ns is the number of images, andMs denotes the number of person identities (classes) in the source domain. The Nt target-domainimages are denoted as Dt = {xti|

Nti=1}, which are not associated with any ground-truth identity label.

3.1 CLUSTERING-BASED UDA METHODS REVISIT

State-of-the-art UDA methods (Fan et al., 2018; Lin et al., 2019; Zhang et al., 2019b; Yang et al.,2019) follow a similar general pipeline. They generally pre-train a deep neural networkF (·|θ) on thesource domain, where θ denotes current network parameters, and the network is then transferred tolearn from the images in the target domain. The source-domain images’ and target-domain images’features encoded by the network are denoted as {F (xsi |θ)}|

Nsi=1 and {F (xti|θ)}|

Nti=1 respectively.

As illustrated in Figure 2 (a), two operations are alternated to gradually fine-tune the pre-trainednetwork on the target domain. (1) The target-domain samples are grouped into pre-defined Mt

classes by clustering the features {F (xti|θ)}|Nti=1 output by the current network. Let yti denotes the

pseudo label generated for image xti. (2) The network parameters θ and a learnable target-domainclassifier Ct : f t → {1, · · · ,Mt} are then optimized with respect to an identity classification (cross-entropy) loss Ltid(θ) and a triplet loss (Hermans et al., 2017) Lttri(θ) in the form of,

Ltid(θ) =

1

Nt

Nt∑i=1

Lce

(Ct(F (xt

i|θ)), yti

), (1)

Lttri(θ) =

1

Nt

Nt∑i=1

max(0, ||F (xt

i|θ)− F (xti,p|θ)||+m− ||F (xt

i|θ)− F (xti,n|θ)||

), (2)

where || · || denotes the L2-norm distance, subscripts i,p and i,n indicate the hardest positive andhardest negative feature index in each mini-batch for the sample xti, and m = 0.5 denotes thetriplet distance margin. Such two operations, pseudo label generation by clustering and featurelearning with pseudo labels, are alternated until the training converges. However, the pseudo labelsgenerated in step (1) inevitably contain errors due to the imperfection of features as well as the errorsof the clustering algorithms, which hinder the feature learning in step (2). To mitigate the pseudolabel noise, we propose the Mutual Mean-Teaching (MMT) framework together with a novel softsoftmax-triplet loss to conduct the pseudo label refinery.

3.2 MUTUAL MEAN-TEACHING (MMT) FRAMEWORK

3.2.1 SUPERVISED PRE-TRAINING FOR SOURCE DOMAIN

UDA task on person re-ID aims at transferring the knowledge from a pre-trained model on the sourcedomain to the target domain. A deep neural network is first pre-trained on the source domain. Giventhe training data Ds, the network is trained to model a feature transformation function F (·|θ) thattransforms each input sample xsi into a feature representation F (xsi |θ). Given the encoded features,the identification classifier Cs outputs anMs-dimensional probability vector to predict the identitiesin the source-domain training set. The neural network is trained with a classification loss Lsid(θ)and a triplet loss Lstri(θ) to separate features belonging to different identities. The overall loss istherefore calculated as

Ls(θ) = Lsid(θ) + λsLs

tri(θ), (3)

where Lsid(θ) and Lstri(θ) are defined similarly to equation 1 and equation 2 but with ground-truthidentity labels {ysi |

Nsi=1}, and λs is the parameter weighting the two losses.

4

Page 5: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Noisy HardPseudo Labels

TemporalAverage

Net 1

MeanNet 1

Predictions

Predictions

Net 2

MeanNet 2

Predictions

Predictions

TemporalAverage

Soft Classification Loss

Soft Triplet Loss

Noisy HardPseudo Labels

Classification Loss

Triplet Loss

Cluster

Net

(b) The proposed Mutual Mean-Teaching (MMT) framework(a) General pipline for clustering-based UDA methods

(c) Inference stage with only one network for MMT

RetrievalMeanNet

Figure 2: (a) The pipeline for existing clustering-based UDA methods on person re-ID with noisyhard pseudo labels. (b) Overall framework of the proposed Mutual Mean-Teaching (MMT) withtwo collaborative networks jointly optimized under the supervisions of off-line refined hard pseudolabels and on-line refined soft pseudo labels. A soft identity classification loss and a novel softsoftmax-triplet loss are adopted. (c) One of the average models with better validated performance isadopted for inference as average models perform better than models with current parameters.

3.2.2 PSEUDO LABEL REFINERY WITH ON-LINE REFINED SOFT PSEUDO LABELS

Our proposed MMT framework is based on the clustering-based UDA methods with off-line refinedhard pseudo labels as introduced in Section 3.1, where the pseudo label generation and refinementare conducted alternatively. However, the pseudo labels generated in this way are hard (i.e., they arealways of 100% confidences) but noisy. In order to mitigate the pseudo label noise, apart from theoff-line refined hard pseudo labels, our framework further incorporates on-line refined soft pseudolabels (i.e., pseudo labels with < 100% confidences) into the training process.

Our MMT framework generates soft pseudo labels by collaboratively training two same networkswith different initializations. The overall framework is illustrated in Figure 2 (b). The pseudoclasses are still generated the same as those by existing clustering-based UDA methods, where eachcluster represents one class. In addition to the hard and noisy pseudo labels, our two collaborativenetworks also generate on-line soft pseudo labels by network predictions for training each other. Theintuition is that, after the networks are trained even with hard pseudo labels, they can roughly capturethe training data distribution and their class predictions can therefore serve as soft class labels fortraining. However, such soft labels are generally not perfect because of the training errors and noisyhard pseudo labels in the first place. To avoid two networks collaboratively bias each other, thepast temporally average model of each network instead of the current model is used to generate thesoft pseudo labels for the other network. Both off-line hard pseudo labels and on-line soft pseudolabels are utilized jointly to train the two collaborative networks. After training, only one of the pastaverage models with better validated performance is adopted for inference (see Figure 2 (c)).

We denote the two collaborative networks as feature transformation functions F (·|θ1) and F (·|θ2),and denote their corresponding pseudo label classifiers as Ct1 and Ct2, respectively. To simulta-neously train the coupled networks, we feed the same image batch to the two networks but withseparately random erasing, cropping and flipping. Each target-domain image can be denoted by xtiand x′ti for the two networks, and their pseudo label confidences can be predicted as Ct1(F (x

ti|θ1))

andCt2(F (x′ti|θ2)). One naıve way to train the collaborative networks is to directly utilize the above

pseudo label confidence vectors as the soft pseudo labels for training the other network. However, insuch a way, the two networks’ predictions might converge to equal each other and the two networkslose their output independences. The classification errors as well as pseudo label errors might beamplified during training. In order to avoid error amplification, we propose to use the temporallyaverage model of each network to generate reliable soft pseudo labels for supervising the other net-work. Specifically, the parameters of the temporally average models of the two networks at currentiteration T are denoted as E(T )[θ1] and E(T )[θ2] respectively, which can be calculated as

E(T )[θ1] = αE(T−1)[θ1] + (1− α)θ1,

E(T )[θ2] = αE(T−1)[θ2] + (1− α)θ2, (4)

where E(T−1)[θ1], E(T−1)[θ2] indicate the temporal average parameters of the two networks in theprevious iteration (T−1), the initial temporal average parameters areE(0)[θ1] = θ1,E(0)[θ2] = θ2,

5

Page 6: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

and α is the ensembling momentum to be within the range [0, 1). The robust soft pseudo labelsupervisions are then generated by the two temporal average models as Ct1(F (x

ti|E(T )[θ1])) and

Ct2(F (x′ti|E(T )[θ2])) respectively. The soft classification loss for optimizing θ1 and θ2 with the

soft pseudo labels generated from the other network can therefore be formulated as

Ltsid(θ1|θ2) = −

1

Nt

Nt∑i=1

(Ct

2(F (x′ti|E(T )[θ2])) · logCt

1(F (xti|θ1))

),

Ltsid(θ2|θ1) = −

1

Nt

Nt∑i=1

(Ct

1(F (xti|E(T )[θ1])) · logCt

2(F (x′ti|θ2))

). (5)

The two networks’ pseudo-label predictions are better dis-related by using other network’s pastaverage model to generate supervisions and can therefore better avoid error amplification.

Generalizing classification cross-entropy loss to work with soft pseudo labels has been well studied(Hinton et al., 2015), (Muller et al., 2019). However, optimizing triplet loss with soft pseudo labelsposes a great challenge as no previous method has investigated soft labels for triplet loss. Fortackling the difficulty, we propose to use softmax-triplet loss, whose hard version is formulated as

Lttri(θ1) =

1

Nt

Nt∑i=1

Lbce

(Ti(θ1),1

), (6)

where

Ti(θ1) =exp(‖F (xt

i|θ1)− F (xti,n|θ1)‖)

exp(‖F (xti|θ1)− F (xt

i,p|θ1)‖) + exp(‖F (xti|θ1)− F (xt

i,n|θ1)‖). (7)

Here Lbce(·, ·) denotes the binary cross-entropy loss, F (xti|θ1) is the encoded feature for target-domain sample xti by network 1, the subscripts i,p and i,n denote sample xti’s hardest positive andnegative samples in the mini-batch, ‖F (xti|θ1) − F (xti,p|θ1)‖ is the L2-norm distance betweensample xti and its positive sample xti,p to measure their similarity, and “1” denotes the ground-truththat the positive sample xti,p should be closer to the sample xti than its negative sample xti,n. Giventhe two collaborative networks, we can utilize the one network’s past temporal average model togenerate soft triplet labels for the other network with the proposed soft softmax-triplet loss,

Ltstri(θ1|θ2) =

1

Nt

Nt∑i=1

Lbce

(Ti(θ1), Ti

(E(T )[θ2])

)),

Ltstri(θ2|θ1) =

1

Nt

Nt∑i=1

Lbce

(Ti(θ2), Ti

(E(T )[θ1])

)), (8)

where Ti(E(T )[θ1]) and Ti(E(T )[θ2]) are the soft triplet labels generated by the two networks’ pasttemporally average models. Such soft triplet labels are fixed as training supervisions. By adoptingthe soft softmax-triplet loss, our MMT framework overcomes the limitation of hard supervisions bythe conventional triple loss (equation 2). It can be successfully trained with soft triplet labels, whichare shown to be important for improving the domain adaptation performance in our experiments.Note that such a softmax-triplet loss was also studied in (Zhang et al., 2019a). However, it has neverbeen used to generate soft labels and was not designed to work with soft pseudo labels before.

3.2.3 OVERALL LOSS AND ALGORITHM

Our proposed MMT framework is trained with both off-line refined hard pseudo labels and on-linerefined soft pseudo labels. The overall loss function L(θ1,θ2) simultaneously optimizes the couplednetworks, which combines equation 1, equation 5, equation 6, equation 8 and is formulated as,

L(θ1,θ2) = (1− λtid)(Lt

id(θ1) + Ltid(θ2)) + λt

id(Ltsid(θ1|θ2) + Lt

sid(θ2|θ1))+ (1− λt

tri)(Lttri(θ1) + Lt

tri(θ2)) + λttri(Lt

stri(θ1|θ2) + Ltstri(θ2|θ1)), (9)

where λtid, λttri are the weighting parameters. The detailed optimization procedures are summarizedin Algorithm 1. The hard pseudo labels are off-line refined after training with existing hard pseudolabels for one epoch. During the training process, the two networks are trained by combining the off-line refined hard pseudo labels and on-line refined soft labels predicted by their peers with proposedsoft losses. The noise and randomness caused by hard clustering, which lead to unstable trainingand limited final performance, can be alleviated by the proposed MMT framework.

6

Page 7: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Require: Target-domain data Dt;Require: Ensembling momentum α for equation 4, weighting factors λt

id, λttri for equation 9;

Require: Initialize pre-trained weights θ1 and θ2 by optimizing with equation 3 on Ds.for n in [1, num epochs] do

Generate hard pseudo labels yti for each sample xt

i in Dt by clustering algorithms.for each mini-batchB ⊂ Dt, iteration T do

1: Generate soft pseudo labels from the collaborative networks by predicting Ti∈B(E(T )[θ1]), Ti∈B(E(T )[θ2]),Ct

1(F (xti∈B |E

(T )[θ1])), Ct2(F (x′t

i∈B |E(T )[θ2]));

2: Joint update parameters θ1 & θ2 by the gradient descent of the objective function equation 9;3: Update temporally average model weightsE(T+1)[θ1] &E(T+1)[θ2] following equation 4.

end forend for

Algorithm 1: Unsupervised Mutual Mean-Teaching (MMT) Training Strategy

4 EXPERIMENTS

4.1 DATASETS

We evaluate our proposed MMT on three widely-used person re-ID datasets, i.e., Market-1501 (Zheng et al., 2015), DukeMTMC-reID (Ristani et al., 2016), and MSMT17 (Wei et al., 2018).The Market-1501 (Zheng et al., 2015) dataset consists of 32,668 annotated images of 1,501 identi-ties shot from 6 cameras in total, for which 12,936 images of 751 identities are used for training and19,732 images of 750 identities are in the test set. DukeMTMC-reID (Ristani et al., 2016) contains16,522 person images of 702 identities for training, and the remaining images out of another 702identities for testing, where all images are collected from 8 cameras. MSMT17 (Wei et al., 2018) isthe most challenging and large-scale dataset consisting of 126,441 bounding boxes of 4,101 identi-ties taken by 15 cameras, for which 32,621 images of 1,041 identities are spitted for training. Forevaluating the domain adaptation performance of different methods, four domain adaptation tasksare set up, i.e., Duke-to-Market, Market-to-Duke, Duke-to-MSMT and Market-to-MSMT, whereonly identity labels on the source domain are provided. Mean average precision (mAP) and CMCtop-1, top-5, top-10 accuracies are adopted to evaluate the methods’ performances.

4.2 IMPLEMENTATION DETAILS

4.2.1 TRAINING DATA ORGANIZATION

For both source-domain pre-training and target-domain fine-tuning, each training mini-batch con-tains 64 person images of 16 actual or pseudo identities (4 for each identity). Note that the generatedhard pseudo labels for the target-domain fine-tuning are updated after each epoch, so the mini-batchof target-domain images needs to be re-organized with updated hard pseudo labels after each epoch.All images are resized to 256× 128 before being fed into the networks.

4.2.2 OPTIMIZATION DETAILS

All the hyper-parameters of the proposed MMT framework are chosen based on a validation set ofthe Duke-to-Market task withMt = 500 pseudo identities and IBN-ResNet-50 backbone. The samehyper-parameters are then directly applied to the other three domain adaptation tasks. We proposea two-stage training scheme, where ADAM optimizer is adopted to optimize the networks with aweight decay of 0.0005. Randomly erasing (Zhong et al., 2017b) is only adopted in target-domainfine-tuning.

Stage 1: Source-domain pre-training. We adopt ResNet-50 (He et al., 2016) or IBN-ResNet-50(Pan et al., 2018) as the backbone networks, where IBN-ResNet-50 achieves better performances byintegrating both IN and BN modules. Two same networks are initialized with ImageNet (Deng et al.,2009) pre-trained weights. Given the mini-batch of images, network parameters θ1, θ2 are updatedindependently by optimizing equation 3 with λs = 1. The initial learning rate is set to 0.00035 andis decreased to 1/10 of its previous value on the 40th and 70th epoch in the total 80 epochs.

Stage 2: End-to-end training with MMT. Based on pre-trained weights θ1 and θ2, the twonetworks are collaboratively updated by optimizing equation 9 with the loss weights λtid = 0.5,λttri = 0.8. The temporal ensemble momentum α in equation 4 is set to 0.999. The learning rateis fixed to 0.00035 for overall 40 training epochs. We utilize k-means clustering algorithm and thenumber Mt of pseudo classes is set as 500, 700, 900 for Market-1501 and DukeMTMC-reID, and500, 1000, 1500, 2000 for MSMT17. Note that actual identity numbers in the target-domain training

7

Page 8: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Methods Market-to-Duke Duke-to-MarketmAP top-1 top-5 top-10 mAP top-1 top-5 top-10

PUL (Fan et al., 2018) (TOMM’18) 16.4 30.0 43.4 48.5 20.5 45.5 60.7 66.7TJ-AIDL (Wang et al., 2018) (CVPR’18) 23.0 44.3 59.6 65.0 26.5 58.2 74.8 81.1SPGAN (Deng et al., 2018) (CVPR’18) 22.3 41.1 56.6 63.0 22.8 51.5 70.1 76.8HHL (Zhong et al., 2018) (ECCV’18) 27.2 46.9 61.0 66.7 31.4 62.2 78.8 84.0CFSM (Chang et al., 2019) (AAAI’19) 27.3 49.8 - - 28.3 61.2 - -BUC (Lin et al., 2019) (AAAI’19) 27.5 47.4 62.6 68.4 38.3 66.2 79.6 84.5ARN (Li et al., 2018) (CVPR’18-WS) 33.4 60.2 73.9 79.5 39.4 70.3 80.4 86.3UDAP (Song et al., 2018) (Arxiv’18) 49.0 68.4 80.1 83.5 53.7 75.8 89.5 93.2ENC (Zhong et al., 2019) (CVPR’19) 40.4 63.3 75.8 80.4 43.0 75.1 87.6 91.6UCDA-CCE (Qi et al., 2019) (ICCV’19) 31.0 47.7 - - 30.9 60.4 - -PDA-Net (Li et al., 2019) (ICCV’19) 45.1 63.2 77.0 82.5 47.6 75.2 86.3 90.2PCB-PAST (Zhang et al., 2019b) (ICCV’19) 54.3 72.4 - - 54.6 78.4 - -SSG (Yang et al., 2019) (ICCV’19) 53.4 73.0 80.6 83.2 58.3 80.0 90.0 92.4Co-teaching (Han et al., 2018)-500 (ResNet-50) 55.7 71.9 83.5 88.1 65.1 82.5 91.8 93.4Co-teaching (Han et al., 2018)-500 (IBN-ResNet-50) 61.7 77.6 88.0 90.7 71.7 87.8 95.0 96.5Pre-trained (ResNet-50) 29.6 46.0 61.5 67.2 31.8 61.9 76.4 82.2Proposed MMT-500 (ResNet-50) 63.1 76.8 88.0 92.2 71.2 87.7 94.9 96.9Proposed MMT-700 (ResNet-50) 65.1 78.0 88.8 92.5 69.0 86.8 94.6 96.9Proposed MMT-900 (ResNet-50) 63.1 77.4 88.1 92.5 66.2 86.8 94.9 96.6Pre-trained (IBN-ResNet-50) 35.4 54.0 67.7 72.9 35.6 65.3 79.7 84.3Proposed MMT-500 (IBN-ResNet-50) 65.7 79.3 89.1 92.4 76.5 90.9 96.4 97.9Proposed MMT-700 (IBN-ResNet-50) 68.7 81.8 91.2 93.4 74.5 91.1 96.5 98.2Proposed MMT-900 (IBN-ResNet-50) 67.3 80.8 90.3 93.0 72.7 91.2 96.3 98.0

Methods Market-to-MSMT Duke-to-MSMTmAP top-1 top-5 top-10 mAP top-1 top-5 top-10

PTGAN (Wei et al., 2018) (CVPR’18) 2.9 10.2 - 24.4 3.3 11.8 - 27.4ENC (Zhong et al., 2019) (CVPR’19) 8.5 25.3 36.3 42.1 10.2 30.2 41.5 46.8SSG (Yang et al., 2019) (ICCV’19) 13.2 31.6 - 49.6 13.3 32.2 - 51.2Pre-trained (ResNet-50) 7.1 19.4 28.9 34.2 9.4 27.0 38.1 43.7Proposed MMT-500 (ResNet-50) 16.6 37.5 50.6 56.5 17.9 41.3 54.2 59.7Proposed MMT-1000 (ResNet-50) 21.6 46.1 59.8 66.1 23.5 50.0 63.6 69.2Proposed MMT-1500 (ResNet-50) 22.9 49.2 63.1 68.8 23.3 50.1 63.9 69.8Proposed MMT-2000 (ResNet-50) 20.8 45.7 59.6 65.6 22.4 49.0 62.5 67.8Pre-trained (IBN-ResNet-50) 9.5 25.3 36.2 41.6 11.9 32.6 44.7 50.4Proposed MMT-500 (IBN-ResNet-50) 19.6 43.3 56.1 61.6 23.3 50.0 62.8 68.4Proposed MMT-1000 (IBN-ResNet-50) 26.3 52.5 66.3 71.7 29.7 58.8 71.0 76.1Proposed MMT-1500 (IBN-ResNet-50) 26.6 54.4 67.6 72.9 29.3 58.2 71.6 76.8Proposed MMT-2000 (IBN-ResNet-50) 25.1 52.7 65.9 71.3 28.1 56.8 70.8 76.0

Table 1: Experimental results of the proposed MMT and state-of-the-art methods on Market-1501 (Zheng et al., 2015), DukeMTMC-reID (Ristani et al., 2016), and MSMT17 (Wei et al., 2018)datasets, where MMT-Mt represents the result withMt pseudo classes. Note that none ofMt valuesequals the actual number of identities but our method still outperforms all state-of-the-arts.

sets are different from Mt. We test different Mt values that are either smaller or greater than actualnumbers.

4.3 COMPARISON WITH STATE-OF-THE-ARTS

We compare our proposed MMT framework with state-of-the-art methods on the four domain adap-tation tasks, Market-to-Duke, Duke-to-Market, Market-to-MSMT and Duke-to-MSMT. The resultsare shown in Table 1. Our MMT framework significantly outperforms all existing approaches withboth ResNet-50 and IBN-ResNet-50 backbones, which verifies the effectiveness of our method.Moreover, we almost approach fully-supervised learning performances (Sun et al., 2018; Ge et al.,2018) without any manual annotations on the target domain. No post-processing technique, e.g.re-ranking (Zhong et al., 2017a) or multi-query fusion (Zheng et al., 2015), is adopted.

Specifically, by adopting the ResNet-50 (He et al., 2016) backbone, we surpass the state-of-the-art clustering-based SSG (Yang et al., 2019) by considerable margins of 11.7% and 12.9% mAPon Market-to-Duke and Duke-to-Market tasks with simpler network architectures and lower outputfeature dimensions. Furthermore, evident 9.7% and 10.2% mAP gains are achieved on Market-to-MSMT and Duke-to-MSMT tasks. Recall that Mt is the number of clusters or number of hardpseudo labels manually specified. More importantly, we achieve state-of-the-art performances onall tested target datasets with different Mt, which are either fewer or more than the actual number ofidentities in the training set of the target domain. Such results prove the necessity and effectivenessof our proposed pseudo label refinery for hard pseudo labels with inevitable noises.

8

Page 9: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Duke-to-Market IBN-ResNet-50 ResNet-50mAP top-1 top-5 top-10 mAP top-1 top-5 top-10

Pre-trained (only Lsid & Ls

tri) 35.6 65.3 79.7 84.3 31.8 61.9 76.4 82.2Baseline (only Lt

id & Lttri) 62.7 84.4 92.7 95.5 53.5 76.0 88.1 91.9

Baseline+MMT-500 (only Ltsid & Lt

stri) 34.5 59.7 73.0 78.0 22.4 46.5 61.5 67.4Baseline+MMT-500 (w/o Lt

id) 38.0 63.4 74.9 79.4 24.9 50.3 64.0 69.8Baseline+MMT-500 (w/o Lt

tri) 76.2 90.8 96.6 97.9 72.0 87.8 95.5 96.9Baseline+MMT-500 (w/o Lt

sid) 69.6 87.4 95.2 96.7 62.6 84.0 93.4 95.4Baseline+MMT-500 (w/o Lt

stri) 71.7 88.5 95.1 96.6 65.9 84.0 93.1 95.5Baseline+MMT-500 (w/o θ2) 72.8 89.1 95.2 97.1 67.5 86.1 94.3 96.1Baseline+MMT-500 (w/oE[θ]) 72.1 88.7 95.4 97.3 62.3 80.5 91.3 94.0Baseline+MMT-500 76.5 90.9 96.4 97.9 71.2 87.7 94.9 96.9

Market-to-Duke IBN-ResNet-50 ResNet-50mAP top-1 top-5 top-10 mAP top-1 top-5 top-10

Pre-trained (only Lsid & Ls

tri) 35.4 54.0 67.7 72.9 29.6 46.0 61.5 67.2Baseline (only Lt

id & Lttri) 55.0 72.3 84.4 88.1 48.2 66.4 79.8 84.0

Baseline+MMT-500 (only Ltsid & Lt

stri) 24.5 38.0 50.1 56.1 13.6 24.3 36.4 42.5Baseline+MMT-500 (w/o Lt

id) 27.5 42.0 53.9 60.3 15.3 25.8 37.7 43.7Baseline+MMT-500 (w/o Lt

tri) 65.6 79.4 89.8 92.3 63.0 77.3 88.3 91.6Baseline+MMT-500 (w/o Lt

sid) 60.3 75.7 86.6 89.9 58.1 74.9 85.2 89.5Baseline+MMT-500 (w/o Lt

stri) 61.7 77.1 86.5 89.6 59.5 73.9 85.5 88.8Baseline+MMT-500 (w/o θ2) 62.1 77.6 86.8 89.7 58.2 74.1 86.0 89.3Baseline+MMT-500 (w/oE[θ]) 61.1 76.3 86.6 89.8 55.7 70.0 83.6 87.2Baseline+MMT-500 65.7 79.3 89.1 92.4 63.1 76.8 88.0 92.2

Table 2: Ablation studies of our proposed MMT on Duke-to-Market and Market-to-Duke tasks withMt of 500. Note that the actual numbers of identities are not equal to 500 for both datasets but ourMMT method still shows significant improvements.

To compare with relevant methods for tackling general noisy label problems, we implement Co-teaching (Han et al., 2018) on unsupervised person re-ID task with 500 pseudo identities on thetarget domain, where the noisy labels are generated by the same clustering algorithm as our MMTframework. The hard classification (cross-entropy) loss is adopted on selected clean batches. All thehyper-parameters are set as the same for fair comparison, and the experimental results are denotedas “Co-teaching (Han et al., 2018)-500” with both ResNet-50 and IBN-ResNet-50 backbones inTable 1. Comparing “Co-teaching (Han et al., 2018)-500 (ResNet-50)” with “Proposed MMT-500(ResNet-50)”, we observe significant 7.4% and 6.1% mAP drops on Market-to-Duke and Duke-to-Market tasks respectively, since Co-teaching (Han et al., 2018) is designed for general close-setrecognition problems with manually generated label noise, which could not tackle the real-worldchallenges in unsupervised person re-ID. More importantly, it does not explore how to mitigate thelabel noise for the triplet loss as our method does.

4.4 ABLATION STUDIES

In this section, we evaluate each component in our proposed framework by conducting ablationstudies on Duke-to-Market and Market-to-Duke tasks with both ResNet-50 (He et al., 2016) andIBN-ResNet-50 (Pan et al., 2018) backbones. Results are shown in Table 2.

Effectiveness of the soft pseudo label refinery. To investigate the necessity of handling noisypseudo labels in clustering-based UDA methods, we create baseline models that utilize only off-linerefined hard pseudo labels, i.e., optimizing equation 9 with λtid = λttri = 0 for the two-step trainingstrategy in Section 3.1. The baseline model performances are present in Table 2 as “Baseline (onlyLtid & Lttri)”. Considerable drops of 17.7% and 14.9% mAP are observed on ResNet-50 for Duke-to-Market and Market-to-Duke tasks. Similarly, 13.8% and 10.7% mAP decreases are shown on theIBN-ResNet-50 backbone. Stable increases achieved by the proposed on-line refined soft pseudolabels on different datasets and backbones demonstrate the necessity of soft pseudo label refineryand the effectiveness of our proposed MMT framework.

Effectiveness of the soft softmax-triplet loss. We also verify the effectiveness of soft softmax-triplet loss with softly refined triplet labels in our proposed MMT framework. Experiments ofremoving the soft softmax-triplet loss, i.e., λttri = 0 in equation 9, but keeping the hard softmax-triplet loss (equation 6) are conducted, which are denoted as “Baseline+MMT-500 (w/o Ltstri)”.All experiments without the supervision of soft triplet loss show distinct drops on Duke-to-Marketand Market-to-Duke tasks, which indicate that the hard pseudo label with hard triplet loss hindersthe feature learning capability because it ignores pseudo label noise by the clustering algorithms.

9

Page 10: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Specifically, the mAP drops are 5.3% on ResNet-50 and 4.8% on IBN-ResNet-50 when evaluatingon the target dataset Market-1501. As for the Market-to-Duke task, similar mAP drops of 3.6% and4.0% on the two network structures can be observed. An evident improvement of up to 5.3% mAPdemonstrates the usefulness of our proposed soft softmax-triplet loss.

Effectiveness of Mutual Mean-Teaching. We propose to generate on-line refined soft pseudo labelsfor one network with the predictions of the past average model of the other network in our MMTframework, i.e., the soft labels for network 1 are output from the average model of network 2 and viceversa. We observe that the soft labels generated in such manner are more reliable due to the betterdecoupling between the past temporally average models of the two networks. Such a frameworkcould effectively avoid bias amplification even when the networks have much erroneous outputsin the early training epochs. There are two possible simplification our MMT framework with lessde-coupled structures. The first one is to keep only one network in our framework and use itspast temporal average model to generate soft pseudo labels for training itself. Such experimentsare denoted as “Baseline+MMT-500 (w/o θ2)”. The second simplification is to naıvely use onenetwork’s current-iteration predictions as the soft pseudo labels for training the other network andvice versa, i.e., α = 0 for equation 4. This set of experiments are denoted as “Baseline+MMT-500(w/o E[θ])”. Significant mAP drops compared to our proposed MMT could be observed in thetwo sets of experiments, especially when using the ResNet-50 backbone, e.g. the mAP drops by8.9% on Duke-to-Market task when removing past average models. This validates the necessity ofemploying the proposed mutual mean-teaching scheme for providing more robust soft pseudo labels.In despite of the large margin of performance declines when removing either the peer network or thepast average model, our proposed MMT outperforms the baseline model significantly, which furtherdemonstrates the importance of adopting the proposed on-line refined soft pseudo labels.

Necessity of hard pseudo labels in proposed MMT. Despite the robust soft pseudo labels bringsignificant improvements, the noisy hard pseudo labels are still essential to our proposed frame-work, since the hard classification loss Ltid is the foundation for capturing the target-domain datadistributions. To investigate the contribution of Ltid in the final training objective function as equa-tion 9, we conduct two experiments. (1) “Baseline+MMT-500 (only Ltsid & Ltstri)” by removingboth hard classification loss and hard triplet loss with λtid = λttri = 1; (2)“Baseline+MMT-500(w/o Ltid)” by removing only hard classification loss with λtid = 1. As illustrated in Table 2, theabove two experiments both result in much lower performances than the model pre-trained on thesource domain (“Pre-trained (only Lsid & Lstri)”), which effectively validate the necessity of Ltid.The initial network usually outputs uniform probabilities for each identity, which act as soft labelsfor soft classification loss, since it could not correctly distinguish between different identities on thetarget domain. Directly training with such smooth and noisy soft pseudo labels, the networks inour framework would soon collapse due to the large bias. One-hot hard labels for classification lossare critical for learning discriminative representations on the target domain. In contrast, the hardtriplet loss Lttri is not absolutely necessary in our framework, as experiments without Lttri, denotedas “Baseline+MMT-500 (w/o Lttri)” with λttri = 1.0, show similar performances as our final resultswith λttri = 0.8. It is much easier to learn to predict robust soft labels for the soft softmax-triplet lossin equation 8 even at early training epochs, which has only two classes, i.e., positive and negative.

5 CONCLUSION

In this work, we propose an unsupervised Mutual Mean-Teaching (MMT) framework to tackle theproblem of noisy pseudo labels in clustering-based unsupervised domain adaptation methods forperson re-ID. The key is to conduct pseudo label refinery to better model inter-sample relations inthe target domain by optimizing with the off-line refined hard pseudo labels and on-line refinedsoft pseudo labels in a collaborative training manner. Moreover, a novel soft softmax-triplet loss isproposed to support learning with softly refined triplet labels for optimal performances. Our methodsignificantly outperforms all existing person re-ID methods on domain adaptation task with up to18.2% improvements.

ACKNOWLEDGMENTS

This work is supported by the General Research Fund sponsored by the Research Grants Council ofHong Kong (Nos. CUHK14208417, CUHK14239816, CUHK14207319), the Hong Kong Innova-tion and Technology Support Program (No. ITS/312/18FX).

10

Page 11: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

REFERENCES

Konstantinos Bousmalis, George Trigeorgis, Nathan Silberman, Dilip Krishnan, and Dumitru Erhan.Domain separation networks. In NIPS, 2016.

Xiaobin Chang, Yongxin Yang, Tao Xiang, and Timothy M Hospedales. Disjoint label space transferlearning with common factorised space. AAAI, 2019.

Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. Imagenet: A large-scalehierarchical image database. 2009.

Jiankang Deng, Jia Guo, Niannan Xue, and Stefanos Zafeiriou. Arcface: Additive angular marginloss for deep face recognition. In CVPR, pp. 4690–4699, 2019.

Weijian Deng, Liang Zheng, Qixiang Ye, Guoliang Kang, Yi Yang, and Jianbin Jiao. Image-image domain adaptation with preserved self-similarity and domain-dissimilarity for person re-identification. In CVPR, 2018.

Hehe Fan, Liang Zheng, Chenggang Yan, and Yi Yang. Unsupervised person re-identification:Clustering and fine-tuning. 2018.

Yixiao Ge, Zhuowan Li, Haiyu Zhao, Guojun Yin, Shuai Yi, Xiaogang Wang, et al. Fd-gan: Pose-guided feature distilling gan for robust person re-identification. In NeurIPS, 2018.

Muhammad Ghifary, W Bastiaan Kleijn, Mengjie Zhang, David Balduzzi, and Wen Li. Deepreconstruction-classification networks for unsupervised domain adaptation. In ECCV, 2016.

Aritra Ghosh, Himanshu Kumar, and PS Sastry. Robust loss functions under label noise for deepneural networks. In AAAI, 2017.

Bo Han, Quanming Yao, Xingrui Yu, Gang Niu, Miao Xu, Weihua Hu, Ivor Tsang, and MasashiSugiyama. Co-teaching: Robust training of deep neural networks with extremely noisy labels. InNIPS, pp. 8527–8537, 2018.

Jiangfan Han, Ping Luo, and Xiaogang Wang. Deep self-learning from noisy labels. In ICCV, 2019.

Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun. Deep residual learning for image recog-nition. In CVPR, 2016.

Alexander Hermans, Lucas Beyer, and Bastian Leibe. In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737, 2017.

Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. Distilling the knowledge in a neural network. arXivpreprint arXiv:1503.02531, 2015.

Samuli Laine and Timo Aila. Temporal ensembling for semi-supervised learning. arXiv preprintarXiv:1610.02242, 2016.

Kuang-Huei Lee, Xiaodong He, Lei Zhang, and Linjun Yang. Cleannet: Transfer learning forscalable image classifier training with label noise. In CVPR, pp. 5447–5456, 2018.

Yu-Jhe Li, Fu-En Yang, Yen-Cheng Liu, Yu-Ying Yeh, Xiaofei Du, and Yu-Chiang Frank Wang.Adaptation and re-identification network: An unsupervised deep transfer learning approach toperson re-identification. In CVPRW, 2018.

Yu-Jhe Li, Ci-Siang Lin, Yan-Bo Lin, and Yu-Chiang Frank Wang. Cross-dataset person re-identification via unsupervised pose disentanglement and adaptation. ICCV, 2019.

Yuncheng Li, Jianchao Yang, Yale Song, Liangliang Cao, Jiebo Luo, and Li-Jia Li. Learning fromnoisy labels with distillation. In ICCV, pp. 1910–1918, 2017.

Yutian Lin, Xuanyi Dong, Liang Zheng, Yan Yan, and Yi Yang. A bottom-up clustering approach tounsupervised person re-identification. In AAAI, 2019.

Mingsheng Long, Yue Cao, Jianmin Wang, and Michael I Jordan. Learning transferable featureswith deep adaptation networks. arXiv preprint arXiv:1502.02791, 2015.

11

Page 12: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Hao Luo, Youzhi Gu, Xingyu Liao, Shenqi Lai, and Wei Jiang. Bag of tricks and a strong baselinefor deep person re-identification. In CVPRW, 2019.

Rafael Muller, Simon Kornblith, and Geoffrey Hinton. When does label smoothing help? arXivpreprint arXiv:1906.02629, 2019.

Xingang Pan, Ping Luo, Jianping Shi, and Xiaoou Tang. Two at once: Enhancing learning andgeneralization capacities via ibn-net. In ECCV, 2018.

Giorgio Patrini, Alessandro Rozza, Aditya Krishna Menon, Richard Nock, and Lizhen Qu. Makingdeep neural networks robust to label noise: A loss correction approach. In CVPR, pp. 1944–1952,2017.

Lei Qi, Lei Wang, Jing Huo, Luping Zhou, Yinghuan Shi, and Yang Gao. A novel unsupervisedcamera-aware domain adaptation framework for person re-identification. ICCV, 2019.

Ergys Ristani, Francesco Solera, Roger Zou, Rita Cucchiara, and Carlo Tomasi. Performance mea-sures and a data set for multi-target, multi-camera tracking. In ECCVW, 2016.

Kuniaki Saito, Kohei Watanabe, Yoshitaka Ushiku, and Tatsuya Harada. Maximum classifier dis-crepancy for unsupervised domain adaptation. In CVPR, 2018.

Liangchen Song, Cheng Wang, Lefei Zhang, Bo Du, Qian Zhang, Chang Huang, and XinggangWang. Unsupervised domain adaptive re-identification: Theory and practice. arXiv preprintarXiv:1807.11334, 2018.

Yifan Sun, Liang Zheng, Yi Yang, Qi Tian, and Shengjin Wang. Beyond part models: Personretrieval with refined part pooling (and a strong convolutional baseline). In ECCV, 2018.

Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. Rethinkingthe inception architecture for computer vision. In CVPR, pp. 2818–2826, 2016.

Antti Tarvainen and Harri Valpola. Mean teachers are better role models: Weight-averaged consis-tency targets improve semi-supervised deep learning results. In NIPS, 2017.

Eric Tzeng, Judy Hoffman, Ning Zhang, Kate Saenko, and Trevor Darrell. Deep domain confusion:Maximizing for domain invariance. arXiv preprint arXiv:1412.3474, 2014.

Eric Tzeng, Judy Hoffman, Trevor Darrell, and Kate Saenko. Simultaneous deep transfer acrossdomains and tasks. In CVPR, 2015.

Eric Tzeng, Judy Hoffman, Kate Saenko, and Trevor Darrell. Adversarial discriminative domainadaptation. In CVPR, 2017.

Arash Vahdat. Toward robustness against label noise in training deep discriminative neural networks.In NIPS, pp. 5596–5605, 2017.

Andreas Veit, Neil Alldrin, Gal Chechik, Ivan Krasin, Abhinav Gupta, and Serge Belongie. Learningfrom noisy large-scale datasets with minimal supervision. In CVPR, pp. 839–847, 2017.

Jingya Wang, Xiatian Zhu, Shaogang Gong, and Wei Li. Transferable joint attribute-identity deeplearning for unsupervised person re-identification. In CVPR, 2018.

Xiaobo Wang, Shuo Wang, Jun Wang, Hailin Shi, and Tao Mei. Co-mining: Deep face recognitionwith noisy labels. In ICCV, pp. 9358–9367, 2019.

Longhui Wei, Shiliang Zhang, Wen Gao, and Qi Tian. Person transfer gan to bridge domain gap forperson re-identification. In CVPR, 2018.

Tong Xiao, Tian Xia, Yi Yang, Chang Huang, and Xiaogang Wang. Learning from massive noisylabeled data for image classification. In CVPR, pp. 2691–2699, 2015.

Hongliang Yan, Yukang Ding, Peihua Li, Qilong Wang, Yong Xu, and Wangmeng Zuo. Mind theclass weight bias: Weighted maximum mean discrepancy for unsupervised domain adaptation. InCVPR, 2017.

12

Page 13: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

Fu Yang, Wei Yunchao, Wang Guanshuo, Zhou Yuqian, Shi Honghui, and Huang Thomas. Self-similarity grouping: A simple unsupervised cross domain adaptation approach for person re-identification. ICCV, 2019.

Hong-Xing Yu, Wei-Shi Zheng, Ancong Wu, Xiaowei Guo, Shaogang Gong, and Jian-Huang Lai.Unsupervised person re-identification by soft multilabel learning. In CVPR, 2019.

Ji Zhang, Yannis Kalantidis, Marcus Rohrbach, Manohar Paluri, Ahmed Elgammal, and MohamedElhoseiny. Large-scale visual relationship understanding. In AAAI, 2019a.

Weichen Zhang, Wanli Ouyang, Wen Li, and Dong Xu. Collaborative and adversarial network forunsupervised domain adaptation. In CVPR, 2018a.

Xinyu Zhang, Jiewei Cao, Chunhua Shen, and Mingyu You. Self-training with progressive augmen-tation for unsupervised cross-domain person re-identification. ICCV, 2019b.

Ying Zhang, Tao Xiang, Timothy M Hospedales, and Huchuan Lu. Deep mutual learning. In CVPR,2018b.

Zhilu Zhang and Mert Sabuncu. Generalized cross entropy loss for training deep neural networkswith noisy labels. In NIPS, pp. 8778–8788, 2018.

Liang Zheng, Liyue Shen, Lu Tian, Shengjin Wang, Jingdong Wang, and Qi Tian. Scalable personre-identification: A benchmark. In ICCV, 2015.

Zhun Zhong, Liang Zheng, Donglin Cao, and Shaozi Li. Re-ranking person re-identification withk-reciprocal encoding. In CVPR, 2017a.

Zhun Zhong, Liang Zheng, Guoliang Kang, Shaozi Li, and Yi Yang. Random erasing data augmen-tation. arXiv preprint arXiv:1708.04896, 2017b.

Zhun Zhong, Liang Zheng, Shaozi Li, and Yi Yang. Generalizing a person retrieval model hetero-and homogeneously. In ECCV, 2018.

Zhun Zhong, Liang Zheng, Zhiming Luo, Shaozi Li, and Yi Yang. Invariance matters: Exemplarmemory for domain adaptive person re-identification. In CVPR, 2019.

A APPENDIX

A.1 FUNCTIONS OF TEMPORAL AVERAGE MODELS IN MMT

Figure 3: The predictions of temporal average models (denoted as “Proposed MMT-500”) serveas more complementary and robust soft pseudo labels than those of ordinary networks (denoted as“Proposed MMT-500 (w/o E[θ])”).

Two temporal average models are introduced in our proposed MMT framework to provide morecomplementary soft labels and avoid training error amplification. Such average models are more

13

Page 14: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

de-coupled by ensembling the past parameters and provide more independent predictions, which isignored by previous methods with peer-teaching strategy (Han et al., 2018; Wang et al., 2019; Zhanget al., 2018b). Despite we have verified the effectiveness of such design in Table 2 by removing thetemporal average model, denoted as “Baseline+MMT-500 (w/o E[θ])”, we would like to visualizethe training process by plotting the KL divergence between peer networks’ predictions for furthercomparison. As illustrated in Figure 3, the predictions by two temporal average models (“ProposedMMT-500”) always keep a larger distance than predictions by two ordinary networks (“ProposedMMT-500 (w/o E[θ])”), which indicates that the temporal average models could prevent the twonetworks in our MMT from converging to each other soon under the collaborative training strategy.

A.2 PARAMETER ANALYSIS

Figure 4: Performance evaluation of our proposed MMT-500 with different values of λttri and λtidin equation 9 on Duke-to-Market and Market-to-Duke tasks in terms of mAP(%) and top-1(%)accuracies. Weighting factors λttri and λtid balance the contributions between hard and soft pseudolabels. Specifically, only hard labels are adopted when the weighting factors are set to 0.0, and onlysoft labels are utilized when the weighting factors are set to 1.0.

We utilize weighting factors of λttri = 0.8, λtid = 0.5 in all our experiments by tuning on Duke-to-Market task with IBN-ResNet-50 backbone and 500 pseudo identities. To further analyse theimpact of different λttri and λtid on different tasks, we conduct comparison experiments by varyingthe value of one parameter and keep the others fixed. Our MMT framework is robust and insensitiveto different parameters except when the hard classification loss is eliminated with λtid = 1.0.The weighting factor of hard and soft triplet losses λttri. In Figure 4 (a-b), we investigate theeffect of the weighting factor λttri in equation 9, where the weight for soft softmax-triplet loss isλttri and the weight for hard triplet loss is (1 − λttri). We test our proposed MMT-500 with bothResNet-50 and IBN-ResNet-50 backbones when λttri is varying from 0.0, 0.3, 0.5, 0.8 and 1.0.Specifically, the soft softmax-triplet loss is removed from the final training objective (equation 9)when λttri is equal to 0.0, and the hard triplet loss is eliminated when λttri is set to 1.0. We observe

14

Page 15: M -TEACHING P L REFINERY FOR UNSUPERVISED DO MAIN ... · labels for both global and local features. However, the training of the neural network was substan-tially hindered by the

Published as a conference paper at ICLR 2020

that the accuracies are almost in direct ratio to the value of λttri which indicate the effectivenessof our proposed novel soft softmax-triplet loss. MMT-500 achieves optimal performances withResNet-50 backbone on both two tasks when λttri = 1.0. With the backbone of IBN-ResNet-50,MMT-500 obtains the best results with λttri = 0.8 on Duke-to-Market and λttri = 0.5 on Market-to-Duke. Despite the performances vary with different values of λttri, all the results by our methodoutperform state-of-the-arts significantly.The weighting factor of hard and soft classification losses λtid. Similar to the comparisons ofλttri, we evaluate our proposed MMT-500 framework with different values of λtid, which is theweighting factor for hard and soft classification losses in equation 9. As illustrated in Figure 4 (c-d),we observe considerable declines when the hard classification loss equation 1 is eliminated withλtid = 1.0. Hard classification loss is essential to our proposed framework, which is fully analysedin Section 4.4. We achieve the optimal performances on both two tasks when λtid = 0.5, while allthe experiments with λtid < 1 outperform state-of-the-arts by large margins.

15


Recommended