+ All Categories
Home > Documents > Eigenfunction-Based Multitask Learning in a Reproducing...

Eigenfunction-Based Multitask Learning in a Reproducing...

Date post: 31-Dec-2020
Category:
Upload: others
View: 3 times
Download: 0 times
Share this document with a friend
13
1818 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019 Eigenfunction-Based Multitask Learning in a Reproducing Kernel Hilbert Space Xinmei Tian , Member, IEEE, Ya Li, Tongliang Liu , Xinchao Wang, and Dacheng Tao , Fellow, IEEE Abstract— Multitask learning aims to improve the performance on related tasks by exploring the interdependence among them. Existing multitask learning methods explore the relatedness among tasks on the basis of the input features and the model parameters. In this paper, we focus on nonparametric multitask learning and propose to measure task relatedness from a novel perspective in a reproducing kernel Hilbert space (RKHS). Past works have shown that the objective function for a given task can be approximated using the top eigenvalues and corresponding eigenfunctions of a predefined integral operator on an RKHS. In our method, we formulate our objective for multitask learning as a linear combination of two sets of eigenfunctions, common eigenfunctions shared by different tasks and unique eigenfunc- tions in individual tasks, such that the eigenfunctions for one task can provide additional information on another and help to improve its performance. We present both theoretical and empiri- cal validations of our proposed approach. The theoretical analysis demonstrates that our learning algorithm is uniformly argument stable and that the convergence rate of the generalization upper bound can be improved by learning multiple tasks. Experiments on several benchmark multitask learning data sets show that our method yields promising results. Index Terms— Eigenfunction-based learning, multitask learning, regression, task relatedness. I. I NTRODUCTION I N RECENT years, multitask learning has been widely studied in various fields, such as metric learning [1]–[3], image and video research [4], and disease prediction [5], [6]. The main advantage of multitask learning is the ability to explore the intrinsic interdependence among different tasks, through which all tasks can benefit each other. As a result, Manuscript received February 27, 2018; revised June 14, 2018 and September 8, 2018; accepted September 25, 2018. Date of publication October 29, 2018; date of current version May 23, 2019. This work was supported in part by the National Key Research and Development Program of China under Grant 2017YFB1002203, in part by the National Natural Science Foundation of China under Grant 61872329 and Grant 61572451, in part by Fok Ying Tung Education Foundation under Grant WF2100060004, and in part by the Australian Research Council Projects under Grant FL-170100117, Grant DP-180103424, and Grant IH180100002. (Corresponding author: Xinmei Tian.) X. Tian and Y. Li are with the CAS Key Laboratory of Technology in Geo-Spatial Information Processing and Application Systems, Univer- sity of Science and Technology of China, Hefei 230027, China (e-mail: [email protected]; [email protected]). T. Liu and D. Tao are with the UBTECH Sydney Artificial Intelligence Centre, School of Information Technologies, Faculty of Engineering and Information Technologies, The University of Sydney, Darlington, NSW 2008, Australia (e-mail: [email protected]; [email protected]). X. Wang is with the Department of Computer Science, Stevens Institute of Technology, Hoboken, NJ 07030, USA (e-mail: [email protected]). Color versions of one or more of the figures in this paper are available online at http://ieeexplore.ieee.org. Digital Object Identifier 10.1109/TNNLS.2018.2873649 multitask learning methods usually achieve better performance than single-task learning methods. The key challenge of multitask learning is measuring the relationships among different tasks. Existing multitask learn- ing methods can be categorized into two classes based on the methods used to measure the task relationships. In the first class of methods, it is assumed that related tasks share a set of common features [7]–[10], while in the second class of methods, it is assumed that different tasks share some common parameters [10]–[14]. Both classes of methods involve the imposition of direct regularizations, on either features or parameters, to learn the relatedness of multiple tasks. However, some of these regularizations are too strong, making the objective functions difficult to solve. In this paper, we propose a novel multitask learning algorithm that utilizes a different measure of task related- ness. Unlike previous methods, in which task relatedness is measured using common features or model parameters, our method measures the interdependence among tasks through the relatedness of eigenfunctions. The objective function for a particular task, in a manner similar to regression, can be approximated as a linear combination of the top eigenvalues of a predefined integral operator on a reproducing kernel Hilbert space (RKHS) [15]–[17]. In our method, we assume that related tasks share a set of common eigenfunctions and that each task also has a set of unique eigenfunctions, which are sparse. We formulate our objective function as a linear combi- nation of both sets of eigenfunctions, such that the functions associated with one task may provide additional information to benefit others. Moreover, since the eigenfunctions can be explicitly computed from the input features with an associated kernel function, our method can be readily extended to any type of kernel version. Please note that our method is not suitable for parametric multitask learning problems [18]. We propose an efficient optimization algorithm for solving our objective function, which has two regularization terms. One is an L1-norm regularization to guarantee the sparsity of the task-specific eigenfunctions. The other is an L2-norm regularization on the shared eigenfunctions to constrain the complexity of the trained model. We present a theoretical analysis to show that our learning algorithm is uniformly argument stable, meaning that the output is not sensitive to subtle changes in the input. In addition, we show that the convergence rate of the generalization upper bound is related to the number of training samples and the number of tasks. This means that when either the training set size or the number 2162-237X © 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission. See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.
Transcript
Page 1: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1818 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

Eigenfunction-Based Multitask Learningin a Reproducing Kernel Hilbert Space

Xinmei Tian , Member, IEEE, Ya Li, Tongliang Liu , Xinchao Wang, and Dacheng Tao , Fellow, IEEE

Abstract— Multitask learning aims to improve the performanceon related tasks by exploring the interdependence among them.Existing multitask learning methods explore the relatednessamong tasks on the basis of the input features and the modelparameters. In this paper, we focus on nonparametric multitasklearning and propose to measure task relatedness from a novelperspective in a reproducing kernel Hilbert space (RKHS). Pastworks have shown that the objective function for a given task canbe approximated using the top eigenvalues and correspondingeigenfunctions of a predefined integral operator on an RKHS.In our method, we formulate our objective for multitask learningas a linear combination of two sets of eigenfunctions, commoneigenfunctions shared by different tasks and unique eigenfunc-tions in individual tasks, such that the eigenfunctions for onetask can provide additional information on another and help toimprove its performance. We present both theoretical and empiri-cal validations of our proposed approach. The theoretical analysisdemonstrates that our learning algorithm is uniformly argumentstable and that the convergence rate of the generalization upperbound can be improved by learning multiple tasks. Experimentson several benchmark multitask learning data sets show that ourmethod yields promising results.

Index Terms— Eigenfunction-based learning, multitasklearning, regression, task relatedness.

I. INTRODUCTION

IN RECENT years, multitask learning has been widelystudied in various fields, such as metric learning [1]–[3],

image and video research [4], and disease prediction [5], [6].The main advantage of multitask learning is the ability toexplore the intrinsic interdependence among different tasks,through which all tasks can benefit each other. As a result,

Manuscript received February 27, 2018; revised June 14, 2018 andSeptember 8, 2018; accepted September 25, 2018. Date of publicationOctober 29, 2018; date of current version May 23, 2019. This work wassupported in part by the National Key Research and Development Programof China under Grant 2017YFB1002203, in part by the National NaturalScience Foundation of China under Grant 61872329 and Grant 61572451,in part by Fok Ying Tung Education Foundation under Grant WF2100060004,and in part by the Australian Research Council Projects underGrant FL-170100117, Grant DP-180103424, and Grant IH180100002.(Corresponding author: Xinmei Tian.)

X. Tian and Y. Li are with the CAS Key Laboratory of Technologyin Geo-Spatial Information Processing and Application Systems, Univer-sity of Science and Technology of China, Hefei 230027, China (e-mail:[email protected]; [email protected]).

T. Liu and D. Tao are with the UBTECH Sydney Artificial IntelligenceCentre, School of Information Technologies, Faculty of Engineering andInformation Technologies, The University of Sydney, Darlington, NSW 2008,Australia (e-mail: [email protected]; [email protected]).

X. Wang is with the Department of Computer Science, Stevens Institute ofTechnology, Hoboken, NJ 07030, USA (e-mail: [email protected]).

Color versions of one or more of the figures in this paper are availableonline at http://ieeexplore.ieee.org.

Digital Object Identifier 10.1109/TNNLS.2018.2873649

multitask learning methods usually achieve better performancethan single-task learning methods.

The key challenge of multitask learning is measuring therelationships among different tasks. Existing multitask learn-ing methods can be categorized into two classes based on themethods used to measure the task relationships. In the firstclass of methods, it is assumed that related tasks share a setof common features [7]–[10], while in the second class ofmethods, it is assumed that different tasks share some commonparameters [10]–[14]. Both classes of methods involve theimposition of direct regularizations, on either features orparameters, to learn the relatedness of multiple tasks. However,some of these regularizations are too strong, making theobjective functions difficult to solve.

In this paper, we propose a novel multitask learningalgorithm that utilizes a different measure of task related-ness. Unlike previous methods, in which task relatedness ismeasured using common features or model parameters, ourmethod measures the interdependence among tasks throughthe relatedness of eigenfunctions. The objective function fora particular task, in a manner similar to regression, can beapproximated as a linear combination of the top eigenvalues ofa predefined integral operator on a reproducing kernel Hilbertspace (RKHS) [15]–[17]. In our method, we assume thatrelated tasks share a set of common eigenfunctions and thateach task also has a set of unique eigenfunctions, which aresparse. We formulate our objective function as a linear combi-nation of both sets of eigenfunctions, such that the functionsassociated with one task may provide additional informationto benefit others. Moreover, since the eigenfunctions can beexplicitly computed from the input features with an associatedkernel function, our method can be readily extended to anytype of kernel version. Please note that our method is notsuitable for parametric multitask learning problems [18].

We propose an efficient optimization algorithm for solvingour objective function, which has two regularization terms.One is an L1-norm regularization to guarantee the sparsityof the task-specific eigenfunctions. The other is an L2-normregularization on the shared eigenfunctions to constrain thecomplexity of the trained model. We present a theoreticalanalysis to show that our learning algorithm is uniformlyargument stable, meaning that the output is not sensitive tosubtle changes in the input. In addition, we show that theconvergence rate of the generalization upper bound is relatedto the number of training samples and the number of tasks.This means that when either the training set size or the number

2162-237X © 2018 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.See http://www.ieee.org/publications_standards/publications/rights/index.html for more information.

Page 2: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1819

of tasks increases, the generalization error will decrease. Ourexperimental results obtained on benchmark data sets furthervalidate our proposed approach.

The remainder of this paper is organized as follows.In Section II, we briefly review related works. We presentthe details of our proposed algorithm and the optimizationalgorithm in Section III, and we derive a theoretical analysisto demonstrate the effectiveness of the proposed multitasklearning method in Section IV. We report experimental resultsobtained on several landmark data sets in Section V. Weconclude this paper and discuss future work in Section VI.

II. RELATED WORK

Recent works have demonstrated the success and devel-opment of multitask learning in various domains [1], [7],[19]–[23]. In traditional single-task learning methods, relatedtasks are learned separately, and the positive interactionsamong different tasks are ignored, which lead to a loss ofvaluable information regarding the data distribution. Given thedrawbacks of single-task learning methods, multitask learninghas been proposed to explore the intrinsic relatedness amongdifferent tasks through the joint learning of multiple tasks.Proper measurements of task relatedness can help to gainadditional information on all tasks, particularly when thenumber of training data is insufficient. Additional informationgained from other tasks can help to compensate for the lack oftraining samples. Consequently, multitask learning is appliedwith the goal of improving the performance on all tasks.

Given the successful applications of multitask learning,various traditional single-task learning methods have beenextended to multitask learning algorithms. For example, sup-port vector machine (SVM) [24], as one of the most popularmachine learning algorithms, has been investigated in vari-ous multitask learning studies [11], [25]–[27]. Evgeniou andPontil [11] proposed a classic SVM-based multitask learn-ing framework, which has been referenced by many otherresearchers. In the proposed regularized multitask learningmethod, it is assumed that the hyperplanes of all tasks areclose to one central hyperplane with an offset. Li et al. [25]extended the proximal SVM approach into a multitask learn-ing framework to improve the efficiency of multitask learn-ing. The proposed multitask proximal vector machine modelcan be solved explicitly with high efficiency and compara-ble performance. Jebara [26] proposed learning a commonfeature selection and kernel selection for multitask SVMswith maximum entropy discrimination. Metric learning hasalso been extended to a multitask learning framework. Para-meswaran and Weinberger [1] studied multitask large marginnearest neighbor metric learning and achieved much betterperformance than that achieved with single-task large mar-gin nearest-neighbor metric learning. Ma et al. [2] appliedmultitask distance metric learning for person reidentificationand achieved considerably improved performance. In theirapproach, multiple distance metrics are learned jointly tomeasure the distances of images from different camera pairs.

In recent years, multitask deep learning has been appliedin various research fields. Zhang et al. [28] utilized multitask

deep learning to improve the robustness of facial land-mark detection by simultaneously considering correlated taskssuch as head pose estimation and facial attribute inference.Liu et al. [29] applied multitask deep learning in video thumb-nail selection, using two highly related data sets to explorequery-thumbnail relevance. In the work of Zhang et al. [30],multitask deep convolutional neural networks were utilized toimprove performance in multiview face detection. The con-structed multitask deep neural networks were simultaneouslytrained for face/nonface decisions, face pose estimation, andfacial landmark localization.

Due to the good performance of multitask learning in vari-ous applications, some researchers have attempted to theoret-ically demonstrate the merits of multitask learning [21], [23],[31], [32]. Liu et al. [31] proposed an algorithm-dependentgeneralization bound for multitask learning based on algo-rithmic stability. Subject to a mild assumption regarding thefeature structures, the authors observed that the functionsassociated with other tasks can be viewed as regularizersfor a given task. Li et al. [21] proposed the use of theRKHS of vector-valued functions as a hypothesis space formultitask classification. They derived an improved empiricalRademacher complexity-based generalization bound and dis-cussed the relationship between a group lasso regularizer andthe proposed hypothesis space. An algorithm for multitasklearning from unlabeled data was proposed by Ando andZhang [32]. Their paper presented a general framework forformulating the structural learning problem and analyzed ittheoretically. Maurer et al. [23] applied sparse coding in mul-titask learning and transfer learning. Their paper adopted theassumption that the parameters of tasks can be approximatedwell through a sparse linear combination of the atoms of ahigh-dimensional dictionary, and a generalization error boundfor the proposed approach was given. All these works havepresented valid theoretical analyses of multitask learning.

Multitask learning is based on the assumption that the tasksto be learned are indeed related. However, the method formeasuring the relatedness among different tasks is alwaysa key problem. Common feature representation sharing andcommon parameter sharing are two popular methods of explor-ing the relatedness among multiple tasks. Among methodsbased on feature sharing, Argyriou et al. [7] proposed a convexmultitask feature learning (CMTL) algorithm with L21-normregularization of the parameters. This regularization has theability to ensure the learning of a sparse feature representationshared across different tasks. Zhang and Yeung [33] proposeda convex formulation for multitask learning that can be used toestimate task relationships automatically. Ciliberto et al. [34]proposed a general computational framework for multitasklearning in which a priori knowledge of the task structureis encoded with a convex penalty. In the setting of that paper,some previous proposals could be recovered as special cases.A nonconvex multitask sparse feature learning method wasproposed by Gong et al. [9]. The authors noted the drawbacksof previous convex formulations of multitask feature learningand argued that the proposed method could achieve a betterparameter estimation error bound that could be achieved witha convex formulation.

Page 3: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1820 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

Fig. 1. Framework of our proposed EMTL method. Three different tasks are considered in this figure. We first explicitly learn the top eigenfunctions fromthe input features for each task. Then, all tasks are learned jointly to identify a set of shared eigenfunctions from among those for all tasks. Because of theuniqueness of each task, each task also has another set of sparse basis eigenfunctions that represent its unique characteristics. The final model for each taskis approximated as a combination of the shared eigenfunctions and the particular eigenfunctions in each task.

Considering methods based on parameter sharing,Rai and Daume [14] proposed a nonparametric Bayesianmodel that captures task relatedness under the assumptionthat the task parameters share a latent subspace. In addition,the proposed method can use both labeled and unlabeleddata to assist in learning this subspace, leading to furtherimprovement in performance. Xue et al. [12] proposed anefficient multitask learning algorithm based on a Dirichletprocess-based statistical model. The proposed algorithm canautomatically group similar tasks whose training data mightbe drawn from similar distributions.

In most of these methods, task relatedness is measureddirectly by means of regularizations applied to features fromthe training data or to the model parameters. However, someregularizations are too strict, and the objective functions aredifficult to optimize. For example, in the method proposedin [7], the features are regularized with an L21-norm reg-ularization, which assumes that all tasks share a subset offeatures. This assumption is too strong because it ignores thepossibility that some tasks may have features that are notshared with other tasks. In addition, the objective functionis nonconvex because of the L21-norm regularization. It isdifficult to solve such nonconvex problems directly. Instead,such a problem must be transformed into an equivalent convexoptimization problem for efficient computation. In addition,the extension of such methods to kernel methods is usuallycomplicated, and thus, the resulting methods are difficult toimplement. For example, the method proposed in [33] has aconvex formulation for multitask learning. However, when itis extended to a kernel version, the objective function mustbe changed, and the optimization procedure becomes morecomplex.

III. EIGENFUNCTION-BASED MULTITASK

LEARNING METHOD

In this section, we present the details of our eigenfunction-based multitask learning (EMTL) method. We first introduce

the algorithm for approximating the target function usingeigenfunctions and then describe our EMTL method, followedby an iterative optimization algorithm for optimizing theobjective function.

The framework of our EMTL method is illustrated in Fig. 1,where we consider three different tasks. Each task is associatedwith a set of features that are used to learn the eigenfunctionsfor that task. All the tasks are then learned jointly using theeigenfunctions from all tasks. We assume that all tasks share aset of eigenfunctions and that each task also has a sparse set oftask-specific eigenfunctions that represent its characteristics.The final model for each task can be approximated as acombination of the shared eigenfunctions and its task-specificeigenfunctions.

A. Explicit Eigenfunction Learning

Here, we give a brief introduction to the algorithm forlearning explicit eigenfunctions from features to approximatethe target function, which is mainly inspired by [15] and [16].Suppose that we have a data set of n samples, Dt ={(x1, y1), (x2, y2), . . . , (xn, yn)}, where xi ∈ X denotes the i thinput feature from a compact manifold in the Euclidean spaceRm and yi ∈ Y is the corresponding output in the Euclideanspace R. Let Z = X ×Y , with a Borel probability measure ρ.In addition, let ρX be the marginal probability on X , and letρ(y|x) be the conditional probability of y given x . In thispaper, we mainly focus on the regression problem y = f (x),where f (·) is our target function. Then, the regression functionfρ(x) can be formulated as follows:

fρ(x) =∫

Yydρ(y|x). (1)

Our goal is to approximate an accurate prediction functionfρ(x) using the given training data in an RKHS. Let K (·, ·) :X × X → R be a Mercer kernel, and let HK be an RKHSassociated with the Mercer kernel K (·, ·). An integral operator

Page 4: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1821

L K on HK is defined as follows:

L K ( f ) =∫X

K (·, x) f (x)dρX (x), f ∈ HK . (2)

Let (�i (x), λi ), i = 1, 2, . . . , n, be the eigenfunctions andeigenvalues of L K ranked in descending order of the eigen-values, where the eigenfunctions �i (x), i = 1, 2, . . . , n, forman orthonormal basis of HK . The regression function f (x)can be approximated by a linear combination of the top meigenfunctions with nonzero eigenvalues of L K [15]. Thisfunction can be formulated as follows:

f (x) =m∑

i=1

Ci�i (x) (3)

where m is the number of top eigenfunctions that are usedto approximate the target regression function and can bedetermined empirically. Ci is the coefficient of the i th eigen-function �i (·). The eigenpairs (λi ,�i (·)) can be explicitlyfound from the given features of the training data as follows.Let K: (K (xi , x j ))

ni, j=1 be the Gramian matrix formed by the

kernel K (·, ·) with the training data, and let d ≤ n be therank of the Gramian matrix. The eigenvalues are arranged indescending order as λ1 ≥ · · · ≥ λd ≥ λd+1 = · · · = λn ,and the corresponding eigenvectors are {ui }n

i=1, which forman orthonormal basis of Rn . We thus have

λi = λi

n

�i (·) = 1√λi

n∑j=1

(ui ) j K (·, x j ), for i = 1, . . . , d. (4)

Since the eigenfunctions can be explicitly computed, our goalis to obtain their corresponding coefficients. In Section III-B,we propose our EMTL algorithm, which uses these explicitlycomputed eigenfunctions. The coefficients for all tasks arejointly learned by means of their shared set of eigenfunctions.

B. Eigenfunction-Based Multitask Learning Algorithm

Suppose that we have T different tasks, each of whichis related to a set of data Dt = {(xt1, yt1), (xt2, yt2), . . . ,(xtnt , ytnt )}, where nt is the number of training samplesfor task t . We first compute the eigenpairs {λt i ,�t i (·)} fortask t according to (4), where λt i is the eigenvalue and �t i

is the corresponding eigenfunction. For clarity of notation,we rewrite the eigenpairs for task t as follows. Let dt ≤ nt bethe rank of the Gramian matrix (K (xti , xt j )

nti, j=1). We order

the eigenvalues as λt1 ≥ · · · ≥ λdt ≥ λdt +1 = · · · = λnt = 0,and the associated eigenvectors are {μt i }nt

t i=1. We have

λt i = λt i

nt

�t i(·) = 1√λt i

nt∑j=1

(μt i) j K (·, xt j ), for i = 1, . . . , dt . (5)

Our method measures the relatedness of different tasksthrough the eigenfunctions {�t i(·)}. We assume that some

eigenfunctions are shared among the tasks and that the eigen-functions for one task may benefit the others. To prevent alltasks from being performed similarly due to the influenceof the shared eigenfunctions, our model maintains a set ofnonshared eigenfunctions for each task. The objective of ourEMTL method is formulated as follows:

minCt ,C0

1

T

T∑t=1

1

nt

nt∑i=1

⎛⎝ d∑

j=1

(Ct j + C0 j )� j (xti) − yti

⎞⎠

2

+ γ ‖C‖1 + β‖��C0‖22 (6)

where C = [C1, C2, . . . , CT ] and d = d1 + d2 + · · · + dT ,with the latter denoting the total number of eigenfunctionsidentified from all tasks. The eigenfunctions from all tasks,{�t i(·)}dt

i=1, t = 1, . . . , T , are combined into a single completeset, {� j (·)}d

j=1.To consider the effects of (1/T ) and (1/nt ), we adopt

the notation �(xti) = [�1(xti) × (1/√

T nt ),�2(xti) ×(1/

√T nt ), . . . ,�d (xti) × (1/

√T nt )]� ∈ Rd to repre-

sent a vector of the values of all eigenfunctions giventhe input xti . In addition, Xt = [xt1, xt2, . . . , xtnt ] ∈Rm×nt denotes the data matrix from task t , and �(Xt ) =[�(xt1),�(xt2), . . . ,�(xtnt )] ∈ Rd×nt is a matrix of thevalues of all eigenfunctions given the training data for task t .Let � denote a matrix with entries corresponding to thevalues of all eigenfunctions given the inputs from all tasks,� = [�(X1),�(X2), . . . ,�(XT )]. C0 ∈ Rd is the vector ofthe coefficients of the shared eigenfunctions, and C0 j denotesthe j th entry of the vector C0. Ct ∈ Rd is the coefficient vectorfor the task-specific eigenfunctions of task t . The first term inthe objective function is the loss between the prediction outputand the ground truth. The second term is the regularizationof the coefficients Ct . We constrain Ct using an L1-normregularization, which leads to a sparse set of coefficients.The third term is a Tikhonov regularization of C0 with aTikhonov matrix �, which controls the complexity of themodel. γ and β are two tradeoff parameters, which can bedetermined empirically. If (γ /β) is set at a large value, thenthe coefficient vector Ct will tend toward zero; in this case, alltasks are closely related and tend to share most eigenfunctions,with few or no task-specific eigenfunctions. By contrast, whenthe value (γ /β) approaches zero, we obtain small values of thecoefficients in C0 that correspond to the shared eigenfunctions,in which case, the above-mentioned objectives can be viewedas T separate single-task learning problems that are veryweakly related.

Let {Ct }dt=1 and C0 be the solutions to the above-mentioned

objective functions. Our target regression function for task tcan be written as follows:

ft (xti) =d∑

j=1

(Ct j + C0 j )�t j (xti) (7)

where xti is the i th input for task t and ft (xti ) is the predictedvalue for task t with input xti .

C. Optimization Algorithm

In this section, we present our iterative algorithm foroptimizing the above-mentioned objective function (6) with

Page 5: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1822 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

respect to C0 and {Ct }Tt=1. The details of the algorithm are

given in Algorithm 1.It is difficult to obtain the closed-form solution

({Ct }Tt=1, C0) to the objective function because of the

shared coefficients C0. We therefore iteratively optimizethe objective function with respect to C0 and {Ct }T

t=1.We first optimize the objective function with respect tothe parameter vector C0 by fixing the parameter vectors{Ct }T

t=1. For simplicity and clarity of notation, we introducesome additional variables and rewrite the formulation givenin (6). Recall that �(xti) = [�1(xti) × (1/

√T nt ),�2(xti) ×

(1/√

T nt ), . . . ,�d (xti) × (1/√

T nt )]� ∈ Rd is a vectorof the values of all eigenfunctions given the input xti ,that Xt = [xt1, xt2, . . . , xtnt ] ∈ Rm×nt , and that�(Xt ) = [�(xt1),�(xt2), . . . ,�(xtnt )] ∈ Rd×nt is amatrix of the values of all eigenfunctions given the trainingdata for task t . The optimization with respect to C0 requiresthe training data for all tasks. Therefore, we adopt the notation�(X) = bdiag(�(X1),�(X2), . . . ,�(XT )) ∈ RdT×N , wherebdiag(�(X1),�(X2), . . . ,�(XT )) is a block diagonal matrixwhose diagonal entries correspond to the outputs of alleigenfunctions given the data for task t , that is,

�(X) =

⎛⎜⎜⎜⎜⎜⎜⎝

�(X1)�(X2)

··

·�(XT )

⎞⎟⎟⎟⎟⎟⎟⎠

.

N denotes the total number of training samples for all tasks,as follows:

N = n1 + n2 + · · · + nT .

The output for all tasks is denoted by Y =[Y �

1 , Y �2 , . . . , Y �

T ]� ∈ RN , where Yt = [yt1×(1/√

T nt ), yt3×(1/

√T nt ), . . . , ytnt × (1/

√T nt )]�, considering the

effects of (1/T ) and (1/nt ) in (6). In addition, letC = [C�

1 , C�2 , . . . , C�

T ]� ∈ RdT , let I be the identitymatrix of dimension d , and let I0 = [I, I, . . . , I︸ ︷︷ ︸

T

]� ∈ RdT×d .

We introduce a new variable C0 = I0 × C0. By ignoring theregularization term γ ‖C‖1, the formulation given in (6) canthen be reformulated as follows:

minC0

‖Y − �(X)�(C0 + C)‖22 + β‖��C0‖2

2. (8)

We replace C0 with I0 × C0 and rewrite the above-mentionedobjective as a standard L2-norm regularized regressionproblem

minC0

‖Y − �(X)�(I0 × C0 + C)‖22 + β‖��C0‖2

2. (9)

The solution to this L2-norm regularized problem can beexplicitly obtained as follows:

C0 = (I�0 �(X)�(X)� I0 + ��

)−1

× (I�0 �(X)Y − I�

0 �(X)�(X)�C). (10)

With the explicit solution for the shared coefficients C0obtained by fixing {Ct }T

t=1, we then optimize {Ct }Tt=1 by

fixing C0. The optimization of {Ct }Tt=1 can be separated into T

different tasks when C0 is fixed. For task t , the optimizationproblem can be reformulated with the additional variables asfollows:

minCt

‖�(Xt )�(Ct + C0) − Yt‖2

2 + γ ‖Ct‖1 (11)

which is a standard L1-norm regularized regression problem.Such L1-norm regularized regression problems have beenextensively investigated in the past and can be solved usingvarious methods, such as those presented in [35]–[38]. Thefinal iterative optimization algorithm is given in Algorithm 1.

D. Time Complexity Analysis

We now present an analysis of the computational complexityof Algorithm 1. In Algorithm 1, the computational cost mainlyarises from the optimization of (9) and (11). Note that theeigenpairs in the first step of Algorithm 1 can be computedahead of time and stored for the following optimization steps.Equation (10) is the closed-form solution to (9), and it can bereformulated as follows:

C0 = (I�0 �(X)�(X)� I0 + ��

)−1

× I�0 �(X) × (Y − �(X)�C). (12)

The computation of (I�0 �(X)�(X)� I0 + ��) has a time

complexity of O(NT d2), and the computation of I�0 �(X)

also has a time complexity of O(NT d2). In addition, thecomputation of (Y − �(X)�C) has a time complexityof O(NT d). Considering that the inversion of the matrix(I�

0 �(X)�(X)� I0+��) has a time complexity of O(d3),the final computational time complexity for solving prob-lem (9) is O(NT d2 + d3), which depends on the number oftraining data for all tasks, the number of tasks and the numberof selected top eigenfunctions. For (11), the computationaltime complexity is O(nt d2 + d3) if we solve it using the leastangle regression algorithm [38]. The optimization of (11) mustbe performed for all tasks; therefore, the total time complexityof solving {Ct }T

t=1 is O(Nd2 +d3). Suppose that Algorithm 1runs for M iterations; then, the final time complexity isM × O(NT d2 + d3). From this time complexity analysis,we can conclude that the time complexity of our proposedalgorithm is independent of the dimensionality of the originaldata. We can control the time complexity by varying thenumber of selected top eigenfunctions. Consequently, ourproposed method is more suitable than other methods forhigh-dimensional multitask learning problems.

IV. THEORETICAL ANALYSIS

In this section, we present a theoretical analysis to demon-strate how the proposed method can better learn sharedinformation. Since C�

0 � represents the commonly sharedparameters and C�

t � represents the specific parameters forthe t th task, we will focus on analyzing the learning propertiesfor C�

0 �. Specifically, we show that the proposed method is

Page 6: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1823

Algorithm 1 Iterative Optimization Algorithm for EMTL

Input: Input data sets Dt , t = 1, 2, . . . , T ; initialize the variables C0 and {Ct }Tt=1 and the trade-off parameters γ and β.

Output: Shared coefficients C0 and coefficients for each specific task {Ct }Tt=1.

1: Explicitly compute the eigenpairs {λt i ,�t i }dti=1 for all tasks using formulation (5)

2: while (not converged) do3: Compute C0 = argmin‖Y − �(X)�(I0 × C0 + C)‖2

2 + β‖��C0‖22.

4: for t =1 to T do5: Compute Ct = argmin ‖�(Xt )

T (Ct + C0) − Yt‖22 + γ ‖Ct‖1.

6: end for7: end while

argument stable [39] for learning C�0 � and that the gener-

alization bound for learning C�0 � has a convergence rate of

O(1/√

nT ), which enables the proposed learning algorithm togeneralize quickly and accurately from a small training samplewhen the number of tasks is large.

We first introduce the notion of argument stability [39],which measures the impact of changing a single trainingexample on the function selected by the learning algorithm.Intuitively, the learning algorithm is stable if its outputs arenot sensitive to subtle changes in the input, or in other words,if the outputs do not change much when the changes in theinput training samples are small.

Definition 1 (Uniform Argument Stability [39]): Let D ={(X1, Y1), . . . , (Xn, Yn)} be a training set consisting of n pairsof independent random variables. Let C�

0,D�D denote theoutput of a learning algorithm obtained by exploiting theinput training set D. We say that the learning algorithm isα(n)-uniformly argument stable if for all i ∈ {1, . . . , n},it holds that ∥∥C�

0,D�D − C�0,Di �Di

∥∥ ≤ α(n) (13)

where α(n) ∈ R+ and Di = {(X1, Y1), . . . , (Xi−1, Yi−1),(X ′

i , Y ′i ), (Xi+1, Yi+1), . . . , (Xn, Yn)} represents the training

set D with the i th sample replaced with an independent copyof (X ′

i , Y ′i ).

We show that the algorithm for learning the commonlyshared parameter C0 in (6) is uniformly argument stable.

Theorem 1: If we assume that the variables ‖C‖2, ‖�(x)‖2,K (x, x), and Y are upper bounded by ∧C , ∧φ , ∧2

K , and ∧Y ,respectively, then the algorithm for learning C0 in (6) isuniformly argument stable, that is,∥∥C�

0,D�D − C�0,Di �Di

∥∥

≤ 2(2 ∧C ∧� + ∧Y )∧K

β min{n1, . . . , nT }T+

√4(2 ∧C ∧� + ∧Y ) ∧C ∧�

β min{n1, . . . , nT }T.

(14)

To prove Theorem 1, we first introduce the notion of theBregman divergence.

Definition 2 (Bregman Divergence): Let f be a convexfunction. For any s and t in its domain, the Bregman diver-gence is defined as

B f (s‖t) = f (s) − f (t) − 〈s − t,∇ f (t)〉 (15)

where ∇ f (t) denotes the gradient of f at t .

It is easy to prove that the Bregman divergence is additiveand nonnegative. For example, if f = f1 + f2 and bothf1 and f2 are convex, then for any s and t in the domain,we have

B f (s‖t) = B f1(s‖t) + B f2(s‖t) (16)

and

B f (s‖t) ≥ 0. (17)

We are now ready to prove Theorem 1. Let

L D(C�

0 �)

= 1

T

T∑t=1

1

nt

nt∑i=1

⎛⎝ d∑

j=1

(Ct j + C0 j )� j (Xti) − Yti

⎞⎠

2

(18)

and

PD(C�

0 �) = β

∥∥C�0 �

∥∥22 + γ ‖C‖1. (19)

The objective in (6) can be written as

OD(C�

0 �) = L D

(C�

0 �) + PD

(C�

0 �). (20)

Note that OD and PD are both convex with respect to C�0 �.

Using the nonnegative and additive properties of the Bregmandivergence, we have

BOD

(C�

0,Di �Di

∥∥C�0,D�D

) + BODi

(C�

0,D�D∥∥C�

0,Di �Di

)≥ BPD

(C�

0,Di �Di

∥∥C�0,D�D

) + BPDi

(C�

0,D�D∥∥C�

0,Di �Di

).

We attempt to upper bound BOD(C�0,Di �Di ‖C�

0,D�D) +BODi (C

�0,D�D‖C�

0,Di �Di ) and lower bound

BPD(C�0,Di �Di ‖C�

0,D�D) + BPDi (C�0,D�D‖C�

0,Di �Di ).

Specifically, let P2D(C�0 �) = β‖C�

0 �‖22; then, we have

BPD

(C�

0,Di �Di

∥∥C�0,D�D

) + BPDi

(C�

0,D�D∥∥C�

0,Di �Di

)≥ BP2D

(C�

0,Di �Di

∥∥C�0,D�D

) + BP2Di

(C�

0,D�D∥∥C�

0,Di �Di

)= β

∥∥C�0,Di �Di

∥∥22 − β

∥∥C�0,D�D

∥∥22

− ⟨C�

0,Di �Di − C�0,D�D, 2βC�

0,D�D⟩ + β

∥∥C�0,D�D

∥∥22

− β∥∥C�

0,Di �Di

∥∥22 − ⟨

C�0,D�D − C�

0,Di �Di , 2βC�0,Di �Di

⟩= 2β

∥∥C�0,Di �Di − C�

0,D�D∥∥2

2.

Page 7: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1824 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

We further upper bound BOD (C�0,Di �Di ‖C�

0,D�D) +BODi (C

�0,D�D‖C�

0,Di �Di )

BOD

(C�

0,Di �Di

∥∥C�0,D�D

)+ BODi

(C�

0,D�D∥∥C�

0,Di �Di

)= OD

(C�

0,Di �Di

) − OD(C�

0,D�D)

− ⟨C�

0,Di �Di − C�0,D�D,∇OD

(C�

0,D�D)⟩

+ ODi

(C�

0,D�D) − ODi

(C�

0,Di �Di

)−⟨

C�0,D�D − C�

0,Di �Di ,∇ODi

(C�

0,Di �Di

)⟩= OD

(C�

0,Di �Di

) − OD(C�

0,D�D) + ODi

(C�

0,D�D)

− ODi

(C�

0,Di �Di

). (21)

The second equality holds because C�0,D�D and C�

0,Di �Di

are the minimizers of OD(C�0,D�D) and ODi (C�

0,Di �Di ),

respectively, and ∇OD(C�0,D�D) = ∇ODi (C�

0,Di �Di ) = 0.Thus

BOD

(C�

0,Di �Di

∥∥C�0,D�D

)+ BODi

(C�

0,D�D∥∥C�

0,Di �Di

)= ODi

(C�

0,D�D) − OD

(C�

0,D�D) + OD

(C�

0,Di �Di

)− ODi

(C�

0,Di �Di

)

= 1

T nt

⎛⎝ d∑

j=1

(Ct j + C0 j,D)� j,D(X ′

t i

) − Y ′t i

⎞⎠

2

− 1

T nt

⎛⎝ d∑

j=1

(Ct j + C0 j,D)� j,D(Xti ) − Yti

⎞⎠

2

+ 1

T nt

⎛⎝ d∑

j=1

(Ct j + C0 j,Di )�′j,Di (Xti ) − Yti

⎞⎠

2

− 1

T nt

⎛⎝ d∑

j=1

(Ct j + C0 j,Di )�′j,Di

(X ′

t i

) − Y ′t i

⎞⎠

2

≤ 2(2 ∧C ∧� + ∧Y )

T nt

∣∣∣∣d∑

j=1

(Ct j + C0 j,D)� j,D(X ′

t i

)

−d∑

j=1

(Ct j + C0 j,Di )�′j,Di

(X ′

t i

)∣∣∣∣ + 2(2 ∧C ∧� + ∧Y )

T nt

×∣∣∣∣∣∣

d∑j=1

(Ct j + C0 j,D)� j,D(Xti )

−d∑

j=1

(Ct j + C0 j,Di )�′j,Di (Xti )

∣∣∣∣∣∣≤ 4(2 ∧C ∧� + ∧Y )∧K

T nt

×∥∥∥∥∥∥

d∑j=1

(C0 j,D�D − C0 j,Di �Di )

∥∥∥∥∥∥2

+8(2 ∧C ∧� + ∧Y ) ∧C ∧�

T nt

≤ 4(2 ∧C ∧� + ∧Y )∧K

T nt

∥∥C�0,D� j,D − C�

0,Di � j,Di

∥∥2

+ 8(2 ∧C ∧� + ∧Y ) ∧C ∧�

T nt. (22)

Combining (21) and (22), we obtain

2β∥∥C�

0,D� j,D − C�0,Di � j,Di

∥∥22

≤ 4(2 ∧C ∧� + ∧Y )∧K

T nt

∥∥C�0,D� j,D − C�

0,Di � j,Di

∥∥2

+ 8(2 ∧C ∧� + ∧Y ) ∧C ∧�

T nt. (23)

We then have∥∥C�

0,D� j,D − C�0,Di � j,Di

∥∥2

≤ 2(2 ∧C ∧� + ∧Y )∧K

βT nt+

√4(2 ∧C ∧� + ∧Y ) ∧C ∧�

βT nt.

(24)

�Theorem 1 implies that when the training set is changed by

one example, the change in the output C�0 � will vanish as the

training set size n or the number of tasks T goes to infinity.This is the property of algorithmic stability, which can beused to derive the generalization bound [40]. By employingthe result of Liu et al. [39] (Theorem 2 therein), we caneasily derive a deformed generalization bound for the pro-posed algorithm with respect to the parameter C�

0,D�D . Thisdeformed generalization upper bound will have a convergencerate of O(1/

√nT ) with respect to the training set size n and

the number of tasks T , which implies that with an increasein either the training set size n or the number of tasks T ,the generalization error will decrease. Specifically, in theproof of Theorem 1, we can see that the convergence rateof O(1/

√nT ) is introduced because of Ct . If Ct = 0, then

the generalization bound for learning the commonly sharedparameter will converge faster, with a rate of O(1/nT ). Theadvantage of multitask learning has thus been demonstratedfor learning C�

0,D�D . The empirical validations presented inSection V also support these theoretical results.

The generalization error measures the difference betweenthe training and testing errors. A small generalization errorbound does not imply a small test error. A small testingerror should additionally be based on a small training error.The choice of � in this paper also essentially guaranteesa small training error in (6) because it guarantees a smallreconstruction error in the feature space. Then, (7) functionssimilar to a representer theorem but with a clear structure ofcommonly shared parameters in the multitask learning setting.

V. EXPERIMENTS

In this section, we present and analyze experimental resultsobtained on three benchmark multitask learning data setsto demonstrate the effectiveness of our proposed multitask

Page 8: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1825

TABLE I

COMPARISON OF RESULTS OBTAINED ON THE SCHOOL DATA SET USING THE MSE

TABLE II

COMPARISON OF RUNNING TIMES (SECONDS) ON THE SCHOOL DATA SET

learning algorithm. The three data sets used in our experi-ments are the School data set,1 the Computer data set [41],and the Isolet data set.2 These three data sets have beenwidely used for evaluating the effectiveness of multitasklearning in various works [1], [7], [9], [11], [25]. The exper-imental results of our proposed EMTL method are com-pared with those of three single-task learning algorithms andseveral state-of-the-art multitask learning methods. The firsttwo single-task learning methods are kernel ridge regres-sion (KRR) and a single-task learning method based onexplicitly learned eigenfunctions feature-based single tasklearning (FSTL) [15]. The third single-task learning methodis FSTL_multiple (FSTL_M), which is based on FSTL butconsiders the eigenfunctions learned from all tasks. Themultitask learning methods are CMTL [7], multitask rela-tionship learning (MTRL) [33], and the convex learning ofmultiple tasks and their structure (CMTS) [34]. These mul-titask learning methods are representative methods that haveachieved promising performance on various multitask learningdata sets. Consequently, comparisons with these methods cansufficiently demonstrate the effectiveness of our proposedmethod.

A. School Data Set

The School data set is one of the most widely used multitasklearning data sets. It was collected from the Inner LondonEducation Authority. This data set consists of 139 tasks,each of which corresponds to the prediction of examinationscores at one secondary school. The provided data includethe examination scores of 15 362 students from 139 secondaryschools in 1985, 1986, and 1987. Each sample includes fourschool-dependent features, three student-dependent features,and the year of the examination. The four school-dependentfeatures are the percentage of students eligible for free schoolmeals, the percentage of students in voltage regulator (VR)band one, the school denomination, and the school’s gendercomposition. The three student-dependent features are gender,ethnic group, and VR band. To ensure fair comparisonswith the other methods, we considered 27-dimensional binaryvariables for each sample, following the same setup as inprevious multitask learning studies [7], [33], [42].

1http://ttic.uchicago.edu/∼argyriou/code/2https://archive.ics.uci.edu/ml/datasets/ISOLET

To evaluate the effectiveness of our proposed multitasklearning method, we randomly selected 10%, 20%, or 30%of the data for each task as the training data. The remainingsamples were split into the validation set and the test set.To avoid statistical outliers, we repeated this selection process10× for all methods, and we reported the mean performanceand the standard deviation across the 10 trials. All meth-ods utilized an radial basis function (RBF) kernel, and thebest parameters for different tasks were selected based onthe validation set. The number of top eigenfunctions wasalso empirically selected based on the validation set. If allnonzero eigenvalues and their corresponding eigenvectors wereused, the best performance would be achieved. However, thisapproach would also increase the computation time of ourproposed algorithm. We therefore attempted to reduce thenumber of eigenfunctions used for each task in our proposedmethod while guaranteeing its performance. For the Schooldata set, the top 10 eigenvectors were used in our method.We evaluated the performances of all regression methods usingmean squared error (MSE). The results are shown in Table I.

From the results shown in Table I, we can concludethat all multitask learning methods outperformed two ofthe single-task learning methods, further demonstrating theeffectiveness of multitask learning compared with single-tasklearning. Notably, our proposed EMTL algorithm consis-tently performed the best as the training ratio increased from10% to 30%. The FSTL_M method, which used the eigenfunc-tions from all tasks without considering how the eigenfunc-tions were shared among tasks, performed the worst. This poorperformance might be caused by the introduction of noise fromother tasks. We can also conclude that the proposed methodeffectively measures the relatedness among tasks, particularlywhen the number of training samples is insufficient. In single-task learning, sufficient information about the distribution ofthe training data cannot be obtained when limited data areprovided. By contrast, our proposed method extracts moreinformation by considering the relatedness among differenttasks. In addition, our proposed method achieves better per-formance using 10% of the training data than other methodsachieve using 30% of the training data.

To better illustrate the computational efficiencies, we com-pared the running times of all methods on a PC witha 4.0-GHz Intel Core CPU and 16 GB of memory. The results

Page 9: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1826 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

TABLE III

COMPARISON OF RESULTS OBTAINED ON THECOMPUTER DATA SET USING THE MSE

are shown in Table II. Note that the running time is thetotal time required to solve all tasks and that the experimentswere repeated 10×. The mean time and standard deviation arereported. We can conclude that multitask learning algorithmsrequire much more time than single-task learning algorithms.Our proposed EMTL method has a much lower time cost thanCMTL and MTRL. CMTS is the most efficient method. Thisresult is mainly because CMTS has closed-form solutions foreach step. However, the cost time of our EMTL method iscomparable to that of CMTS as the number of training samplesincreases, and our proposed EMTL method achieves muchbetter performance.

B. Computer Data Set

In this section, we report on the experiments conductedon the Computer data set, which contains people’s ratings ofcomputer products [41]. This data set includes the results of asurvey of 180 people who rated their tendency to buy 20 differ-ent computers. Each person is regarded as one task, followingthe same experimental setup as in previous works. Each com-puter is represented by a 13-dimensional binary feature vector,which includes telephone hotline availability, amount of mem-ory, screen size, CPU speed, hard disk, CD-ROM/multimedia,cache, color, availability, warranty, software, guarantee, andprice. The output is an integer rating that scales from 0 to 10.To facilitate computation, one dimension representing the biasterm was added. Following the same setup used in previousworks, the first eight examples from each person were usedas training data, and the last four examples were used as testdata. We chose the top eight eigenfunctions from each task toapproximate the final regression function. We used the sameevaluation measurement, the MSE, as used on the School dataset to evaluate the performance. All methods used an RBFkernel, and the best parameters were selected based on thevalidation set.

The results are reported in Table III, based on which weconclude that the multitask learning methods outperform thesingle-task learning methods. Our proposed EMTL methodand CMTL exhibit the best performance, with similar MSEvalues. In addition, we show the running times of all methodson the Computer data set in Table IV, and conclusions similarto those obtained on the School data set can be drawn.

We also illustrate the learned shared coefficients C0 andtask-specific coefficients C = [C1, C2, . . . , CT ]. For theComputer data set, the training data for each task are thesame. Consequently, the explicitly learned eigenfunctionsfor each task are also the same. This is the reason whyFSTL and FSTL_M achieve the same MSE and the same

TABLE IV

COMPARISON OF RUNNING TIMES (SECONDS)ON THE COMPUTER DATA SET

Fig. 2. Illustration of the absolute values of the eight shared coefficients (C0)learned from the Computer data set. Black areas denote zero values, and thevalue increases as the color changes from black to white. Only the coefficientof the 6th eigenfunction is close to zero, which means that all tasks shareseven eigenfunctions, leading to high relatedness.

Fig. 3. Illustration of the absolute values of the nonshared coefficients forindividual tasks learned from the Computer data set. The coefficients of thetask-specific eigenfunctions are sparse.

computation time. We only have to learn the coefficients for alltasks from the selected top eight eigenfunctions. The absolutevalues of the learned coefficients are shown in Figs. 2 and 3.

Page 10: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1827

TABLE V

COMPARISON OF RESULTS OBTAINED ON THE ISOLET DATA SET USING THE MSE

TABLE VI

COMPARISON OF RUNNING TIMES (SECONDS) ON THE ISOLET DATA SET

TABLE VII

COMPARISON OF RESULTS OBTAINED ON FIVE ISOLET TASKS USING THE MSE

Black areas denote zero values, and the value increases asthe color changes from black to white. From Fig. 2, we findthat the coefficient of the 6th eigenfunction is close to zero,which means that seven of the eight eigenfunctions are shared.From Fig. 3, we find that the coefficients of the task-specificeigenfunctions are sparse. We can conclude that the differenttasks in the Computer data set are closely related. The reasonthe coefficient value for the second eigenfunction appears quitehigh compared to the rest of the values may be that the secondeigenfunction is important to all tasks. This can also be seenin the results in Fig. 3. The task-specific coefficient for thesecond eigenfunction appears to be large for almost all tasks.Therefore, we can conclude that the second eigenfunction isthe most important of the eigenfunctions to most of the tasksand should be shared among the tasks with a large coefficient.

C. Isolet Data Set

We report the results of testing the performance of ourproposed multitask learning method on the Isolet data set inthis section. This data set concerns the pronunciation of theletters in the alphabet by 150 speakers. Each speaker spokeeach letter twice; thus, 52 examples were collected from eachspeaker. The speakers are grouped into five groups: Isolet-1,Isolet-2, Isolet-3, Isolet-4, and Isolet-5. Thus, there are fivetasks, one corresponding to each of these five groups, whichcontain 1560, 1560, 1560, 1558, and 1559 samples, respec-tively. Each letter is related to a label (1–26), and we treatall tasks as regression problems, following [9]. We randomlyselected 10%, 20%, or 30% of the data as the training set, andthe rest of the data was split into the validation set and thetest set. To avoid statistical outliers in the experimental results,we repeated all experiments five times, and we reported themean performance with the standard deviation. An RBF kernel

was used in all methods, and the best parameters were selectedbased on the validation set. We again used the MSE to evaluatethe performance of each method.

Based on the results shown in Table V, we can againconclude that all of the multitask learning methods exceptCMTL outperform the single-task learning methods. Thisresult is because considering the sharing of common featuresacross tasks is not an adequate means of measuring therelatedness among tasks in the Isolet data set. In addition,the performance of FSTL_M is similar to that of FSTL. Thisfinding indicates that FSTL_M cannot learn additional infor-mation about the data from the eigenfunctions of other taskswithout considering the shared eigenfunctions. Our proposedEMTL method consistently achieved the best performanceacross the different training ratios. These findings demonstratethat measuring task relatedness through eigenfunctions enablesbetter exploration of the information contained in this dataset that can be achieved with the other methods. In addition,we present the running times of all methods in Table VI, usingthe same settings as for the School data set. We can againobtain conclusions similar to those found in the School dataset and the Computer data set.

In Table VII, we present additional experimental results toenable an analysis of the performance improvement on eachtask when our proposed method is used. This experiment wasconducted with 30% of the data as the training set and theremaining data split into the validation set and the test set.All experiments were repeated five times to avoid statisticaloutliers, and the best parameters were selected based on thevalidation set. Based on the performance of the single-tasklearning methods, we can conclude that the difficulty variesamong the different tasks and that task 4 is the most difficultone. Compared with the single-task learning methods, allof the multitask learning methods except CMTL showed

Page 11: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1828 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

Fig. 4. Sensitivity analysis of EMTL with respect to the parameter γ .

Fig. 5. Sensitivity analysis of EMTL with respect to the parameter β.

improved performance on all tasks. For the more difficulttasks, limited information about the distribution of the data wasavailable from their training data. However, multitask learningmethods can extract more shared information relevant to thesetasks, leading to performance improvements. Our proposedmultitask learning method effectively measures the relatednessamong tasks and significantly improves the performance on alltasks, particularly difficult ones.

D. Sensitivity Analysis of EMTL

In this section, we report experiments conducted to analyzethe sensitivity of our proposed EMTL method to the regular-ization parameters β and γ . All experiments were conductedon the School data set with a training ratio of 10%.

All parameters, including β and γ , were selected to achievethe best performance on the validation set for all experimentsin this paper. We conducted a grid search of β and γ withinthe set {10−3, 10−2, 10−1, 1, 10, 102, 103}. For the Schooldata set, we used an RBF kernel with a bandwidth of 10.As seen from experiments on the validation set, the bestperformance was achieved with parameter values of γ = 100and β = 1. Consequently, we analyzed the sensitivity ofEMTL to γ with a fixed value of β = 1 and analyzed thesensitivity of EMTL to β with a fixed value of γ = 100. Theresults are shown in Figs. 4 and 5. From these results, we can

TABLE VIII

COMPARISONS BETWEEN OUR PROPOSED METHOD ANDTHE SECOND BEST METHOD IN TERMS OF P -VALUES

conclude that the performances in the experiments were betterfor γ values larger than 10 and β values near 1. These findingsindicate that the coefficients {Ct }T

t=1 of the task-specific eigen-functions tend to be smaller than the coefficients C0 of theshared eigenfunctions. Therefore, the shared eigenfunctionsplay a more important role, and for each task, additionalvaluable information can be obtained from the training dataassociated with other tasks. The performances on all tasksshould improve in such a situation.

E. Analysis of P-Values

In this section, we present an analysis of p-values obtainedusing the t-test to show that our proposed method is statisti-cally significantly better than the next best method. We per-formed t-tests only on the School data set and the Isoletdata set because the training and test samples in the Computerdata set are fixed.

From Table I, we can see that on the School data set,MTRL performs the second best when the training ratio is 10%or 20% and that CMTS performs the second-best when thetraining ratio is 30%. We therefore compare our EMTL methodwith MTRL for training ratios of 10% and 20% and withCMTS for a training ratio of 30%. Similarly, we compare ourEMTL method with MTRL for training ratios of 10% and 20%and with CMTS for a training ratio of 30% on the Isolet dataset. The results are shown in Table VIII. We can conclude thatour proposed method performs significantly better than thesecond-best methods, as the p-values are substantially smallerthan 0.05 for all training ratios on both data sets.

VI. CONCLUSION

In this paper, we propose a method for learning multipletasks from a new perspective. Unlike previous multitask learn-ing methods, in which task relatedness is measured throughparameter sharing or feature sharing, our proposed multitasklearning method learns task relationships by considering ashared set of eigenfunctions. These eigenfunctions can beexplicitly learned and easily extended to any kernel type.Consequently, we only have to learn a set of shared coefficientsfor all tasks and a set of task-specific coefficients for eachtask. The objective function can be optimized by means of aniterative algorithm, which divides the optimization probleminto two subproblems: L2-norm regularized regression andL1-norm regularized regression. We also present a detailedtheoretical analysis to demonstrate that our proposed algorithmis uniformly argument stable and that the convergence rateof the generalization upper bound is related to the numberof training samples and the number of tasks. The findings

Page 12: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

TIAN et al.: EMTL IN AN RKHS 1829

imply that learning multiple tasks simultaneously can helpimprove performance. Various experiments were conductedon several multitask learning data sets, and the experimentalresults demonstrate the effectiveness of our proposed method.

REFERENCES

[1] S. Parameswaran and K. Q. Weinberger, “Large margin multi-task metriclearning,” in Proc. Adv. Neural Inf. Process. Syst., 2010, pp. 1867–1875.

[2] L. Ma, X. Yang, and D. Tao, “Person re-identification over cameranetworks using multi-task distance metric learning,” IEEE Trans. ImageProcess., vol. 23, no. 8, pp. 3656–3670, Aug. 2014.

[3] Y. Luo, Y. Wen, and D. Tao, “Heterogeneous multitask metric learningacross multiple domains,” IEEE Trans. Neural Netw. Learn. Syst.,vol. 29, no. 1, pp. 154–167, Sep. 2018.

[4] X. Wang, C. Zhang, and Z. Zhang, “Boosted multi-task learning for faceverification with applications to Web image and video search,” in Proc.Comput. Vis. Pattern Recognit., 2009, pp. 142–149.

[5] D. Zhang and D. Shen, “Multi-modal multi-task learning for joint pre-diction of multiple regression and classification variables in Alzheimer’sdisease,” NeuroImage, vol. 59, no. 2, pp. 895–907, 2012.

[6] L. Nie, L. Zhang, L. Meng, X. Song, X. Chang, and X. Li, “Modelingdisease progression via multisource multitask learners: A case study withAlzheimer’s disease,” IEEE Trans. Neural Netw. Learn. Syst., vol. 28,no. 7, pp. 1508–1519, Jul. 2017.

[7] A. Argyriou, T. Evgeniou, and M. Pontil, “Convex multi-task featurelearning,” Mach. Learn., vol. 73, no. 3, pp. 243–272, 2008.

[8] G. Obozinski, B. Taskar, and M. I. Jordan, “Joint covariate selectionand joint subspace selection for multiple classification problems,” Statist.Comput., vol. 20, no. 2, pp. 231–252, Apr. 2010.

[9] P. Gong, J. Ye, and C.-S. Zhang, “Multi-stage multi-task feature learn-ing,” in Proc. Adv. Neural Inf. Process. Syst., 2012, pp. 1988–1996.

[10] J. Liu, S. Ji, and J. Ye, “Multi-task feature learning via efficient 2,1-norm minimization,” in Proc. 25th Conf. Uncertainty Artif. Intell., 2009,pp. 339–348.

[11] T. Evgeniou and M. Pontil, “Regularized multi–task learning,” in Proc.10th ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining, 2004,pp. 109–117.

[12] Y. Xue, X. Liao, L. Carin, and B. Krishnapuram, “Multi-task learningfor classification with Dirichlet process priors,” J. Mach. Learn. Res.,vol. 8, pp. 35–63, May 2007.

[13] K. Yu, V. Tresp, and A. Schwaighofer, “Learning Gaussian processesfrom multiple tasks,” in Proc. 22nd Int. Conf. Mach. Learn., 2005,pp. 1012–1019.

[14] P. Rai and H. Daumé, III, “Infinite predictor subspace models formultitask learning,” in Proc. AISTATS, 2010, pp. 613–620.

[15] X. Guo and D.-X. Zhou, “An empirical feature-based learning algo-rithm producing sparse approximations,” Appl. Comput. Harmon. Anal.,vol. 32, no. 3, pp. 389–400, 2012.

[16] L. Zwald, G. Blanchard, P. Massart, and R. Vert, “Kernel projectionmachine: A new tool for pattern recognition,” in Proc. Adv. Neural Inf.Process. Syst., 2005, pp. 1649–1656.

[17] M. Ji, T. Yang, B. Lin, R. Jin, and J. Han. (2012). “A simple algorithmfor semi-supervised learning with improved generalization error bound.”[Online]. Available: https://arxiv.org/abs/1206.6412

[18] I. Takeuchi, T. Hongo, M. Sugiyama, and S. Nakajima, “Parametric tasklearning,” in Proc. Adv. Neural Inf. Process. Syst., 2013, pp. 1358–1366.

[19] X. Mei, Z. Hong, D. Prokhorov, and D. Tao, “Robust multitask multiviewtracking in videos,” IEEE Trans. Neural Netw. Learn. Syst., vol. 26,no. 11, pp. 2874–2890, Nov. 2015.

[20] X. Chang and Y. Yang, “Semisupervised feature analysis by miningcorrelations among multiple tasks,” IEEE Trans. Neural Netw. Learn.Syst., vol. 28, no. 10, pp. 2294–2305, Oct. 2017.

[21] C. Li, M. Georgiopoulos, and G. C. Anagnostopoulos, “Multitaskclassification hypothesis space with improved generalization bounds,”IEEE Trans. Neural Netw. Learn. Syst., vol. 26, no. 7, pp. 1468–1479,Jul. 2015.

[22] Y. Kong, M. Shao, Y. Fu, and K. Li, “Probabilistic low-rank multitasklearning,” IEEE Trans. Neural Netw. Learn. Syst., vol. 29, no. 1,pp. 670–680, Mar. 2018.

[23] A. Maurer, M. Pontil, and B. Romera-Paredes, “Sparse coding for mul-titask and transfer learning,” Proc. 30th Int. Conf. Mach. Learn. (ICML),2013, pp. 343–351.

[24] C. Cortes and V. Vapnik, “Support-vector networks,” Mach. Learn.,vol. 20, no. 3, pp. 273–297, 1995.

[25] Y. Li, X. Tian, M. Song, and D. Tao, “Multi-task proximal support vectormachine,” Pattern Recognit., vol. 48, no. 10, pp. 3249–3257, 2015.

[26] T. Jebara, “Multi-task feature and kernel selection for SVMs,” in Proc.21st Int. Conf. Mach. Learn., 2004, pp. 55–63.

[27] J. Tang, Y. Tian, P. Zhang, and X. Liu, “Multiview privileged supportvector machines,” IEEE Trans. Neural Netw. Learn. Syst., vol. 29, no. 8,pp. 3463–3477, Aug. 2018.

[28] Z. Zhang, P. Luo, C. C. Loy, and X. Tang, “Facial landmark detectionby deep multi-task learning,” in Proc. Eur. Conf. Comput. Vis., 2014,pp. 94–108.

[29] W. Liu, T. Mei, Y. Zhang, C. Che, and J. Luo, “Multi-task deepvisual-semantic embedding for video thumbnail selection,” in Proc.IEEE Conf. Comput. Vis. Pattern Recognit., Jun. 2015, pp. 3707–3715.

[30] C. Zhang and Z. Zhang, “Improving multiview face detection withmulti-task deep convolutional neural networks,” in Proc. IEEE WinterConf. Appl. Comput. Vis. (WACV), Mar. 2014, pp. 1036–1041.

[31] T. Liu, D. Tao, M. Song, and S. J. Maybank, “Algorithm-dependentgeneralization bounds for multi-task learning,” IEEE Trans. PatternAnal. Mach. Intell., vol. 39, no. 2, pp. 227–241, Feb. 2017.

[32] R. K. Ando and T. Zhang, “A framework for learning predictivestructures from multiple tasks and unlabeled data,” J. Mach. Learn. Res.,vol. 6, pp. 1817–1853, Nov. 2005.

[33] Y. Zhang and D.-Y. Yeung, “A convex formulation for learning taskrelationships in multi-task learning,” in Proc. 26th Conf. UncertaintyArtif. Intell. (UAI), 2010, pp. 733–742.

[34] C. Ciliberto, Y. Mroueh, T. Poggio, and L. Rosasco, “Convex learningof multiple tasks and their structure,” in Proc. Int. Conf. Mach. Learn.,2015, pp. 1548–1557.

[35] T. Goldstein and S. Osher, “The split Bregman method forL1-regularized problems,” SIAM J. Imag. Sci., vol. 2, no. 2, pp. 323–343,2009.

[36] M. Schmidt, G. Fung, and R. Rosales, “Optimization methods for1-regularization,” Univ. Brit. Columbia, Vancouver, BC, Canada,Tech. Rep. TR-2009, 2009, vol. 19.

[37] S. Boyd, N. Parikh, E. Chu, B. Peleato, and J. Eckstein, “Distributedoptimization and statistical learning via the alternating direction methodof multipliers,” Found. Trends Mach. Learn., vol. 3, no. 1, pp. 1–122,Jan. 2011.

[38] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, “Least angleregression,” Ann. Statist., vol. 32, no. 2, pp. 407–499, 2004.

[39] T. Liu, G. Lugosi, G. Neu, and D. Tao, “Algorithmic stability andhypothesis complexity,” in Proc. Int. Conf. Mach. Learn., vol. 70, 2017,pp. 2159–2167.

[40] S. Shalev-Shwartz, O. Shamir, N. Srebro, and K. Sridharan, “Learnabil-ity, stability and uniform convergence,” J. Mach. Learn. Res., vol. 11,pp. 2635–2670, Oct. 2010.

[41] P. J. Lenk, W. S. DeSarbo, P. E. Green, and M. R. Young, “Hier-archical bayes conjoint analysis: Recovery of partworth heterogeneityfrom reduced experimental designs,” Marketing Sci., vol. 15, no. 2,pp. 173–191, 1996.

[42] T. Evgeniou, C. A. Micchelli, and M. Pontil, “Learning multiple taskswith kernel methods,” J. Mach. Learn. Res., vol. 6, pp. 615–637,Apr. 2005.

Xinmei Tian (M’13) received the B.E. and Ph.D.degrees from the University of Science and Tech-nology of China, Hefei, China, in 2005 and 2010,respectively.

She is currently an Associate Professor with theCAS Key Laboratory of Technology in Geo-SpatialInformation Processing and Application System,University of Science and Technology of China.Her current research interests include multimediainformation retrieval and machine learning.

Dr. Tian was a recipient of the Excellent DoctoralDissertation of Chinese Academy of Sciences Award in 2012 and theNomination of National Excellent Doctoral Dissertation Award in 2013.

Page 13: Eigenfunction-Based Multitask Learning in a Reproducing ...staff.ustc.edu.cn/~xinmei/publications_pdf/2019...based on feature sharing, Argyriouet al. [7] proposed a convex multitask

1830 IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, VOL. 30, NO. 6, JUNE 2019

Ya Li received the B.S. and Ph.D. degrees fromthe Department of Electronic Engineering and Infor-mation Science, University of Science and Tech-nology of China, Hefei, China, in 2013 and 2018,respectively.

His current research interests include machinelearning and computer vision.

Tongliang Liu received the B.Eng. degree fromthe University of Science and Technology of China,Hefei, China, and the Ph.D. degree from the Univer-sity of Technology Sydney, Ultimo, NSW, Australia.

He is currently a Lecturer with the School ofInformation Technologies, Faculty of Engineeringand Information Technologies, The University ofSydney, Darlington, NSW, where he is a CoreMember with the UBTECH Sydney Artificial Intelli-gence Centre. He has authored or co-authored morethan 40 research papers including the IEEE TRANS-

ACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, the IEEETRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, theIEEE TRANSACTIONS ON IMAGE PROCESSING, International Conferenceon Machine Learning, IEEE Conference on Computer Vision and PatternRecognition, and Knowledge Discovery in Database. His current researchinterests include statistical learning theory, computer vision, and optimization.

Xinchao Wang received the first class honorabledegree from The Hong Kong Polytechnic Univer-sity, Hong Kong, in 2010, and the Ph.D. degreefrom the École Polytechnique Fédérale de Lausanne,Lausanne, Switzerland, in 2015.

He was a Post-Doctoral Fellow with the Universityof Illinois at Urbana–Champaign, Champaign, IL,USA, with Prof. T. S. Huang. He is currently aTenure-Track Assistant Professor with the Depart-ment of Computer Science, Stevens Institute ofTechnology, Hoboken, NJ, USA. He has authored or

co-authored various venues including the IEEE TRANSACTIONS ON PATTERN

ANALYSIS AND MACHINE INTELLIGENCE, the IEEE TRANSACTIONS ON

IMAGE PROCESSING, the IEEE TRANSACTIONS ON MEDICAL IMAGING, theIEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS,IEEE Conference on Computer Vision and Pattern Recognition, EuropeanConference on Computer Vision, IEEE International Conference on ComputerVision, Neural Information Processing Systems, and International ConferenceOn Medical Image Computing & Computer Assisted Intervention. His currentresearch interests include computer vision, machine learning, and artificialintelligence.

Dr. Wang is an Associate Editor of the Journal of Visual Communicationand Image Representation.

Dacheng Tao (F’15) is currently a Professor ofcomputer science and an ARC Laureate Fellow withthe School of Information Technologies, Facultyof Engineering and Information Technologies, andthe Inaugural Director of the UBTECH SydneyArtificial Intelligence Centre, The University of Syd-ney, Darlington, NSW, Australia. He mainly appliesstatistics and mathematics to Artificial Intelligenceand Data Science. He has authored or co-authored1 monograph and more than 200 publications atprestigious journals and prominent conferences, such

as the IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTEL-LIGENCE, the IEEE TRANSACTIONS ON IMAGE PROCESSING, the IEEETRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, Inter-national Journal of Computer Vision, Journal of Machine Learning Research,Neural Information Processing Systems, International Conference on MachineLearning, IEEE Conference on Computer Vision and Pattern Recognition,IEEE International Conference on Computer Vision, European Conference onComputer Vision, IEEE International Conference on Data Mining (ICDM),and ACM Special Interest Group on Knowledge Discovery and Data Mining.

Mr. Tao is a Fellow of the Australian Academy of Science, AAAS, IAPR,OSA, and SPIE. He was a recipient of several best paper awards, such as theBest Theory/Algorithm Paper Runner-Up Award in IEEE ICDM in 2007, theBest Student Paper Award in IEEE ICDM in 2013, the Distinguished PaperAward in the 2018 International Joint Conference on Artificial Intelligence,the 2014 ICDM 10-Year Highest-Impact Paper Award, and the 2017 IEEESignal Processing Society Best Paper Award.


Recommended