+ All Categories
Home > Documents > Target Detection Via Network Filtering

Target Detection Via Network Filtering

Date post: 24-Sep-2016
Category:
Upload: ed
View: 213 times
Download: 0 times
Share this document with a friend
14
2502 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010 Target Detection Via Network Filtering Shu Yang and Eric D. Kolaczyk, Senior Member, IEEE Abstract—A method of network filtering has been proposed re- cently to detect the effects of certain external perturbations on the interacting members in a network. However, with large networks, the goal of detection seems a priori difficult to achieve, especially since the number of observations available often is much smaller than the number of variables describing the effects of the under- lying network. Under the assumption that the network possesses a certain sparsity property, we provide a formal characterization of the accuracy with which the external effects can be detected, using a network filtering system that combines Lasso regression in a sparse simultaneous equation model with simple residual anal- ysis. We explore the implications of the technical conditions un- derlying our characterization, in the context of various network topologies, and we illustrate our method using simulated data. Index Terms—Lasso regression, network topology, sparse net- work, target detection. I. INTRODUCTION A CANONICAL problem in statistical signal and image processing is the detection of localized targets against complex backgrounds, which often is likened to the proverbial task of “finding a needle in a haystack.” In this paper, we con- sider the task of detecting such targets when the “background” is neither a one-dimensional signal nor a two-dimensional image, but rather consists of the ‘typical’ behavior of inter- acting units in a network system. More specifically, we assume network-indexed data, where measurements are made on each of the units in the system and the interaction among these units manifests itself through the correlations among these measurements. Then, given the possible presence of an external effect applied to a unit(s) of this system, we take as our goal the task of identifying the location and magnitude of this effect. It is expected that evidence of this effect be diffused throughout the system, to an extent determined by the underlying network of interactions among system units, like the blurring of a point source in an image. As a result, an appropriate filtering of the observed measurements is necessary. These ideas are illustrated schematically in Fig. 1. While networks have been an important topic of study for some time, in recent years there has been an enormous surge in interest in the topic, across various diverse areas of science. Examples include computer traffic networks (e.g., [10]), biolog- ical networks (e.g., [1]), social networks (e.g., [25]), and sensor Manuscript received February 22, 2009; revised December 07, 2009. Current version published April 21, 2010. This work was supported in part by the NIH under award GM078987. The material in this paper was presented in part at the Joint Statistical Meeting, Denver, CO, August 2008. The authors are with the Department of Mathematics and Statistics, Boston University, Boston, MA 02215 USA (e-mail: [email protected]; ko- [email protected]). Communicated by J. Romberg, Associate Editor for Signal Processing. Digital Object Identifier 10.1109/TIT.2010.2043770 networks (e.g., [23]). Our network filtering problem was formu- lated by, and is largely motivated by the work of, Cosgrove et al. [8], who used it to tackle the problem of predicting genetic targets of biochemical compounds proposed as candidates for drug development. However, the problem is clearly general and it is easy to conceive of applications in other domains. The authors in [8] model the acquisition of network data, including the potential presence of targets, using a system of sparse simultaneous equation models (SSEMs), and propose to search for targets using a simple two-step procedure. In the first step, sparse statistical inference techniques are used to remove the background of network effects, while in the second step, out- lier detection methods are applied to the resulting network-fil- tered data. Empirical work presented in [8], using both simu- lated data and real data from microarray studies, demonstrates that such network filtering can be highly effective. However, there is no accompanying theoretical work in [8]. In this paper, we present a formal characterization of the per- formance of network filtering, exploring under what conditions the methodology can be expected to work well. A collection of theoretical results are provided, which in turn are supported by an extensive numerical study. Particular attention is paid to the question of how network structure influences our ability to de- tect external effects. The technical aspects of our work draw on a combination of tools from the literatures on sparse regression and compressed sensing, complex networks, and spectral graph theory. The remainder of the paper is organized as follows. The basic SSEM model and two-step network filtering methodology are presented formally in Section II. In Section III we characterize the accuracy with which the network effects can be learned from training data, while in Section IV, we use these results to quan- tify the extent to which external effects will be evident in test data after filtering out the learned network effects. Numerical results, based on simulations under various choices of network structure, are presented in Section V. Finally, we conclude with some additional discussion in Section VI. Proofs of all formal results are gathered in the Appendices. II. NETWORK FILTERING:MODEL AND METHODOLOGY Consider a system of units (e.g., genes, people, sensors, etc.). We will assume that we can take measurements at each unit, and that these measurements are likely correlated with measurements at some (small) fraction of other units, such as might occur through “interaction” among the units. For example, in [8], where the units are genes, the measurements are gene expression levels from microarray experiments. Genes regulating other genes can be expected to have expression profiles correlated across experiments. Alternatively, we could envision measuring environmental variables (e.g., temperature) at nodes in a sensor network. Sensors located sufficiently close 0018-9448/$26.00 © 2010 IEEE
Transcript
Page 1: Target Detection Via Network Filtering

2502 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Target Detection Via Network FilteringShu Yang and Eric D. Kolaczyk, Senior Member, IEEE

Abstract—A method of network filtering has been proposed re-cently to detect the effects of certain external perturbations on theinteracting members in a network. However, with large networks,the goal of detection seems a priori difficult to achieve, especiallysince the number of observations available often is much smallerthan the number of variables describing the effects of the under-lying network. Under the assumption that the network possessesa certain sparsity property, we provide a formal characterizationof the accuracy with which the external effects can be detected,using a network filtering system that combines Lasso regression ina sparse simultaneous equation model with simple residual anal-ysis. We explore the implications of the technical conditions un-derlying our characterization, in the context of various networktopologies, and we illustrate our method using simulated data.

Index Terms—Lasso regression, network topology, sparse net-work, target detection.

I. INTRODUCTION

A CANONICAL problem in statistical signal and imageprocessing is the detection of localized targets against

complex backgrounds, which often is likened to the proverbialtask of “finding a needle in a haystack.” In this paper, we con-sider the task of detecting such targets when the “background”is neither a one-dimensional signal nor a two-dimensionalimage, but rather consists of the ‘typical’ behavior of inter-acting units in a network system. More specifically, we assumenetwork-indexed data, where measurements are made on eachof the units in the system and the interaction among theseunits manifests itself through the correlations among thesemeasurements. Then, given the possible presence of an externaleffect applied to a unit(s) of this system, we take as our goal thetask of identifying the location and magnitude of this effect. Itis expected that evidence of this effect be diffused throughoutthe system, to an extent determined by the underlying networkof interactions among system units, like the blurring of a pointsource in an image. As a result, an appropriate filtering of theobserved measurements is necessary. These ideas are illustratedschematically in Fig. 1.

While networks have been an important topic of study forsome time, in recent years there has been an enormous surgein interest in the topic, across various diverse areas of science.Examples include computer traffic networks (e.g., [10]), biolog-ical networks (e.g., [1]), social networks (e.g., [25]), and sensor

Manuscript received February 22, 2009; revised December 07, 2009. Currentversion published April 21, 2010. This work was supported in part by the NIHunder award GM078987. The material in this paper was presented in part at theJoint Statistical Meeting, Denver, CO, August 2008.

The authors are with the Department of Mathematics and Statistics, BostonUniversity, Boston, MA 02215 USA (e-mail: [email protected]; [email protected]).

Communicated by J. Romberg, Associate Editor for Signal Processing.Digital Object Identifier 10.1109/TIT.2010.2043770

networks (e.g., [23]). Our network filtering problem was formu-lated by, and is largely motivated by the work of, Cosgrove etal. [8], who used it to tackle the problem of predicting genetictargets of biochemical compounds proposed as candidates fordrug development. However, the problem is clearly general andit is easy to conceive of applications in other domains.

The authors in [8] model the acquisition of network data,including the potential presence of targets, using a system ofsparse simultaneous equation models (SSEMs), and propose tosearch for targets using a simple two-step procedure. In the firststep, sparse statistical inference techniques are used to removethe background of network effects, while in the second step, out-lier detection methods are applied to the resulting network-fil-tered data. Empirical work presented in [8], using both simu-lated data and real data from microarray studies, demonstratesthat such network filtering can be highly effective. However,there is no accompanying theoretical work in [8].

In this paper, we present a formal characterization of the per-formance of network filtering, exploring under what conditionsthe methodology can be expected to work well. A collection oftheoretical results are provided, which in turn are supported byan extensive numerical study. Particular attention is paid to thequestion of how network structure influences our ability to de-tect external effects. The technical aspects of our work draw ona combination of tools from the literatures on sparse regressionand compressed sensing, complex networks, and spectral graphtheory.

The remainder of the paper is organized as follows. The basicSSEM model and two-step network filtering methodology arepresented formally in Section II. In Section III we characterizethe accuracy with which the network effects can be learned fromtraining data, while in Section IV, we use these results to quan-tify the extent to which external effects will be evident in testdata after filtering out the learned network effects. Numericalresults, based on simulations under various choices of networkstructure, are presented in Section V. Finally, we conclude withsome additional discussion in Section VI. Proofs of all formalresults are gathered in the Appendices.

II. NETWORK FILTERING: MODEL AND METHODOLOGY

Consider a system of units (e.g., genes, people, sensors,etc.). We will assume that we can take measurements at eachunit, and that these measurements are likely correlated withmeasurements at some (small) fraction of other units, suchas might occur through “interaction” among the units. Forexample, in [8], where the units are genes, the measurementsare gene expression levels from microarray experiments. Genesregulating other genes can be expected to have expressionprofiles correlated across experiments. Alternatively, we couldenvision measuring environmental variables (e.g., temperature)at nodes in a sensor network. Sensors located sufficiently close

0018-9448/$26.00 © 2010 IEEE

Page 2: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2503

Fig. 1. Schematic illustration of the network filtering process proposed in thispaper, shown in two stages. In the first stage, the aim is to recover informationon the correlation (i.e., ��) among the five network units, given training data � .In the second stage, that information is used to filter new data �� , produced inthe presence of an effect external to the system (i.e., �), so as to detect the targetof that effect.

to each other, with respect to the dynamics of the environmentalprocess of interest, can be expected to yield correlated readings.

We will also assume that there are two possible types of mea-surements: a training set, obtained under ‘standard’ conditions,and a test set measured under the influence of additional “ex-ternal” effects. The training set will be used to learn patterns ofinteraction among the units (i.e., our “network”), and with thatknowledge, we will seek to identify in the test data those unitstargeted by the external effects.

We model these two types of measurements using systems ofsimultaneous equation models (SEMs). Formally, suppose thatfor each of the units, we have in the training set replicatedmeasurements, which are assumed to be realizations of the el-ements of a random vector . Let be theth element of , and let denote all elements of except

. We specify a conditional linear relationship among these el-ements, in the form

(1)

where represents the strength of association of the measure-ment for the i th unit with that of the th unit, and are errorterms, assumed to be independently distributed as .That is, we specify a so-called “conditional Gaussian model”for the , which in turn yields a joint distribution for in theform

(2)

with being a matrix whose entry is , for ,and zero otherwise. See [9, ch. 6.3.2]. The matrix is as-sumed to be positive definite. In addition, we will assume (andhence ) to be sparse, in the sense of having a substantial

proportion of its entries equal to zero. A more precise charac-terization of this assumption is given below, in the statement ofTheorem 1.

We can associate a network with this model using the frame-work of graphical models (e.g., [19]). Let each unit in oursystem correspond to a vertex in a vertex set ,and define an edge set such that if and onlyif . Then the model in (2), paired with the graph

, is a Gaussian graphical model, with concentration(or “precision”) matrix and concentrationgraph . Since we assume to be sparse, the graph likewisewill be sparse. Gaussian graphical models are a common choicefor modeling multivariate data with potentially complicatedcorrelation (and hence dependency) structure. This structureis encoded in the graph , and questions regarding the natureof this structure often can be rephrased in terms of questionsinvolving the topology of . In recent years, there has beenincreased interest in modeling and inference for large, sparseGaussian graphical models of the type we consider here (e.g.,[11] and [21]).

For the test set, our observations are assumed to be realiza-tions of another random vector, say , the ele-ments of which differ from those of only through the possiblepresence of an additive perturbation. That is, we model each ,conditional on the others, as

(3)

where denotes the effect of the external perturbation for theth unit, and the error terms are again independently dis-

tributed as . Similar to (2), we have in this scenario

(4)

where .The external effects are assumed unknown to us but sparse.

That is, we expect only a relatively small proportion of units tobe perturbed. Our objective is to estimate the external effectsand to detect which units were perturbed i.e., to detect thoseunits for which stands out from zero above noise. But wedo not observe the external effects directly. Rather, these ef-fects are ‘blurred’ by the network of interactions captured in ,as indicted by the expression for the mean vector in (4). Ifwere known, however, it would be natural to filter the data ,producing

(5)

The random vector has a multivariate Gaussian distribu-tion, with expectation and covariance . Hence, el-ement-wise, each is distributed as , and there-fore the detection of perturbed units reduces to detection of asparse signal against a uniform additive Gaussian noise, whichis a well-studied problem. Note that under this model, we ex-pect the noise in to be correlated. However, given the as-sumptions of sparsity on , these correlations will be relativelylocalized.

Page 3: Target Detection Via Network Filtering

2504 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Of course, typically is not known in practice, and soin (5) is an unobtainable ideal. Studying the same problem, Cos-grove et al. [8] proposed a two-stage procedure in which: (i)

simultaneous sparse regressions are performed to infer ,row-by-row, yielding an estimate ; and (ii) the ideal residualsin (5) are predicted by the values

(6)

after which detection is carried out.1 They dubbed this overallprocess “network filtering.” A schematic illustration of networkfiltering is shown in Fig. 1.

Our central concern in this paper is with characterizing theconditions under which network filtering can be expected towork well. Motivated by the original context of Cosgrove et al.,involving a network of gene interactions and measurementsbased on microarray technology, we assume here that: (i)

; (ii) the matrix is sparse; and (iii) the vectoris sparse. In carrying out our study, we adopt a strategy forestimating based on Lasso regression [24], a now-canonicalexample of sparse regression. Specifically, motivated by (1),we estimate each row as

(7)

where is a regularization parameter. Following this es-timation stage, we carry out detection using simple rank-basedprocedures.

We present our results in two stages, first describing condi-tions under which estimates accurately, given the system ofsparse simultaneous equation models (SSEMs) defined by (1),and then discussing the nature of the resulting vector . In bothstages, we explore the implications of the topological structureof on our results.

III. ACCURACY IN ESTIMATION OF

At first glance, accurate estimation of seems impossible,since even if the error terms are small, this noise typicallywill be inflated by naive inversion of our systems of equations(i.e., because ). However, recent work on analogous prob-lems in other models has shown that under certain conditions,and using tools of sparse inference, it is indeed possible to ob-tain good estimates. Results of this nature have appeared underthe names “compressed sensing,” “compressive sampling,” andsimilar. See the recent review [6], and the references therein.The following result is similar in spirit to these others, for theparticular sparse simultaneous equation models we study here.

Theorem 1: Assume the training model defined in (1) and(2), and set . Let be the largest number of

1Technically, Cosgrove et al. work under a model that differs slightly fromours, sharing the same conditional distributions, but arrived at through specifi-cation of a different joint distribution. See [9, ch. 6.3] for discussion comparingsuch “simultaneous Gaussian models” with our conditional Gaussian model.

nonzero entries in any row of , and suppose that andsatisfy

(8)

and

(9)

Here and refer to maximum and minimum eigen-values and , where

and is a function to be defined later.2 Finally, as-sume that , for a constantand . Let , for .Then it follows that, with overwhelming probability, for everyrow of the estimator in (7) satisfies

(10)

where and is a constant.

Remark 1: The accuracy of is seen in (10) to dependprimarily on the product and on . The constant canbe bounded by an expression of the form times aconstant depending only on the structure of . The magnitude of

therefore is controlled essentially by the extent to whichis less than 1, which in turn is a rough reflection of the sparsityof the network. Hence, in order to have good accuracy, mustbe small compared to . In particular, if , for

, then the error in (10) behaves roughly like .

Remark 2: Clearly, it cannot be expected that we estimatewith high accuracy in all situations. The expressions in (8)

and (9) dictate sufficient conditions under which, with over-whelming probability (meaning with probability decaying ex-ponentially in ), we can expect to do well. Due to the inti-mate connection between the covariance and the concentra-tion graph , these conditions effectively place restrictions onthe structure of the network we seek to filter, with (8) controllingthe relative magnitude of the eigenvalues of the matrix ,and (9), its sparseness. Note that since is simply the maximumdegree of , (9) relates the maximum extent of the degree dis-tribution of to the sample size . We explore the nature ofthese conditions in more detail immediately here.

Remark 3: In general, of course, choice of the Lasso regular-ization parameter in (7) matters. The statement of The-orem 1 includes constraints on the range of acceptable valuesfor this parameter. In particular, it suggests that should varylike , which for means we want

. The theorem does not, however, provide ex-plicit guidance on how to set this parameter in practice. Forthe empirical work shown later in this paper, we have usedcross-validation, which we find yields results like those pre-dicted by the theorem over a broad range of scenarios.

Remark 4: There are results in the literature that addressother problems sharing certain aspects of our network filtering

2Specifically, � is defined in Section III, Section III-B, immediately following(14).

Page 4: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2505

Fig. 2. Plots of ER, BA, and geometric random graphs of size � � ��� andaverage degree �� � �.

problem, but none that address all together. For example, thebound in (10) is like that in work by Candès and Tao andcolleagues (e.g., [4] and [5]), although for a single regression,rather than a system of simultaneous regressions. In addition,those authors use constrained minimization for parameter esti-mation, rather than Lasso-based optimization. As Zhu [26] hasrecently pointed out, there are small but important differencesin these closely related problems. Our proof makes use of Zhu’sresults. Similarly, Greenshtein and Ritov [15] present resultsfor models that—in principle at least—include the individualunivariate regressions in (1), although again their results do notencompass a system of such regressions. Furthermore, theirresults are in terms of mean-squared prediction error, ratherthan in terms of the regression coefficients themselves. Finally,Meinshausen and Bühlmann [21] have studied the use of Lassoin the context of Gaussian graphical models, but for the purposeof recovering the topology of i.e., for variable selection,rather than parameter estimation. The proof of Theorem 1 maybe found in Appendix A.

In the remainder of this section, we examine conditions (8)and (9) in greater depth. These conditions derive from our use ofcertain concentration inequalities, which—although central tothe proof of our result—can be expected to be somewhat conser-vative. Our numerical results, shown later, confirm this expec-tation. Nonetheless, these conditions are useful in that they helpprovide insight into the way that the network topology struc-ture, on the one hand, and the sample size , on the other, canbe expected to interact in determining the performance of ournetwork filtering methodology.

A. Eigenvalue Constraint

Recall that the covariance matrix is proportional to. In order to better understand the condition on the covari-

ance matrix in (8), consider the special case of

(11)

where is the adjacency matrix for a graphis a diagonal matrix, is the degree (i.e., the

number of neighbors) of vertex , and is a constant. Herethe covariance is defined entirely in terms of the topology ofthe concentration graph . While later, in Sections 3 and 4, weuse simulation to explore more complicated covariance struc-tures, where the are assigned randomly according to certaindistributions, the simplified form in (11) is useful in allowingus to produce analytical results. In particular, conditions onreduce to conditions on our network topology.3 For example,the following theorem describes a sufficient condition underwhich (8) holds for this model.

Theorem 2: Suppose that the covariance matrix from The-orem 1 is defined through (11), with . Denote

(12)

where indicates that the vertices and are neighborsin and is the maximum vertex degree.Then (8) on the eigenvalues of is satisfied if

(13)

Proof of this result may be found in Appendix A. The restric-tion on ensures that the matrix is diagonally dominant,which is needed for our proof, although it likely could be weak-ened. Note that (13) involves the graph only through the de-gree sequence . More precisely, this condition re-lates the average harmonic mean of neighbor degrees (i.e., )and the maximum degree to the sample size and the constant. Accordingly, given a network, it is straightforward to explore

the implications of this condition numerically. For example, wecan explore the range of values for which the condition holds,given .

Fig. 2 shows examples of three network topologies. The firstis an Erdös-Rényi (ER) random graph [13], a classical form ofrandom graph in which vertex pairs are assigned edges ac-cording to independently and identically distributed Bernoullirandom variables (i.e., coin flips). The degree distribution of anER network is concentrated around its mean and has tails thatdecay exponentially fast. The second is a random graph gener-ated according to the Barabási and Albert model [2], which was

3We note that (11) can be rewritten in the form � � �� � �� � � � �,where � is the (normalized) Laplacian matrix of the graph �. In other words,the precision matrix � � in this simple model is just a modified Laplacianmatrix.

Page 5: Target Detection Via Network Filtering

2506 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Fig. 3. Plots of eigenvalue ratio for ER, BA, and geometry graph under different values of �.

originally motivated by observed structure in the World WideWeb. The defining characteristic of the BA model is that the de-rived network has a degree distribution of a power-law form,with tails decreasing like for large . Therefore, the BAnetworks tend to contain many vertices with only a few neigh-bors, and a few vertices with many neighbors. Lastly, we alsouse a geometric random graph model, such as might be appro-priate for modeling spatial networks. Following [21], vertices inthe graph are uniformly distributed throughout the unit square

, and each vertex pair has an edge with probability, where is the standard normal density func-

tion and is the Euclidean distance in between and. In all three cases, the random graph was of size and

had average degree .In Fig. 3, we show the eigenvalue ratio in (8), under the sim-

plified covariance structure in (11), for these ER, BA, and geo-metric random graphs, as a function of . The horizontal linesrepresent the theoretical eigenvalue ratio bound given by The-orem 1. The open symbols (including the “plus” symbol) indi-cate graphs that satisfy the condition in Theorem 2, while thefilled symbols indicate graphs that do not satisfy the condition.We can see from the plot that the condition in Theorem 2 clearlyis conservative, since as a function of it ceases to hold long be-fore the inequality in (8) is violated.

B. Sparsity Constraint

The second condition in Theorem 1, given in (9), can be readas a condition on the sparsity of the precisionmatrix , and therefore a condition on the sparsity ofour network graph . The analytical form of the functionis

(14)

where andis the entropy function.

While it is not feasible to produce a closed-form solution into the inequality (9), it is straightforward to explore the spaceof solutions numerically.

Note that actually is a function of the three parameters, and through the two ratios and . In prac-

tice we expect both ratios to be in the interval . Shown inFig. 4 is , as a function of , for a handful of representativechoices of . We see from the plot that the theory suggests,through (9), that the sparsity should be bounded by roughly

. Our numerical results, however, shown later, indicatethat the theory is quite conservative, in that, for example, for oursimulations we successfully used networks with sparsity on theorder of . Analogous observations have been made in[4]. Also shown in Fig. 4 is a 3-D plot of , as a functionof both and . In this plot, the dark area corresponding tothe innermost contour line satisfies the condition that .Again, the value of the information shown here is primarily asan indication of the existence of feasible combinations of ,and allowing for the accurate estimation of the rows of .

IV. ACCURACY OF THE NETWORK FILTERING

With the accuracy of quantified, we turn our attention tothe effectiveness of our filtering of the network effects. Specifi-cally, in the following theorem we characterize the behavior of

, defined in (6), as a predictor of , defined in (5).

Theorem 3: Suppose is a vector of test data, obtainedaccording to the model defined in (3) and (4). Letbe defined as in (6) and let . Then conditional on

has a multivariate normal distribution, with expectation andvariance

(15)

(16)

Page 6: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2507

Fig. 4. 2-D plot and 3-D plot showing the behavior of � ��� for three valuesof the ratio ���.

Furthermore, under the conditions of Theorem 1, element-wisewe have

(17)

and

(18)

with overwhelming probability, where and are as inTheorem 1.

Proof of this theorem may be found in Appendix B. Recallthat in (5) is distributed as a multivariate normal randomvector, with expectation and variance . Equations(15) and (16) show that our predictor mimics well tothe extent that our error in estimating —that is, those termsinvolving —are small. Theorem 1 quantifies the magnitude ofthe rows of , from which we obtain theterm in our bounds on the element-wise predictive biasin (17) and variance in (18).

Remark 4: In the case that there are no external effects ex-erted upon our system i.e., , the elements of the

ideal estimate are just identically distributednoise. This case corresponds to the intuitive null distributionwe might use to formulate our detection problem as a statisticalhypothesis testing problem. The implication of the theorem isthat, in using rather than , following substitution offor , the price we pay is that the elements are instead dis-tributed as , where the differ from by no morethan . Treating asa constant for the moment, this term is dominated by ,i.e., our error in estimating the rows of . Hence, for example,if with , as in Remark 1, then the variances

will also be .

Remark 5: Suppose instead that, for some .

This case corresponds to the simplest alternative hypothesiswe might use, involving a nontrivial perturbation, and is areasonable proxy for the type of genetic perturbations (e.g.,from gene knock-out experiments) considered in Cosgrove etal. [8]. Now the bias is potentially nonzero, even for units

with . But, again treating as aconstant, and assuming , this bias will be onlynegligibly worse than the magnitude of theideal standard deviation . And the variance will again be

. Therefore, we should be able to detect single-unitperturbations well for sufficiently above the noise. Oursimulation results in the next section confirm this expectation.

Now consider the term , which reflects theeffect of the topology of on our ability to do detection withnetwork filtering. This term will not necessarily be a constant in

, due to the role of in the bounds (8) and (9) of Theorem 1,constraining the behavior of . The followinglemma lends some insight into the behavior of this term in thecase where the precision matrix again has the simpleform specified in (11). The proof may be found in Appendix B.

Lemma 1: Suppose that , as in(11). Then

(19)

Remark 6: Because we assume that the network will besparse, and that , the above result indicates that theterm can be treated under our simplified co-variance as a constant essentially with respect to in ex-pressions like (17) and (18).

V. SIMULATION RESULTS

A. Background

In this section, we use simulated network data to further studythe performance of our proposed network filtering method. Thedata are drawn from the models for training and test data definedin Section II, with randomly generated covariance matrices .We define these covariances through their corresponding pre-cision matrices , which are obtained in turn by: (i)generating a random network topology ; and then

Page 7: Target Detection Via Network Filtering

2508 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

(ii) assigning random weights to entries in corresponding topairs with edges . These collections of weightsare then rescaled in a final step to coerce into the form

and, if necessary, to enforce positive definiteness. For thetopology , we use the three classes of random network topolo-gies described above in Section III-A i.e., the ER,BA, and geometric networks. For each choice of network, weuse nodes, each of which has an average degree of

. The adjacency matrices of the ER and BA model aregenerated randomly using the algorithms listed in [3], while thatof the geometric network is generated according to the methoddescribed in [21].

In implementing our network filtering method, the Lars [12]implementation of the Lasso optimization in (7) was used, ontraining data sets of various sample sizes for each network. TheLasso regularization parameter was chosen by cross-valida-tion. To generate testing data, we used single-unit perturbationsof the form , where isin the th position, for each . Since in oursimulation is effectively set to 1, can be interpreted as thesignal-to-noise ratio (SNR) of the underlying perturbation. Inour simulations, we let range over integers from 1 to 20. Ourfinal objective of detection is to find the position of the unitat which the external perturbation occurred. In our proposednetwork filtering method, we declare the perturbed unit to bethat corresponding to the entry of with largest magnitude i.e.,

.In each experiment described later, our method is compared

with two other methods. The first, called True, is that in whichthe ideal is used instead of , which presumes knowledgeof the true . The second, called Direct, is that in which the ac-tual testing data , i.e., the data without network filtering, areused instead of . In both cases, we declare the perturbed unitto be that corresponding to the entry of largest magnitude. TheTrue method gives us a benchmark for the detection error underthe ideal situation that we already have all of the network in-formation, while the Direct method is a natural approach in theface of having no information on the network. By comparingour method with the two, we may gauge how much is gainedby using the network filtering method. In all cases, performanceerror is quantified as the fraction of times a perturbed unit is notcorrectly identified, i.e., the proportion of misdetections. Resultsreported below for all three methods are based in each case upon30 replicates of the testing data. Our plots show average propor-tions of misdetections and one standard deviation.

B. Results

First we present the results from an experiment where isdefined according to the simple formulation given in (11), thedefinition that underlays the results in Theorem 2 and Lemma1. That is, we define in terms of just the (random) adjacencystructure of our three underlying networks, scaled by an appro-priate choice of to ensure positive definiteness. We may thinkof this case, from the perspective of the simulation design afore-described, as one with a particular nonrandom choice of weightsfor edges in the network , i.e., where .

Fig. 5. Plots of the proportion of misdetections versus SNR, for the BA, ER,and geometric random networks, based on the simplified covariance model in(11), using � � ����� � � ��� and � � ��. Error bars indicate one standarderror over 30 test datasets.

Fig. 5 shows the average proportions of misdetections, as afunction the SNR, for these three models. Note that since the

Page 8: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2509

underlying graphs are random, there is some variability in suchdetection results from simulation to simulation. However, theseplots and the others below like them are representative in our ex-perience. From the plots in the figure, we can see that in all casesthe network filtering offers a significant improvement over theDirect method, and in fact comes reasonably close to matchingthe performance of the True method, with misdetections at arate of roughly 5%–25% for high SNR. Performance differssomewhat with respect to networks of different topology. Thenetwork filtering method shows the most gain over the Directmethod with the BA network. This phenomenon is consistentwith our intuition: the distribution of edges in the BA network isthe least uniform one, and certain choices of perturbed unit (i.e.,perturbed units with large degree ) will enable the effects ofperturbation to spread comparatively widely. Hence, obtainingand correcting for the internal interactions among units in thenetwork is particularly helpful in this case.

Now consider the assignment of random weights to edgesin , which allows us to generate a richer variety of models.For this purpose, we choose the family of beta distributions

from which to draw weights independently foreach edge . Three different classes of distributionswere used, i.e., , and ,which gives flat (uniform), U-shape, and peaked shape forms.Shown in Fig. 6 are the results of our network filtering method,the True method, and the Direct method, for each of these threechoices of weight distributions, for each of the three networktopologies. The same (random) network topology is used ineach plot for each type of network.

Broadly speaking, these plots show that the performanceof network filtering in the context of randomly generatededge weights , as compared to that of the True and Di-rect methods, is essentially consistent with the case of fixededge-weights underlying the plots in Fig. 5. However, thereare some interesting nuances. For example, in the case of Flatweights, network filtering in fact is able to match the perfor-mance of True for all three classes of graphs. On the other hand,in the ER random network topology this matching occurs onlywhen the edge-weight distribution is flat (i.e., ), andin the BA random network topology, when the distribution iseither -shaped (i.e., ) or flat (i.e., ).Nevertheless, the qualitatively similar performance acrosschoice of edge-weight distribution suggests that most importantelement here is the network structure, indicating connectionbetween pairs of units, with the strength of connection beingsecondary.

Finally, we consider the effect of sample size and, there-fore, implicitly, the extent to which the condition in (8) on thestructure of the covariance matrix may be relaxed. For thesame networks used in the simulations described above, with

units, we varied the sample size to range over 20, 50,100, and 150. Weights of the network edges are set accordingto a distribution, which is the “peak” case. Trainingand testing data were generated as before. The results of usingnetwork filtering in these different settings are shown in Fig. 7.

Again, our network filtering method is seen to work similar toabove. Even for a sample size as small as , our methodstill does better than the Direct method in all three models, par-ticularly under the BA and ER models.4

On a final note, we point out that in all of the experiments therichness of network models studied is much broader than sug-gested by our theory. As was mentioned earlier, the concentra-tion inequalities we use can be expected to be conservative in na-ture, and therefore some of the bounds are more restrictive thanpractice seems to indicate is necessary. For example, in our sim-ulations involving the geometric random graph in Fig. 2, with asample size of and , the theoretical bound (8) forthe eigenvalue ratio is 6.12, while the actual value achieved bythis ratio is 219 in this instance. Also, the maximum degrees ofthe graphs in most of the simulations are larger than the averagedegree 4, and hence the sparseness rate , whichis already larger than those theoretical sparse rates suggested byFig. 1. Yet still we observed the network filtering method to per-form quite well. It is an interesting open question to see if thetheory can be extended to produce bounds like (8) that more ac-curately reflect practice, to serve as a better practical guides forusers.

VI. DISCUSSION

The concept of network filtering considered in this paper wasfirst proposed by Cosgrove et al. [8], as a methodology for fil-tering out the effects of “typical” gene regulatory patterns inDNA microarray expression data, so as to enhance the potentialsignal from genetic targets of putative drug compounds. Here,we have formalized the methodology of Cosgrove et al. andestablished basic conditions under which it may be expectedto perform well. Furthermore, we have explored the implica-tions of these conditions on the topology of the network un-derlying the data (i.e., a Gaussian concentration graph). Proofof our results rely on principles and techniques central to theliterature on compressed sensing and, therefore, like other re-sults in that literature, make performance statements that holdwith overwhelming probability. Numerical simulation resultsstrongly suggest a high degree of robustness of the methodologyto departure from certain of the basic conditions stated in ourtheorems regarding network topology. Our current work is nowfocused on the development of adaptive learning strategies thatintentionally utilize perturbations (i.e., in the form of the vectors

) to more efficiently explore network effects (i.e., the matrix).

APPENDIX APROOF OF THEOREM 1

4We note that some care must be used in fitting Lasso with � � �, due tonumerical instabilities that can arise. This issue affects any method attemptingto estimate the inverse of a covariance matrix (as is implicitly being done here).Krämer [17] describes how a reparameterization of the Lasso penalty can beused to avoid this problem.

Page 9: Target Detection Via Network Filtering

2510 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Fig. 6. Plots of proportions of mis-detections versus signal-to-noise ratio. Columns: BA (left), ER (middle), and geometric (right) random networks. Rows:U-shaped (top), flat (middle), and peaked (bottom) choice of weight distributions, generated according to ��������� �������������� and ��������� distribu-tions, respectively.

Theorem 1 jointly characterizes the accuracy of simulta-neous regressions, each based on the model in (1), i.e., for

(20)

For convenience, we reexpress the above model for a single re-gression in the generic form

(21)

Here is a design matrix with rows sampled i.i.d.from a multivariate normal distribution , with

covariance matrix is an error vector, inde-pendent of , with i.i.d. elements; and is theresponse vector.

We will make use of a result of Zhu [26], which requires thenotion of restricted isometry constants. Following Zhu,5 we de-fine the -restricted isometry constant of the matrix as thesmallest quantity such that

(22)

for all index subsets with . Zhu’s resultis then as follows.

Lemma 2 (Zhu): If: (i) the number of nonzero entries of isno more than ; (ii) the isometry constants and obeythe inequality ; and (iii) the Lasso regularizationparameter obeys the constraint , for ,then

(23)

5This definition differs slightly from that in Candés [4]. See [26] for discus-sion.

Page 10: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2511

Fig. 7. Plots of proportion of misdetections versus SNR ratio for the BA (left),ER (middle), and geometric (right) random network models, for sample sizes� � ��� ������, and 150.

where

(24)

Zhu’s first condition is assumed in our statement of Theorem1. Therefore, to prove Theorem 1 we need to show, under theother conditions stated in our theorem, that Zhu’s second andthird conditions above hold simultaneously for each of our re-gressions, with overwhelming probability. In addition, we needto show that the right-hand side (RHS) of (23) is bounded aboveby the RHS of (10).

A. Verification of Lemma 2, Condition (ii)

The essence of what is needed for the restricted isometry con-stants is contained in the following lemma.

Lemma 3: Suppose is a submatrix of the covariance ma-trix with columns variables corresponding to these in of theindices in set , where . Denote the largest and smallesteigenvalues of any such matrix as and ,respectively. Suppose too that

and (25)

where is defined as in (14). Then,holds with overwhelming probability.

The covariance matrix corresponding to any single regres-sion is a submatrix of in Theorem 1, and, hence, so is .By the interlacing property of eigenvalues (e.g., Golub and vanLoan [14, Thm 8.1.7]), which relates the eigenvalues of a sym-metric matrix to those of its principle submatrices, as long assatisfies the eigenvalue constraint (8), the matrices will aswell. So it is sufficient to prove Lemma 3.

Proof of Lemma 3: Let denote the submatrix ofcorresponding to the subset of indices . Since the rows ofare independent samples from , the rows of are

independent samples from . Let be the th largestsingular value of and be the th largest singular valueof . The eigenvalue condition in the lemma reducesto . Without lossof generality, therefore, assume that while

.Note we can express as , where

. Then and, hence, theeigenvalues of are the same as those of . Thus,we have

(26)

(27)

Let denote the th largest singular value of . There-fore, we have6

(28)

(29)

6��� � equals the smallest eigenvalue of � ���, which is � in [4]. Sim-ilar for ��� � .

Page 11: Target Detection Via Network Filtering

2512 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Denote by the smallest and largest singularvalues of their argument. Notice that for any index set ,we have

Thus, we need only to consider the situation whereand choose as the smallest constant that satisfies (22) for anysubmatrix of size . Therefore, we set

where and . Itthen follows that .

Now, by the large deviation results in [20], [4], for a standardGaussian random matrix , there are two relevantconcentration inequalities:

(30)

(31)

where is an term.We can then use the above tools and concentration inequal-

ities to see how behaves under the conditions described inLemma 3. Notice that, for , we have

(32)

(33)

(34)

(35)

(36)

(37)

Denoting , we have by (29) that .Therefore, for the term with in (37), we have by Bonferroni’sinequality that

As in [4], we fix , from which

it follows that and

. Hence the above inequality is equivalent to

For the term with in (37), the analogous inequality

may be verified using a similar argument. Combining these twoprobability inequalities for and , we have that

. Ignoring the negli-gible terms, it follows that when is large enough

holds with overwhelming probability. Defining, we have that

. Imposing the condition , we obtainthat , and therefore Lemma 3 is proved.

B. Verification of Lemma 2, Condition (iii) and the RHS of (10)

Let denote the regression (21) for theth of the simultaneous regressions in (20). Condition (iii) of

Lemma 2 requires that the regularization parameter be suchthat , where . If so, and assumingof course that conditions (i) and (ii) are satisfied as well, thenthe inequality in (23) says that . Weshow that the condition on in the statement of Theorem 1 i.e.,that , guarantees that Condition (iii) holdsfor every with overwhelming probability. In addi-tion, we show that with overwhelming prob-ability, where , and is boundedby times a constant, as claimed in Remark 1.

Notice that if we have , then is dis-tributed as chi-square on degrees of freedom. By [18, Sec 4.1,Lemma 4], for

and

Therefore

and similarly

We choose so that uniformly in withprobability at least , so as to match the rate inSection A. Specifically, we set .Hence, for sufficiently small we have .Therefore, as long as ,then with probability exceeding , theinequality holds uniformly in .Suppose , with . Under this condition,

and our requirement thusreduces to . Similarly,

Page 12: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2513

with , we also have with probability exceedingthat .

Let and be defined as in the statement of Theorem 1.Then by requiring that , from the above re-sults it follows that Condition (iii) of Lemma 2 holds for all

, with high probability. Furthermore, with highprobability, , for . Therefore, bythe bound (23) in Lemma 2, we have established the bound (10)in Theorem 1, except for the constant .

Specifically, it remains for us to establish that forall . Denote the value for an arbitrary regression by .Note that the here in our paper corresponds to the square ofwhat is called “ ” in Zhu [26]. Hence by [26, eq. (17)], issmaller than the larger root of a quadratic equation of the form

where is the argument and and are positiveparameters.7 For our purposes, it is enough to remark that

is bounded by a constant pro-portional to relates to through the expression

, and is a constant greater thanfour.

As the larger root of the above quadratic equation

Note that , whichis bounded by , becauseand are all positive. Hence, we have

(38)

It remains for us to bound the RHS of (38). Recall that, byconstruction, , and that is assumedstrictly less than 1. Thus, because is bounded by a term propor-tional to , the second term in the RHS of (38) is boundedby a term proportional to . Furthermore, if ,then the first term is bounded by a term proportional to

. Last, therefore, suppose that and consider theterm

(39)

The term involving in the RHS of (39) is easily bounded,per our reasoning above, while—taking, for example,

in the condition of Lemma 2, the other term is equalto .

7Our notation here is slightly different from that of [26].

Hence, returning to the context of our original problem, foreach , the constant is bounded by some con-stant times . Letting be the largest of thesebounds, our proof of Theorem 1 is complete.

APPENDIX BPROOF FOR THEOREM 2

To show that condition (8) of Theorem 1 holds in the contextof Theorem 2, we first bound the eigenvalue ratio of the covari-ance matrix . For , the matrix is diagonallydominant, and hence, by the Levy-Desplanques Theorem, non-singular. Furthermore, since is real and symmetric, it is anormal matrix. Therefore

where is condition number of the precision matrix. As a result, by an inequality of Guggenheimer, Edelman,

and Johnson [16, p. 4] for condition numbers, we can bound oureigenvalue ratio as

(40)

Since , direct calculation showsthat

where is the th element of the diagonal matrix . As for thequantity , we note that

Denoting , and applying a result of Ostrowskifor determinants of diagonally dominant matrices (e.g., [22]),we find that

Hence, we have .

Page 13: Target Detection Via Network Filtering

2514 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 56, NO. 5, MAY 2010

Combining the relevant expressions above, we have that

(41)

(42)

Denoting and as in (12), bounding the RHS of (42) by, to enforce the bound (8),

and some trivial manipulation of the resulting inequality, yieldsthe condition in (13), as desired.

APPENDIX CPROOF FOR THEOREM 3

Note that the difference of the predictor and the true ex-ternal effect is given as .So the bias term is just

where , and hence for the th component of the biasterm, we have

Here the first term in brackets follows from Theorem 1, whilethe second follows from the definition of the matrix norm.

For the variance of the predictor , we have

So the variance of the th element of is

since . The absolute value of the second term isbounded by , and thus (18) follows.

APPENDIX DPROOF FOR LEMMA 1

Under the model , the largestnumber of nonzero entries is , the maximal degree of thenetwork. So by the eigenvalue condition in Theorem 1, we have

Now. And clearly

. Furthermore

By [7, Lem. 8.6], . Therefore,.

Combining these results, we have

REFERENCES

[1] U. Alon, An Introduction to Systems Biology: Design Principles of Bi-ological Circuits. Boca Raton, FL: Chapman and Hall/CRC, 2007.

[2] A. L. Barabasi and R. Albert, “Emergence of scaling in random net-works,” Science, vol. 286, pp. 509–512, 1999.

[3] V. Batagelj and U. Brandes, “Efficient generation of large random net-works,” Phys. Rev. E, vol. 71, no. 3, pp. 1–5, 2005.

[4] E. Candès and T. Tao, “Decoding by linear programming,” IEEE Trans.Inf. Theory, vol. 51, pp. 4203–4215, 2004.

[5] E. Candès, J. Romberg, and T. Tao, “Stable signal recovery from in-complete and inaccurate measurements,” Commun. Pure Appl. Math.,vol. 59, pp. 1207–1223, 2006.

[6] E. Candés and M. Wakin, “An introduction to compressive sampling,”IEEE Signal Process. Mag., vol. 25, no. 2, pp. 21–30, 2008.

[7] F. Chung and L. Lu, Complex Graphs and Networks. Boston, MA:Amer. Math. Soc., 2006.

[8] E. Cosgrove, Y. Zhou, T. Gardner, and E. D. Kolaczyk, “Predictinggene targets of perturbations via network-based filtering of mRNA ex-pression compendia,” Bioinformatics, vol. 24, no. 21, pp. 2483–2490,2008.

[9] N. A. C. Cressie, Statistics for Spatial Data, revised ed. New York:Wiley , 1993.

[10] M. Crovella and B. Krishnamurthy, Internet Measurement. NewYork: Wiley, 2006.

[11] A. Dobra, C. Hans, B. Jones, J. R. Nevins, G. Yao, and M. West,“Sparse graphical models for exploring gene expression data,” J.Multivar. Anal., no. 90, pp. 196–212, 2004.

Page 14: Target Detection Via Network Filtering

YANG AND KOLACZYK: TARGET DETECTION VIA NETWORK FILTERING 2515

[12] B. Efron, T. Hastie, I. Johnstone, and R. Tibshirani, “Least angle re-gression,” Ann. Statist., vol. 32, pp. 407–451, 2004.

[13] P. Erdös and A. Rényi, “Random graphs,” Math. Inst. Hungarian Acad.Sci., vol. 5, pp. 17–61, 1960.

[14] G. H. Golub and C. F. van Loan, Matrix Computations, 3rd ed. Bal-timore, MD: The Johns Hopkins Univ. Press, 1996.

[15] E. Greenshtein and Y. Ritov, “Persistence in high-dimensional linearpredictor selection and the virtue of overparametrization,” Bernoulli,vol. 10, no. 6, pp. 971–988, 2004.

[16] H. W. Guggenheimer, A. S. Edelman, and C. R. Johnson, “A simpleestimate of the condition number of a linear system,” College Math, J.,vol. 26, 1995.

[17] N. Krämer, On the Peaking Phenomenon of the Lasso in Model Selec-tion [Online]. Available: http://arxiv.org/abs/0904.4416 (unpublished).Available at

[18] B. Laurent and P. Massart, “Adaptive estimation of a quadratic func-tional by model selection,” Ann. Statist., vol. 28, no. 5, 2000.

[19] S. L. Lauritzen, Graphical Models. Oxford, U.K.: Oxford Univ.Press, 1996.

[20] M. Ledoux, “The concentration of measure phenomenon,” in Mathe-matical Surveys and Monographs 89. Providence, RI: Amer. Math.Soc., 2001.

[21] N. Meinshausen and P. Bühlmann, “High dimensional graphs andvariable selection with the Lasso,” Ann. Statist., vol. 34, no. 3, pp.1436–1462, 2006.

[22] A. Ostrowski, “Sur la détermination des bornes inférieures pour uneclasse des déterminants,” Bull. Sci. Math., vol. 61, pp. 19–32, 1937.

[23] B. Krishnamachari, Networking Wireless Sensors. Cambridge, U.K.:Cambridge Univ. Press, 2006.

[24] R. Tibshirani, “Regression shrinkage and selection via the lasso,” J.Royal Statist. Soc.: Ser. B, vol. 58, no. 1, pp. 267–288, 1996.

[25] S. Wasserman and K. Faust, Social Network Analysis: Methods andApplications. Cambridge, U.K.: Cambridge Univ. Press, 1994.

[26] C. Zhu, “Stable recovery of sparse signals via regularized minimiza-tion,” IEEE Trans. Inf. Theory, vol. 54, pp. 3364–3367, Jul. 2008.

Shu Yang received the B.S. degree from the Department of Artificial Intelli-gence and the M.S. degree from the School of Mathematical Science, PekingUniversity, Beijing, China, in 2004 and 2006, respectively.

Currently, she is working toward the Ph.D. degree with the Department ofMathematics and Statistics, Boston University, Boston, MA.

Eric D. Kolaczyk (SM’06) received the B.S. degree in mathematics from theUniversity of Chicago, Chicago, IL, in 1990, and the M.S. and Ph.D. degreesin statistics from Stanford University, Stanford, CA, in 1992 and 1994, respec-tively.

He was Assistant Professor of statistics at the University of Chicago from1994 to 1998, at which time he joined the Department of Mathematics and Sta-tistics, Boston University, where he is now Professor of Mathematics and Sta-tistics and Director of the Program in Statistics. He is also an affiliated facultymember with the Center for Biodynamics, the Program in Bioinformatics, andthe Division of Systems Engineering, Boston University, and has been a visitingfaculty member with Harvard University, Cambridge, MA, and l’Université deParis 7, France. Previously his research interests were in statistical multiscalemodeling. His current research interests are in the area of networks, with appli-cations in computer network traffic analysis, biological network modeling, andsocial networks. He is the author of a recent textbook on the topic “StatisticalAnalysis of Network Data: Methods and Models” (Springer, 2009).


Recommended