+ All Categories
Home > Documents > arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and...

arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and...

Date post: 16-Jul-2020
Category:
Upload: others
View: 1 times
Download: 0 times
Share this document with a friend
12
A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad 1 , Ahmed Roman 2 , Mahesh C. Shastry 3 , Mihir Gadgil 4 and Ayanendranath Basu 5 1 Department of Physics, Indian Institute of Technology Gandhinagar, Ahmedabad, Gujarat, INDIA 2 Department of Mathematics, Virginia Tech , Blacksburg, VA, USA. 3 Department of Physics, Indian Institute of Science Education and Research Bhopal, Bhopal, Madhya Pradesh, INDIA. 4 Biomedical Engineering Department, Oregon Health & Science University, Portland, OR, USA. 5 Indian Statistical Institute, Kolkata, West Bengal-700108, INDIA [email protected], [email protected], [email protected], [email protected], [email protected] Keywords: Divergence Measures, Bhattacharyya Distance, Error Probability, F-divergence, Pattern Recognition, Signal Detection, Signal Classification. Abstract: We introduce a new one-parameter family of divergence measures, called bounded Bhattacharyya distance (BBD) measures, for quantifying the dissimilarity between probability distributions. These measures are bounded, symmetric and positive semi-definite and do not require absolute continuity. In the asymptotic limit, BBD measure approaches the squared Hellinger distance. A generalized BBD measure for multiple distributions is also introduced. We prove an extension of a theorem of Bradt and Karlin for BBD relating Bayes error probability and divergence ranking. We show that BBD belongs to the class of generalized Csiszar f-divergence and derive some properties such as curvature and relation to Fisher Information. For distributions with vector valued parameters, the curvature matrix is related to the Fisher-Rao metric. We derive certain inequalities between BBD and well known measures such as Hellinger and Jensen-Shannon divergence. We also derive bounds on the Bayesian error probability. We give an application of these measures to the problem of signal detection where we compare two monochromatic signals buried in white noise and differing in frequency and amplitude. 1 INTRODUCTION Divergence measures for the distance between two probability distributions are a statistical approach to comparing data and have been extensively studied in the last six decades [Kullback and Leibler, 1951, Ali and Silvey, 1966, Kapur, 1984, Kullback, 1968, Ku- mar et al., 1986]. These measures are widely used in varied fields such as pattern recognition [Basseville, 1989, Ben-Bassat, 1978, Choi and Lee, 2003], speech recognition [Qiao and Minematsu, 2010, Lee, 1991], signal detection [Kailath, 1967, Kadota and Shepp, 1967,Poor, 1994], Bayesian model validation [Tumer and Ghosh, 1996] and quantum information theory [Nielsen and Chuang, 2000, Lamberti et al., 2008]. Distance measures try to achieve two main objectives (which are not mutually exclusive): to assess (1) how “close” two distributions are compared to others and (2) how “easy” it is to distinguish between one pair than the other [Ali and Silvey, 1966]. There is a plethora of distance measures available to assess the convergence (or divergence) of proba- bility distributions. Many of these measures are not metrics in the strict mathematical sense, as they may not satisfy either the symmetry of arguments or the triangle inequality. In applications, the choice of the measure depends on the interpretation of the metric in terms of the problem considered, its analytical prop- erties and ease of computation [Gibbs and Su, 2002]. One of the most well-known and widely used di- vergence measures, the Kullback-Leibler divergence (KLD) [Kullback and Leibler, 1951, Kullback, 1968], can create problems in specific applications. Specif- ically, it is unbounded above and requires that the distributions be absolutely continuous with respect to each other. Various other information theoretic measures have been introduced keeping in view ease of computation ease and utility in problems of sig- nal selection and pattern recognition. Of these mea- sures, Bhattacharyya distance [Bhattacharyya, 1946, Kailath, 1967, Nielsen and Boltz, 2011] and Chernoff arXiv:1201.0418v9 [math.ST] 10 Apr 2016
Transcript
Page 1: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

A New Family of Bounded Divergence Measures and Application toSignal Detection

Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir Gadgil4 and Ayanendranath Basu 5

1Department of Physics, Indian Institute of Technology Gandhinagar, Ahmedabad, Gujarat, INDIA2 Department of Mathematics, Virginia Tech , Blacksburg, VA, USA.

3 Department of Physics, Indian Institute of Science Education and Research Bhopal, Bhopal, Madhya Pradesh, INDIA.4 Biomedical Engineering Department, Oregon Health & Science University, Portland, OR, USA.

5 Indian Statistical Institute, Kolkata, West Bengal-700108, [email protected], [email protected], [email protected], [email protected], [email protected]

Keywords: Divergence Measures, Bhattacharyya Distance, Error Probability, F-divergence, Pattern Recognition, SignalDetection, Signal Classification.

Abstract: We introduce a new one-parameter family of divergence measures, called bounded Bhattacharyya distance(BBD) measures, for quantifying the dissimilarity between probability distributions. These measures arebounded, symmetric and positive semi-definite and do not require absolute continuity. In the asymptoticlimit, BBD measure approaches the squared Hellinger distance. A generalized BBD measure for multipledistributions is also introduced. We prove an extension of a theorem of Bradt and Karlin for BBD relatingBayes error probability and divergence ranking. We show that BBD belongs to the class of generalized Csiszarf-divergence and derive some properties such as curvature and relation to Fisher Information. For distributionswith vector valued parameters, the curvature matrix is related to the Fisher-Rao metric. We derive certaininequalities between BBD and well known measures such as Hellinger and Jensen-Shannon divergence. Wealso derive bounds on the Bayesian error probability. We give an application of these measures to the problemof signal detection where we compare two monochromatic signals buried in white noise and differing infrequency and amplitude.

1 INTRODUCTION

Divergence measures for the distance between twoprobability distributions are a statistical approach tocomparing data and have been extensively studied inthe last six decades [Kullback and Leibler, 1951, Aliand Silvey, 1966, Kapur, 1984, Kullback, 1968, Ku-mar et al., 1986]. These measures are widely used invaried fields such as pattern recognition [Basseville,1989, Ben-Bassat, 1978, Choi and Lee, 2003], speechrecognition [Qiao and Minematsu, 2010, Lee, 1991],signal detection [Kailath, 1967, Kadota and Shepp,1967,Poor, 1994], Bayesian model validation [Tumerand Ghosh, 1996] and quantum information theory[Nielsen and Chuang, 2000, Lamberti et al., 2008].Distance measures try to achieve two main objectives(which are not mutually exclusive): to assess (1) how“close” two distributions are compared to others and(2) how “easy” it is to distinguish between one pairthan the other [Ali and Silvey, 1966].

There is a plethora of distance measures available

to assess the convergence (or divergence) of proba-bility distributions. Many of these measures are notmetrics in the strict mathematical sense, as they maynot satisfy either the symmetry of arguments or thetriangle inequality. In applications, the choice of themeasure depends on the interpretation of the metric interms of the problem considered, its analytical prop-erties and ease of computation [Gibbs and Su, 2002].One of the most well-known and widely used di-vergence measures, the Kullback-Leibler divergence(KLD) [Kullback and Leibler, 1951,Kullback, 1968],can create problems in specific applications. Specif-ically, it is unbounded above and requires that thedistributions be absolutely continuous with respectto each other. Various other information theoreticmeasures have been introduced keeping in view easeof computation ease and utility in problems of sig-nal selection and pattern recognition. Of these mea-sures, Bhattacharyya distance [Bhattacharyya, 1946,Kailath, 1967, Nielsen and Boltz, 2011] and Chernoff

arX

iv:1

201.

0418

v9 [

mat

h.ST

] 1

0 A

pr 2

016

Page 2: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

distance [Chernoff, 1952, Basseville, 1989, Nielsenand Boltz, 2011] have been widely used in signalprocessing. However, these measures are again un-bounded from above. Many bounded divergence mea-sures such as Variational, Hellinger distance [Bas-seville, 1989, DasGupta, 2011] and Jensen-Shannonmetric [Burbea and Rao, 1982,Rao, 1982b,Lin, 1991]have been studied extensively. Utility of these mea-sures vary depending on properties such as tightnessof bounds on error probabilities, information theoreticinterpretations, and the ability to generalize to multi-ple probability distributions.

Here we introduce a new one-parameter (α) fam-ily of bounded measures based on the Bhattacharyyacoefficient, called bounded Bhattacharyya distance(BBD) measures. These measures are symmet-ric, positive-definite and bounded between 0 and 1.In the asymptotic limit (α → ±∞) they approachsquared Hellinger divergence [Hellinger, 1909,Kaku-tani, 1948]. Following Rao [Rao, 1982b] and Lin[Lin, 1991], a generalized BBD is introduced to cap-ture the divergence (or convergence) between multi-ple distributions. We show that BBD measures belongto the generalized class of f-divergences and inherituseful properties such as curvature and its relation toFisher Information. Bayesian inference is useful inproblems where a decision has to be made on clas-sifying an observation into one of the possible arrayof states, whose prior probabilities are known [Hell-man and Raviv, 1970, Varshney and Varshney, 2008].Divergence measures are useful in estimating the er-ror in such classification [Ben-Bassat, 1978, Kailath,1967, Varshney, 2011]. We prove an extension of theBradt Karlin theorem for BBD, which proves the exis-tence of prior probabilities relating Bayes error prob-abilities with ranking based on divergence measure.Bounds on the error probabilities Pe can be calcu-lated through BBD measures using certain inequal-ities between Bhattacharyya coefficient and Pe. Wederive two inequalities for a special case of BBD(α = 2) with Hellinger and Jensen-Shannon diver-gences. Our bounded measure with α = 2 has beenused by Sunmola [Sunmola, 2013] to calculate dis-tance between Dirichlet distributions in the context ofMarkov decision process. We illustrate the applicabil-ity of BBD measures by focusing on signal detectionproblem that comes up in areas such as gravitationalwave detection [Finn, 1992]. Here we consider dis-criminating two monochromatic signals, differing infrequency or amplitude, and corrupted with additivewhite noise. We compare the Fisher Information ofthe BBD measures with that of KLD and Hellingerdistance for these random processes, and highlight theregions where FI is insensitive large parameter devia-

tions. We also characterize the performance of BBDfor different signal to noise ratios, providing thresh-olds for signal separation.

Our paper is organized as follows: Section I isthe current introduction. In Section II, we recall thewell known Kullback-Leibler and Bhattacharyya di-vergence measures, and then introduce our boundedBhattacharyya distance measures. We discuss somespecial cases of BBD, in particular Hellinger distance.We also introduce the generalized BBD for multi-ple distributions. In Section III, we show the posi-tive semi-definiteness of BBD measure, applicabilityof the Bradt Karl theorem and prove that BBD be-longs to generalized f-divergence class. We also de-rive the relation between curvature and Fisher Infor-mation, discuss the curvature metric and prove someinequalities with other measures such as Hellingerand Jensen Shannon divergence for a special case ofBBD. In Section IV, we move on to discuss applica-tion to signal detection problem. Here we first brieflydescribe basic formulation of the problem, and thenmove on computing distance between random pro-cesses and comparing BBD measure with Fisher In-formation and KLD. In the Appendix we provide theexpressions for BBD measures , with α = 2, for somecommonly used distributions. We conclude the paperwith summary and outlook.

2 DIVERGENCE MEASURES

In the following subsection we consider a measur-able space Ω with σ-algebra B and the set of all prob-ability measures M on (Ω,B). Let P and Q denoteprobability measures on (Ω,B) with p and q denotingtheir densities with respect to a common measure λ.We recall the definition of absolute continuity [Roy-den, 1986]:

Absolute Continuity A measure P on the Borel sub-sets of the real line is absolutely continuous with re-spect to Lebesgue measure Q, if P(A) = 0, for everyBorel subset A ∈ B for which Q(A) = 0, and is de-noted by P << Q.

2.1 Kullback-Leibler divergence

The Kullback-Leibler divergence (KLD) (or rela-tive entropy) [Kullback and Leibler, 1951, Kullback,1968] between two distributions P,Q with densities pand q is given by:

I(P,Q)≡∫

p log(

pq

)dλ. (1)

Page 3: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

The symmetrized version is given by

J(P,Q)≡ (I(P,Q)+ I(Q,P))/2

[Kailath, 1967], I(P,Q) ∈ [0,∞]. It diverges if∃ x0 : q(x0) = 0 and p(x0) 6= 0.

KLD is defined only when P is absolutely contin-uous w.r.t. Q. This feature can be problematic in nu-merical computations when the measured distributionhas zero values.

2.2 Bhattacharyya Distance

Bhattacharyya distance is a widely used measurein signal selection and pattern recognition [Kailath,1967]. It is defined as:

B(P,Q)≡− ln(∫ √

pqdλ

)=− ln(ρ), (2)

where the term in parenthesis ρ(P,Q) ≡∫ √

pqdλ

is called Bhattacharyya coefficient [Bhattacharyya,1943, Bhattacharyya, 1946] in pattern recognition,affinity in theoretical statistics, and fidelity in quan-tum information theory. Unlike in the case of KLD,the Bhattacharyya distance avoids the requirement ofabsolute continuity. It is a special case of Chernoffdistance

Cα(P,Q)≡− ln(∫

pα(x)q1−α(x)dx),

with α = 1/2. For discrete probability distribu-tions, ρ ∈ [0,1] is interpreted as a scalar product ofthe probability vectors P = (

√p1,√

p2, . . . ,√

pn) andQ = (

√q1,√

q2, . . . ,√

qn). Bhattacharyya distanceis symmetric, positive-semidefinite, and unbounded(0 ≤ B ≤ ∞). It is finite as long as there exists someregion S⊂ X such that whenever x∈ S : p(x)q(x) 6= 0.

2.3 Bounded Bhattacharyya DistanceMeasures

In many applications, in addition to the desirableproperties of the Bhattacharyya distance, bounded-ness is required. We propose a new family of boundedmeasure of Bhattacharyya distance as below,

Bψ,b(P,Q)≡− logb(ψ(ρ)) (3)

where, ρ = ρ(P,Q) is the Bhattacharyya coefficient,ψb(ρ) satisfies ψ(0) = b−1 , ψ(1) = 1. In particularwe choose the following form :

ψ(ρ) =

[1− (1−ρ)

α

b =

α−1

, (4)

where α ∈ [−∞,0)∪ (1,∞]. This gives the measure

Bα(ρ(P,Q))≡− log(1− 1

α )−α

[1− (1−ρ)

α

, (5)

which can be simplified to

Bα(ρ) =log[1− (1−ρ)

α

]log[1− 1

α

] . (6)

It is easy to see that Bα(0) = 1, Bα(1) = 0.

2.4 Special cases

1. For α = 2 we get,

B2(ρ) =− log22

[1+ρ

2

]2

=− log2

(1+ρ

2

).

(7)We study some of its special properties in Sec.3.7.

2. α→ ∞

B∞(ρ) =− loge e−(1−ρ) = 1−ρ = H2(ρ), (8)

where H(ρ) is the Hellinger distance [Basseville,1989, Kailath, 1967, Hellinger, 1909, Kakutani,1948]

H(ρ)≡√

1−ρ(P,Q). (9)

3. α =−1

B−1(ρ) =− log2

(1

2−ρ

). (10)

4. α→−∞

B−∞(ρ) = loge e(1−ρ) = 1−ρ = H2(ρ). (11)

We note that BBD measures approach squaredHellinger distance when α→ ±∞. In general, theyare convex (concave) when α > 1 (α < 0) in ρ, asseen by evaluating second derivative

∂2Bα(ρ)

∂ρ2 =−1

α2 log(1− 1

α

)(1− 1−ρ

α

)2 =

=

> 0 α > 1< 0 α < 0 .

(12)

From this we deduce Bα>1(ρ)≤H2(ρ)≤Bα<0(ρ) forρ ∈ [0,1]. A comparison between Hellinger and BBDmeasures for different values of α are shown in Fig.1.

Page 4: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

0 0.2 0.4 0.6 0.8 10

0.2

0.4

0.6

0.8

1

ρ

Bα(ρ)

α=−10

−1

1.01

2

10

H2(ρ)

Figure 1: [Color Online] Comparison of Hellinger andbounded Bhattacharyya distance measures for different val-ues of α.

2.5 Generalized BBD measure

In decision problems involving more than two ran-dom variables, it is very useful to have divergencemeasures involving more than two distributions [Lin,1991, Rao, 1982a, Rao, 1982b]. We use the general-ized geometric mean (G) concept to define boundedBhattacharyya measure for more than two distribu-tions. The Gβ(pi) of n variables p1, p2, . . . , pn withweights β1,β2, . . . ,βn, such that βi ≥ 0, ∑i βi = 1, isgiven by

Gβ(pi) =n

∏i=1

pβii .

For n probability distributions P1,P2, . . . ,Pn , withdensities p1, p2, . . . , pn , we define a generalized Bhat-tacharyya coefficient, also called Matusita measure ofaffinity [Matusita, 1967, Toussaint, 1974]:

ρβ(P1,P2, . . . ,Pn) =∫

Ω

n

∏i=1

pβii dλ. (13)

where βi ≥ 0, ∑i βi = 1. Based on this, we define thegeneralized bounded Bhattacharyya measures as:

α(ρβ(P1,P2, . . . ,Pn))≡log(1− 1−ρβ

α)

log(1−1/α)(14)

where α∈ [−∞,0)∪(1,∞]. For brevity we denote it asBβ

α(ρ). Note that, 0≤ ρβ≤ 1 and 0≤Bβ

α(ρ)≤ 1, sincethe weighted geometric mean is maximized when allthe pi’s are the same, and minimized when any twoof the probability densities pi’s are perpendicular toeach other.

3 PROPERTIES

3.1 Symmetry, Boundedness andPositive Semi-definiteness

Theorem 3.1. Bα(ρ(P,Q)) is symmetric, positivesemi-definite and bounded in the interval [0,1] forα ∈ [−∞,0)∪ (1,∞].

Proof. Symmetry: Since ρ(P,Q) = ρ(Q,P), it fol-lows that

Bα(ρ(P,Q)) = Bα(ρ(Q,P)).

Positive-semidefinite and boundedness: SinceBα(0) = 1, Bα(1) = 0 and

∂Bα(ρ)

∂ρ=

1α log(1−1/α) [1− (1−ρ)/α]

< 0

for 0≤ ρ≤ 1 and α ∈ [−∞,0)∪ (1,∞], it follows that

0≤ Bα(ρ)≤ 1. (15)

3.2 Error Probability and DivergenceRanking

Here we recap the definition of error probability andprove the applicability of Bradt and Karlin [Bradt andKarlin, 1956] theorem to BBD measure.Error probability: The optimal Bayes error proba-bilities (see eg: [Ben-Bassat, 1978, Hellman and Ra-viv, 1970,Toussaint, 1978]) for classifying two eventsP1,P2 with densities p1(x) and p2(x) with prior prob-abilities Γ = π1,π2 is given by

Pe =∫

min[π1 p1(x),π2 p2(x)]dx. (16)

Error comparison: Let pβ

i (x) (i = 1,2) be pa-rameterized by β (Eg: in case of Normal distributionβ = µ1,σ1;µ2,σ2 ). In signal detection literature,a signal set β is considered better than set β′ for thedensities pi(x) , when the error probability is less forβ than for β′ (i.e. Pe(β)< Pe(β

′)) [Kailath, 1967].

Divergence ranking: We can also rank theparameters by means of some divergence D. Thesignal set β is better (in the divergence sense) than β′,if Dβ(P1,P2)> Dβ′(P1,P2).

In general it is not true that Dβ(P1,P2) >Dβ′(P1,P2) =⇒ Pe(β) < Pe(β

′). Bradt and Karlinproved the following theorem relating error probabil-ities and divergence ranking for symmetric KullbackLeibler divergence J:

Page 5: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

Theorem 3.2 (Bradt and Karlin [Bradt and Karlin,1956]). If Jβ(P1,P2) > Jβ′(P1,P2), then ∃ a set ofprior probabilities Γ = π1,π2 for two hypothesisg1,g2, for which

Pe(β,Γ)< Pe(β′,Γ) (17)

where Pe(β,Γ) is the error probability with parameterβ and prior probability Γ.

It is clear that the theorem asserts existence, butno method of finding these prior probabilities. Kailath[Kailath, 1967] proved the applicability of the BradtKarlin Theorem for Bhattacharyya distance measure.We follow the same route and show that the Bα(ρ)measure satisfies a similar property using the follow-ing theorem by Blackwell.

Theorem 3.3 (Blackwell [Blackwell, 1951]).Pe(β

′,Γ) ≤ Pe(β,Γ) for all prior probabilities Γ ifand only if

Eβ′ [Φ(Lβ′)|g]≤ Eβ[Φ(Lβ)|g],

∀ continuous concave functions Φ(L), whereLβ = p1(x,β)/p2(x,β) is the likelihood ratio andEω[Φ(Lω)|g] is the expectation of Φ(Lω) under thehypothesis g = P2.

Theorem 3.4. If Bα(ρ(β)) > Bα(ρ(β′)), or equiva-

lently ρ(β)< ρ(β′) then ∃ a set of prior probabilitiesΓ = π1,π2 for two hypothesis g1,g2, for which

Pe(β,Γ)< Pe(β′,Γ). (18)

Proof. The proof closely follows Kailath [Kailath,1967]. First note that

√L is a concave function of

L (likelihood ratio) , and

ρ(β) = ∑x∈X

√p1(x,β)p2(x,β)

= ∑x∈X

√p1(x,β)p2(x,β)

p2(x,β)

= Eβ

[√Lβ|g2

]. (19)

Similarly

ρ(β′) = Eβ′

[√Lβ′ |g2

](20)

Hence, ρ(β)< ρ(β′)⇒

[√Lβ|g2

]< Eβ′

[√Lβ′ |g2

]. (21)

Suppose assertion of the stated theorem is not true,then for all Γ, Pe(β

′,Γ)≤ Pe(β,Γ). Then by Theorem3.3, Eβ′ [Φ(Lβ′)|g2]≤Eβ[Φ(Lβ)|g2] which contradictsour result in Eq. 21.

3.3 Bounds on Error Probability

Error probabilities are hard to calculate in general.Tight bounds on Pe are often extremely useful in prac-tice. Kailath [Kailath, 1967] has shown bounds on Pein terms of the Bhattacharyya coefficient ρ:

12

[2π1−

√1−4π1π2ρ2

]≤Pe≤

(π1−

12

)+√

π1π2ρ,

(22)with π1 +π2 = 1. If the priors are equal π1 = π2 =

12 ,

the expression simplifies to

12

[1−√

1−ρ2

]≤ Pe ≤

12

ρ. (23)

Inverting relation in Eq. 6 for ρ(Bα), we can get thebounds in terms of Bα(ρ) measure. For the equal priorprobabilities case, Bhattacharyya coefficient gives atight upper bound for large systems when ρ→ 0 (zerooverlap) and the observations are independent andidentically distributed. These bounds are also usefulto discriminate between two processes with arbitrar-ily low error probability [Kailath, 1967]. We supposethat tighter upper bounds on error probability can bederived through Matusita’s measure of affinity [Bhat-tacharya and Toussaint, 1982, Toussaint, 1977, Tous-saint, 1975], but is beyond the scope of present work.

3.4 f-divergence

A class of divergence measures called f-divergenceswere introduced by Csiszar [Csiszar, 1967, Csiszar,1975] and independently by Ali and Silvey [Ali andSilvey, 1966] (see [Basseville, 1989] for review).It encompasses many well known divergence mea-sures including KLD, variational, Bhattacharyya andHellinger distance. In this section, we show thatBα(ρ) measure for α ∈ (1,∞], belongs to the genericclass of f-divergences defined by Basseville [Bas-seville, 1989].

f-divergence [Basseville, 1989] Consider a measur-able space Ω with σ-algebra B . Let λ be a measureon (Ω,B) such that any probability laws P and Q areabsolutely continuous with respect to λ, with densitiesp and q. Let f be a continuous convex real function onR+, and g be an increasing function on R. The classof divergence coefficients between two probabilities:

d(P,Q) = g(∫

Ω

f(

pq

)qdλ

)(24)

are called the f-divergence measure w.r.t. functions( f ,g) . Here p/q = L is the likelihood ratio. The termin the parenthesis of g gives the Csiszar’s [Csiszar,1967, Csiszar, 1975] definition of f-divergence.

Page 6: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

The Bα(ρ(P,Q)) , for α ∈ (1,∞] measure can be writ-ten as the following f divergence:

f (x) =−1+1−√

, g(F) =log(−F)

log(1−1/α), (25)

where,

F =∫

Ω

[−1+

(1−√

pq

)]qdλ

=∫

Ω

[q(−1+

)− 1

α

√pq]

= −1+1−ρ

α. (26)

and

g(F) =log(1− 1−ρ

α)

log(1−1/α)= Bα(ρ(P,Q)). (27)

3.5 Curvature and Fisher Information

In statistics, the information that an observable ran-dom variable X carries about an unknown parameter θ

(on which it depends) is given by the Fisher informa-tion. One of the important properties of f-divergenceof two distributions of the same parametric familyis that their curvature measures the Fisher informa-tion. Following the approach pioneered by Rao [Rao,1945], we relate the curvature of BBD measures tothe Fisher information and derive the differential cur-vature metric. The discussions below closely follow[DasGupta, 2011].

Definition Let f (x|θ);θ ∈ Θ ⊆ R, be a family ofdensities indexed by real parameter θ, with some reg-ularity conditions ( f (x|θ) is absolutely continuous).

Zθ(φ)≡ Bα(θ,φ) =log(1− 1−ρ(θ,φ)

α)

log(1−1/α)(28)

where ρ(θ,φ) =∫√

f (x|θ) f (x|φ)dx

Theorem 3.5. Curvature of Zθ(φ)|φ=θ is the Fisherinformation of f (x|θ) up to a multiplicative constant.

Proof. Expand Zθ(φ) around theta

Zθ(φ)=Zθ(θ)+(φ−θ)dZθ(φ)

dφ+(φ−θ)2

2d2Zθ(φ)

dφ2 +. . .

(29)Let us observe some properties of Bhattacharyya co-efficient : ρ(θ,φ) = ρ(φ,θ), ρ(θ,θ) = 1, and itsderivatives:

∂ρ(θ,φ)

∂φ

∣∣∣φ=θ

=12

∂θ

∫f (x|θ)dx = 0, (30)

∂2ρ(θ,φ)

∂φ2

∣∣∣φ=θ

=− 14

∫ 1f

(∂ f∂θ

)2

dx+12

∂2

∂θ2

∫f dx

=− 14

∫f (x|θ)

(∂ log f (x|θ)

∂θ

)2

dx

=− 14

I f (θ). (31)

where I f (θ) is the Fisher Information of distributionf (x|θ)

I f (θ) =∫

f (x|θ)(

∂ log f (x|θ)∂θ

)2

dx. (32)

Using the above relationships, we can write downthe terms in the expansion of Eq. 29 Zθ(θ) =

0 , ∂Zθ(φ)∂φ

∣∣∣φ=θ

= 0, and

∂2Zθ(φ)

∂φ2

∣∣∣φ=θ

=C(α)I f (θ)> 0, (33)

where C(α) = −14α log(1−1/α) > 0

The leading term of Bα(θ,φ) is given by

Bα(θ,φ)∼(φ−θ)2

2C(α)I f (θ). (34)

3.6 Differential Metrics

Rao [Rao, 1987] generalized the Fisher informationto multivariate densities with vector valued parame-ters to obtain a “geodesic” distance between two para-metric distributions Pθ,Pφ of the same family. TheFisher-Rao metric has found applications in many ar-eas such as image structure and shape analysis [May-bank, 2004, Peter and Rangarajan, 2006] , quantumstatistical inference [Brody and Hughston, 1998] andBlackhole thermodynamics [Quevedo, 2008]. We de-rive such a metric for BBD measure using property off-divergence.

Let θ,φ ∈ Θ ⊆ Rp, then using the fact that∂Z(θ,φ)

∂θi

∣∣∣φ=θ

= 0, we can easily show that

dZθ =p

∑i, j=1

∂2Zθ

∂θi∂θ jdθidθ j+ · · ·=

p

∑i, j=1

gi jdθidθ j+ . . . .

(35)The curvature metric gi j can be used to find thegeodesic on the curve η(t), t ∈ [0,1] with

C = η(t) : η(0) = θ η(1) = φ. (36)

Details of the geodesic equation are given in manystandard differential geometry books. In the con-text of probability distance measures reader is re-ferred to (see 15.4.2 in A DasGupta [DasGupta, 2011]

Page 7: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

for details) The curvature metric of all Csiszar f-divergences are just scalar multiple KLD measure[DasGupta, 2011, Basseville, 1989] given by:

g fi j(θ) = f ′′(1)gi j(θ). (37)

For our BBD measure

f ′′(x) =

(−1+

1−√

)′′=

14αx3/2

f ′′(1) = 1/4α. (38)

Apart from the −1/ log(1− 1α), this is same as C(α)

in Eq. 34. It follows that the geodesic distance for ourmetric is same KLD geodesic distance up to a multi-plicative factor. KLD geodesic distances are tabulatedin DasGupta [DasGupta, 2011].

3.7 Relation to other measures

Here we focus on the special case α = 2, i.e. B2(ρ)

Theorem 3.6.

B2 ≤ H2 ≤ log4 B2 (39)

where 1 and log4 are sharp.

Proof. Sharpest upper bound is achieved via takingsupρ∈[0,1)

H2(ρ)B2(ρ)

. Define

g(ρ) ≡ 1−ρ

− log2 (1+ρ)/2. (40)

We note that g(ρ) is continuous and has no singulari-ties whenever ρ ∈ [0,1). Hence

g′(ρ) =1−ρ

1+ρ+ log( 1+ρ

2 )

log2 ρ+12

log2≥ 0.

It follows that g(ρ) is non-decreasing and hencesupρ∈[0,1)g(ρ) = limρ→1 g(ρ) = log(4). Thus

H2/B2 ≤ log4. (41)

Combining this with convexity property of Bα(ρ) forα > 1, we get

B2 ≤ H2 ≤ log4 B2

Using the same procedure we can prove a generic ver-sion of this inequality for α ∈ (1,∞] , given by

Bα(ρ)≤ H2 ≤−α log(

1− 1α

)Bα(ρ) (42)

Jensen-Shannon Divergence: The Jensen differ-ence between two distributions P,Q, with densitiesp,q and weights (λ1,λ2); λ1 +λ2 = 1, is defined as,

Jλ1,λ2(P,Q) = Hs(λ1 p+λ2q)−λ1Hs(p)−λ2Hs(q),(43)

where Hs is the Shannon entropy. Jensen-Shannondivergence (JSD) [Burbea and Rao, 1982,Rao, 1982b,Lin, 1991] is based on the Jensen difference and isgiven by:

JS(P,Q) =J1/2,1/2(P,Q)

=12

∫ [p log

(2p

p+q

)+q log

(2q

p+q

)]dλ (44)

The structure and goals of JSD and BBD measuresare similar. The following theorem compares the twometrics using Jensen’s inequality.

Lemma 3.7. Jensen’s Inequality: For a convex func-tion ψ, E[ψ(X)]≥ ψ(E[X ]).

Theorem 3.8 (Relation to Jensen-Shannon measure).JS(P,Q)≥ 2

log2 B2(P,Q)− log2

We use the un-symmetrized Jensen-Shannon met-ric for the proof.

Proof.

JS(P,Q) =∫

p(x) log2p(x)

p(x)+q(x)dx

=−2∫

p(x) log

√p(x)+q(x)√

2p(x)dx

≥−2∫

p(x) log

√p(x)+

√q(x)√

2p(x)dx

(since√

p+q≤√p+√

q)

=EP

[−2log

√p(X)+

√q(X)√

2p(X)

]

By Jensen’s inequalityE[− log f (X)]≥− logE[ f (X)], we have

EP

[−2log

√p(X)+

√q(X)√

2p(X)

]≥

−2logEP

[√p(X)+

√q(X)√

2p(X)

].

Page 8: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

Hence,

JS(P,Q) ≥ −2log∫

p(x)

(√p(x)+

√q(x)

)√

2p(x)dx

= −2log

(1+

∫√p(x)q(x)2

)− log2

= 2(

B2(p(x),q(x))log2

)− log2

=2

log2B2(P,Q)− log2. (45)

4 APPLICATION TO SIGNALDETECTION

Signal detection is a common problem occurringin many fields such as communication engineering,pattern recognition, and Gravitational wave detection[Poor, 1994]. In this section, we briefly describe theproblem and terminology used in signal detection. Weillustrate though simple cases how divergence mea-sures, in particular BBD can be used for discriminat-ing and detecting signals buried in white noise of cor-relator receivers (matched filter). For greater detailsof the formalism used we refer the reader to reviewarticles in the context of Gravitational wave detectionby Jaranowski and Krolak [Jaranowski and Krolak,2007] and Sam Finn [Finn, 1992].

One of the central problem in signal detection is todetect whether a deterministic signal s(t) is embeddedin an observed data x(t), corrupted by noise n(t). Thiscan be posed as a hypothesis testing problem wherethe null hypothesis is absence of signal and alternativeis its presence. We take the noise to be additive , sothat x(t) = n(t)+ s(t). We define the following termsused in signal detection: Correlation G (also calledmatched filter) between x and s, and signal to noiseratio ρ [Finn, 1992, Budzynski et al., 2008].

G = (x|s), ρ =√(s,s), (46)

where the scalar product (.|.) is defined by

(x|y) := 4ℜ

∫∞

0

x( f )y∗( f )N( f )

d f . (47)

ℜ denotes the real part of a complex expression, tildedenotes the Fourier transform and the asterisk * de-notes complex conjugation. N is the one-sided spec-tral density of the noise.For white noise, the probability densities of G when

respectively signal is present and absent are given by[Budzynski et al., 2008]

p1(G) =1√2πρ

exp(− (G−ρ2)2

2ρ2

), (48)

p0(G) =1√2πρ

exp(− G2

2ρ2

)(49)

4.1 Distance between Gaussianprocesses

Consider a stationary Gaussian random process X,which has signals s1 or s2 with probability densitiesp1 and p2 respectively of being present in it. Thesedensities follow the form Eq. 48 with signal to noiseratios ρ2

1 and ρ22 respectively. The probability den-

sity p(X) of Gaussian process can modeled as limitof multivariate Gaussian distributions. The diver-gence measures between these processes d(s1,s2) arein general functions of the correlator (s1− s2|s1− s2)[Budzynski et al., 2008]. Here we focus on distin-guishing monochromatic signal s(t) = Acos(ωt + φ)and filter sF(t) = AF cos(ωF t + φ) (both buried innoise), separated in frequency or amplitude.

The Kullback-Leibler divergence between the sig-nal and filter I(s,sF) is given by the correlation (s−sF |s− sF):

I(s,sF) =(s− sF |s− sF) = (s|s)+(sF |sF)−2(s|sF)

=ρ2 +ρ

2F −2ρρF [〈cos(∆ωt)〉cos(∆φ)

−〈sin(∆ωt)〉sin(∆φ)], (50)

where 〈〉 is the average over observation time [0,T ].Here we have assumed that noise spectral densityN( f ) = N0 is constant over the frequencies [ω,ωF ].The SNRs are given by

ρ2 =

A2TN0

, ρ2F =

A2F TN0

. (51)

(for detailed discussions we refer the reader toBudzynksi et. al [Budzynski et al., 2008]).

The Bhattacharyya distance between Gaussianprocessors with signals of same energy is ( Eq 14 in[Kailath, 1967]) just a multiple of the KLD B = I/8.We use this result to extract the Bhattacharyya coeffi-cient :

ρ(s,sF) = exp(− (s− sF |s− sF)

8

)(52)

4.1.1 Frequency difference

Let us consider the case when the SNRs of signal andfilter are equal, phase difference is zero, but frequen-cies differ by ∆ω. The KL divergence is obtained by

Page 9: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

evaluating the correlator in Eq. 50

I(∆ω) = (s− sF |s− sF) = 2ρ2(

1− sin(∆ωT )∆ωT

).

(53)by noting 〈cos(∆ωt)〉 = sin(∆ωT )

∆ωT and 〈sin(∆ωt)〉 =1−cos(∆ωT )

∆ωT . Using this, the expression for BBD familycan be written as

Bα(∆ω) =

log(

1− 1α

[1− e−

ρ24

(1− sin(∆ωT )

∆ωT

)])log(1− 1

α

) .

(54)As we have seen in section 3.4, both BBD and KLDbelong to the f-divergence family. Their curvature fordistributions belonging to same parametric family is aconstant times the Fisher information (FI) (see Theo-rem: 3.5). Here we discuss where the BBD and KLDdeviates from FI, when we account for higher termsin the expansion of these measures.

The Fisher matrix element for frequency gω,ω =

E[(

∂ logΛ

∂ω

)2]= ρ2T 2/3 [Budzynski et al., 2008],

where Λ is the likelihood ratio. Using the relation forline element ds2 = ∑i, j gi jdθidθ j and noting that onlyfrequency is varied, we get

ds =ρT ∆ω√

3. (55)

Using the relation between curvature of BBD measureand Fisher’s Information in Eq.34, we can see that forlow frequency differences the line element varies as:√

2Bα(∆ω)

C(α)∼ ds.

Similarly√

dKL ∼ ds at low frequencies. However, athigher frequencies both KLD and BBD deviate fromthe Fisher information metric. In Fig. 2, we have plot-ted ds,

√dKL and

√2Bα(∆ω)/C(α) with α = 2 and

Hellinger distance (α→∞) for ∆ω ∈ (0,0.1). We ob-serve that till ∆ω = 0.01 (i.e. ∆ωT ∼ 1), KLD andBBD follows Fisher Information and after that theystart to deviate. This suggests that Fisher Informa-tion is not sensitive to large deviations. There is notmuch difference between KLD, BBD and Hellingerfor large frequencies due to the correlator G becom-ing essentially a constant over a wide range of fre-quencies.

4.1.2 Amplitude difference

We now consider the case where the frequency andphase of the signal and the filter are same but they dif-fer in amplitude ∆A (which reflects in differing SNR).

Figure 2: Comparison of Fisher Information, KLD, BBDand Hellinger distance for two monochromatic signals dif-fering by frequency ∆ω, buried in white noise. Inset showswider range ∆ω ∈ (0,1) . We have set ρ = 1 and chosenparameters T = 100 and N0 = 104.

Figure 3: Comparison of Fisher information line elementwith KLD, BBD and Hellinger distance for signals differingin amplitude and buried in white noise. We have set A = 1,T = 100 and N0 = 104.

The correlation reduces to

(s− sF |s− sF) =A2TN0

+(A+∆A)2T

N0−2

A(A+∆A)TN0

=(∆A)2T

N0. (56)

This gives us I(∆A) = (∆A)2TN0

, which is the sameas the line element ds2 with Fisher metric ds =√

T/2N0∆A. In Fig. 3, we have plotted ds,√

dKL

and√

2Bα(∆ω)/C(α) for ∆A ∈ (0,40). KLD and FIline element are the same. Deviations of BBD and

Page 10: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

Hellinger can be observed only for ∆A > 10.Discriminating between two signals s1,s2 requires

minimizing the error probability between them. ByTheorem 3.4, there exists priors for which the prob-lem translates into maximizing the divergence forBBD measures. For the monochromatic signals dis-cussed above, the distance depends on parameters(ρ1,ρ2,∆ω,∆φ). We can maximize the distance for agiven frequency difference by differentiating with re-spect to phase difference ∆φ [Budzynski et al., 2008].In Fig. 4, we show the variation of maximized BBDfor different signal to noise ratios (ρ1,ρ2), for a fixedfrequency difference ∆ω = 0.01. The intensity mapshows different bands which can be used for settingthe threshold for signal separation.

Detecting signal of known form involves minimiz-ing the distance measure over the parameter space ofthe signal. A threshold on the maximum “distance”between the signal and filter can be put so that a detec-tion is said to occur whenever the measures fall withinthis threshold. Based on a series of tests, Receiver Op-erating Characteristic (ROC) curves can be drawn tostudy the effectiveness of the distance measure in sig-nal detection. We leave such details for future work.

Figure 4: BBD with different signal to noise ratio for a fixed. We have set T = 100 and ∆ω = 0.01.

5 SUMMARY AND OUTLOOK

In this work we have introduced a new familyof bounded divergence measures based on the Bhat-tacharyya distance, called bounded Bhattacharyyadistance measures. We have shown that it belongsto the class of generalized f-divergences and inher-its all its properties, such as those relating Fishers In-formation and curvature metric. We have discussed

several special cases of our measure, in particularsquared Hellinger distance, and studied relation withother measures such as Jensen-Shannon divergence.We have also extended the Bradt Karlin theorem onerror probabilities to BBD measure. Tight bounds onBayes error probabilities can be put by using proper-ties of Bhattacharyya coefficient.

Although many bounded divergence measureshave been studied and used in various applications, nosingle measure is useful in all types of problems stud-ied. Here we have illustrated an application to signaldetection problem by considering “distance” betweenmonochromatic signal and filter buried in white Gaus-sian noise with differing frequency or amplitude, andcomparing it to Fishers Information and Kullback-Leibler divergence.

A detailed study with chirp like signal and colorednoise occurring in Gravitational wave detection willbe taken up in a future study. Although our measureshave a tunable parameter α, here we have focused ona special case with α = 2. In many practical appli-cations where extremum values are desired such asminimal error, minimal false acceptance/rejection ra-tio etc, exploring the BBD measure by varying α maybe desirable. Further, the utility of BBD measures isto be explored in parameter estimation based on min-imal disparity estimators and Divergence informationcriterion in Bayesian model selection [Basu and Lind-say, 1994]. However, since the focus of the currentpaper is introducing a new measure and studying itsbasic properties, we leave such applications to statis-tical inference and data processing to future studies.

ACKNOWLEDGEMENTS

One of us (S.J) thanks Rahul Kulkarni for insight-ful discussions, Anand Sengupta for discussions onapplication to signal detection, and acknowledge thefinancial support in part by grants DMR-0705152 andDMR-1005417 from the US National Science Foun-dation. M.S. would like to thank the Penn State Elec-trical Engineering Department for support.

Page 11: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

REFERENCES

Ali, S. M. and Silvey, S. D. (1966). A general class of co-efficients of divergence of one distribution from an-other. Journal of the Royal Statistical Society. SeriesB (Methodological), 28(1):131–142.

Basseville, M. (1989). Distance measures for signal pro-cessing and pattern recognition. Signal processing,18:349–369.

Basu, A. and Lindsay, B. G. (1994). Minimum disparityestimation for continuous models: efficiency, distri-butions and robustness. Annals of the Institute of Sta-tistical Mathematics, 46(4):683–705.

Ben-Bassat, M. (1978). f-entropies, probability of er-ror, and feature selection. Information and Control,39(3):227–242.

Bhattacharya, B. K. and Toussaint, G. T. (1982). An upperbound on the probability of misclassification in termsof matusita’s measure of affinity. Annals of the Insti-tute of Statistical Mathematics, 34(1):161–165.

Bhattacharyya, A. (1943). On a measure of divergencebetween two statistical populations defined by theirprobability distributions. Bull. Calcutta Math. Soc,35(99-109):4.

Bhattacharyya, A. (1946). On a measure of divergence be-tween two multinomial populations. Sankhya: The In-dian Journal of Statistics (1933-1960), 7(4):401–406.

Blackwell, D. (1951). Comparison of experiments. In Sec-ond Berkeley Symposium on Mathematical Statisticsand Probability, volume 1, pages 93–102.

Bradt, R. and Karlin, S. (1956). On the design and compar-ison of certain dichotomous experiments. The Annalsof mathematical statistics, pages 390–409.

Brody, D. C. and Hughston, L. P. (1998). Statistical geom-etry in quantum mechanics. Proceedings of the RoyalSociety of London. Series A: Mathematical, Physicaland Engineering Sciences, 454(1977):2445–2475.

Budzynski, R. J., Kondracki, W., and Krolak, A. (2008).Applications of distance between probability distribu-tions to gravitational wave data analysis. Classicaland Quantum Gravity, 25(1):015005.

Burbea, J. and Rao, C. R. (1982). On the convexity ofsome divergence measures based on entropy func-tions. IEEE Transactions on Information Theory,28(3):489 – 495.

Chernoff, H. (1952). A measure of asymptotic efficiency fortests of a hypothesis based on the sum of observations.The Annals of Mathematical Statistics, 23(4):pp. 493–507.

Choi, E. and Lee, C. (2003). Feature extraction basedon the Bhattacharyya distance. Pattern Recognition,36(8):1703–1709.

Csiszar, I. (1967). Information-type distance measuresand indirect observations. Stud. Sci. Math. Hungar,2:299–318.

Csiszar, I. (1975). I-divergence geometry of probability dis-tributions and minimization problems. The Annals ofProbability, 3(1):pp. 146–158.

DasGupta, A. (2011). Probability for Statistics and Ma-chine Learning. Springer Texts in Statistics. SpringerNew York.

Finn, L. S. (1992). Detection, measurement, and gravita-tional radiation. Physical Review D, 46(12):5236.

Gibbs, A. and Su, F. (2002). On choosing and boundingprobability metrics. International Statistical Review,70(3):419–435.

Hellinger, E. (1909). Neue begrundung der theo-rie quadratischer formen von unendlichvielenveranderlichen. Journal fur die reine und angewandteMathematik (Crelle’s Journal), (136):210–271.

Hellman, M. E. and Raviv, J. (1970). Probability of Error,Equivocation, and the Chernoff Bound. IEEE Trans-actions on Information Theory, 16(4):368–372.

Jaranowski, P. and Krolak, A. (2007). Gravitational-wavedata analysis. formalism and sample applications: thegaussian case. arXiv preprint arXiv:0711.1115.

Kadota, T. and Shepp, L. (1967). On the best finite setof linear observables for discriminating two gaussiansignals. IEEE Transactions on Information Theory,13(2):278–284.

Kailath, T. (1967). The Divergence and Bhattacharyya Dis-tance Measures in Signal Selection. IEEE Transac-tions on Communications, 15(1):52–60.

Kakutani, S. (1948). On equivalence of infinite productmeasures. The Annals of Mathematics, 49(1):214–224.

Kapur, J. (1984). A comparative assessment of various mea-sures of directed divergence. Advances in Manage-ment Studies, 3(1):1–16.

Kullback, S. (1968). Information theory and statistics. NewYork: Dover, 1968, 2nd ed., 1.

Kullback, S. and Leibler, R. A. (1951). On informationand sufficiency. The Annals of Mathematical Statis-tics, 22(1):pp. 79–86.

Kumar, U., Kumar, V., and Kapur, J. N. (1986). Somenormalized measures of directed divergence. Inter-national Journal of General Systems, 13(1):5–16.

Lamberti, P. W., Majtey, A. P., Borras, A., Casas, M., andPlastino, A. (2008). Metric character of the quan-tum Jensen-Shannon divergence . Physical Review A,77:052311.

Lee, Y.-T. (1991). Information-theoretic distortion mea-sures for speech recognition. Signal Processing, IEEETransactions on, 39(2):330–335.

Lin, J. (1991). Divergence measures based on the shannonentropy. IEEE Transactions on Information Theory,37(1):145 –151.

Matusita, K. (1967). On the notion of affinity of several dis-tributions and some of its applications. Annals of theInstitute of Statistical Mathematics, 19(1):181–192.

Maybank, S. J. (2004). Detection of image structures usingthe fisher information and the rao metric. IEEE Trans-actions on Pattern Analysis and Machine Intelligence,26(12):1579–1589.

Nielsen, F. and Boltz, S. (2011). The burbea-rao and bhat-tacharyya centroids. IEEE Transactions on Informa-tion Theory, 57(8):5455–5466.

Page 12: arXiv:1201.0418v9 [math.ST] 10 Apr 2016 · A New Family of Bounded Divergence Measures and Application to Signal Detection Shivakumar Jolad1, Ahmed Roman2, Mahesh C. Shastry 3, Mihir

Nielsen, M. and Chuang, I. (2000). Quantum computationand information. Cambridge University Press, Cam-bridge, UK, 3(8):9.

Peter, A. and Rangarajan, A. (2006). Shape analysis usingthe fisher-rao riemannian metric: Unifying shape rep-resentation and deformation. In Biomedical Imaging:Nano to Macro, 2006. 3rd IEEE International Sympo-sium on, pages 1164–1167. IEEE.

Poor, H. V. (1994). An introduction to signal detection andestimation. Springer.

Qiao, Y. and Minematsu, N. (2010). A study on in-variance of-divergence and its application to speechrecognition. Signal Processing, IEEE Transactionson, 58(7):3884–3890.

Quevedo, H. (2008). Geometrothermodynamics of blackholes. General Relativity and Gravitation, 40(5):971–984.

Rao, C. (1982a). Diversity: Its measurement, decomposi-tion, apportionment and analysis. Sankhya: The In-dian Journal of Statistics, Series A, pages 1–22.

Rao, C. R. (1945). Information and the accuracy attainablein the estimation of statistical parameters. Bull. Cal-cutta Math. Soc., 37:81–91.

Rao, C. R. (1982b). Diversity and dissimilarity coefficients:A unified approach. Theoretical Population Biology,21(1):24 – 43.

Rao, C. R. (1987). Differential metrics in probabilityspaces. Differential geometry in statistical inference,10:217–240.

Royden, H. (1986). Real analysis. Macmillan PublishingCompany, New York.

Sunmola, F. T. (2013). Optimising learning with trans-ferable prior information. PhD thesis, University ofBirmingham.

Toussaint, G. T. (1974). Some properties of matusita’s mea-sure of affinity of several distributions. Annals of theInstitute of Statistical Mathematics, 26(1):389–394.

Toussaint, G. T. (1975). Sharper lower bounds for discrimi-nation information in terms of variation (corresp.). In-formation Theory, IEEE Transactions on, 21(1):99–100.

Toussaint, G. T. (1977). An upper bound on the probabilityof misclassification in terms of the affinity. Proceed-ings of the IEEE, 65(2):275–276.

Toussaint, G. T. (1978). Probability of error, expecteddivergence and the affinity of several distributions.IEEE Transactions on Systems, Man and Cybernetics,8(6):482–485.

Tumer, K. and Ghosh, J. (1996). Estimating the Bayeserror rate through classifier combining. Proceedingsof 13th International Conference on Pattern Recogni-tion, pages 695–699.

Varshney, K. R. (2011). Bayes risk error is a bregman di-vergence. IEEE Transactions on Signal Processing,59(9):4470–4472.

Varshney, K. R. and Varshney, L. R. (2008). Quanti-zation of prior probabilities for hypothesis testing.IEEE TRANSACTIONS ON SIGNAL PROCESSING,56(10):4553.

APPENDIX

BBD measures of some common distributionsHere we provide explicit expressions for BBD B2,

for some common distributions. For brevity we de-note ζ≡ B2.

• Binomial :

P(k) =(n

k

)pk(1− p)n−k, Q(k) =

(nk

)qk(1−q)n−k.

ζbin(P,Q)=− log2

(1+[√

pq+√(1− p)(1−q)]n

2

).

(57)

• Poisson :

P(k) =λk

pe−λp

k! , Q(k) =λk

qe−λq

k! .

ζpoisson(P,Q) =− log2

(1+ e−(

√λp−√

λq)2/2

2

).

(58)

• Gaussian :

P(x) =1√

2πσpexp

(−(x− xp)

2

2σ2p

),

Q(x) =1√

2πσqexp

(−(x− xq)

2

2σ2q

).

ζGauss(P,Q)= 1−log2

[1+

2σpσq

σ2p +σ2

qexp

(−

(xp− xq)2

4(σ2p +σ2

q)

)].

(59)

• Exponential : P(x) = λpe−λpx, Q(x) = λqe−λqx.

ζexp(P,Q) =− log2

[(√

λp +√

λq)2

2(λp +λq)

]. (60)

• Pareto : Assuming the same cut off xm,

P(x) =

αp

xαpm

xαp+1 for x≥ xm

0 for x < xm,(61)

Q(x) =

αq

xαqm

xαq+1 for x≥ xm

0 for x < xm.(62)

ζpareto(P,Q) =− log2

[(√

αp +√

αq)2

2(αp +αq)

]. (63)


Recommended