+ All Categories
Home > Documents > Bayesian Watermark Attacks · Ivo D. Shterev [email protected] Duke University, Durham NC 27705,...

Bayesian Watermark Attacks · Ivo D. Shterev [email protected] Duke University, Durham NC 27705,...

Date post: 18-Jan-2021
Category:
Upload: others
View: 2 times
Download: 0 times
Share this document with a friend
8
Bayesian Watermark Attacks Ivo D. Shterev [email protected] Duke University, Durham NC 27705, USA David B. Dunson [email protected] Duke University, Durham NC 27708, USA Abstract This paper presents an application of statis- tical machine learning to the field of water- marking. We propose a new attack model on additive spread-spectrum watermarking systems. The proposed attack is based on Bayesian statistics. We consider the scenario in which a watermark signal is repeatedly em- bedded in specific, possibly chosen based on a secret message bitstream, segments (signals) of the host data. The host signal can rep- resent a patch of pixels from an image or a video frame. We propose a probabilistic model that infers the embedded message bit- stream and watermark signal, directly from the watermarked data, without access to the decoder. We develop an efficient Markov chain Monte Carlo sampler for updating the model parameters from their conjugate full conditional posteriors. We also provide a variational Bayesian solution, which further increases the convergence speed of the algo- rithm. Experiments with synthetic and real image signals demonstrate that the attack model is able to correctly infer a large part of the message bitstream and obtain a very accurate estimate of the watermark signal. 1. Introduction Watermarking is the process of imperceptibly em- bedding a watermark signal into a host signal (au- dio segment, pixel patch from image or video frame). The watermark signal should only introduce tolera- ble distortion to the host signal and it should be recoverable by the intended receiver. Watermark- Appearing in Proceedings of the 29 th International Confer- ence on Machine Learning, Edinburgh, Scotland, UK, 2012. Copyright 2012 by the author(s)/owner(s). ing techniques differ by the way they modulate the host signal to embed information. There are two ma- jor classes of watermark embedding schemes, namely spread spectrum and quantization index modulation (QIM) (Chen & Wornell, 2001). Spread spectrum watermarking (Cox et al., 2007; Hartung et al., 1999) constitutes a popular class of watermarking algorithms. In their simplest form, the watermarked signal is constructed by adding the host and watermark signals together, i.e. additive water- mark embedding. Although, in terms of additive noise attacks they have been outperformed by the more ro- bust QIM watermarking techniques (Chen & Wornell, 2001), spread-spectrum techniques have advantageous features that make them preferable in some water- marking scenarios. Examples of such inherent features include their simplicity and robustness to removal at- tacks. Another advantage is that spread-spectrum wa- termarking can be applied in different forms (multi- plicative watermarking (Huang & Zhang, 2007)) that can further improve performance in some cases. They can also effectively exploit the human visual system (HVS) (Podilchuk & Zheng, 1998) to reduce percep- tual degradation of the host signal. Many attacks have been designed to hamper the performance of watermarking in general and spread- spectrum watermarking in particular. The attacks are usually classified with respect to the attacker’s as- sumed knowledge about the watermark scheme. Ro- bustness attacks pertain to the class of attacks under which the attacker has no knowledge of the watermark scheme. Examples of such attacks include adding ran- dom noise (Chen & Wornell, 2001) to the watermarked signal, replacing signal blocks with perceptually simi- lar blocks computed in a certain way (Kirovski et al., 2007), applying a geometric transformation (cropping, scaling, translation, etc.) to the watermarked signal, or applying a malicious filtering operation (Su et al., 2001), to name a few. Other attacks belong to the
Transcript
Page 1: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

Ivo D. Shterev [email protected]

Duke University, Durham NC 27705, USA

David B. Dunson [email protected]

Duke University, Durham NC 27708, USA

Abstract

This paper presents an application of statis-tical machine learning to the field of water-marking. We propose a new attack modelon additive spread-spectrum watermarkingsystems. The proposed attack is based onBayesian statistics. We consider the scenarioin which a watermark signal is repeatedly em-bedded in specific, possibly chosen based on asecret message bitstream, segments (signals)of the host data. The host signal can rep-resent a patch of pixels from an image ora video frame. We propose a probabilisticmodel that infers the embedded message bit-stream and watermark signal, directly fromthe watermarked data, without access to thedecoder. We develop an efficient Markovchain Monte Carlo sampler for updating themodel parameters from their conjugate fullconditional posteriors. We also provide avariational Bayesian solution, which furtherincreases the convergence speed of the algo-rithm. Experiments with synthetic and realimage signals demonstrate that the attackmodel is able to correctly infer a large partof the message bitstream and obtain a veryaccurate estimate of the watermark signal.

1. Introduction

Watermarking is the process of imperceptibly em-bedding a watermark signal into a host signal (au-dio segment, pixel patch from image or video frame).The watermark signal should only introduce tolera-ble distortion to the host signal and it should berecoverable by the intended receiver. Watermark-

Appearing in Proceedings of the 29 th International Confer-ence on Machine Learning, Edinburgh, Scotland, UK, 2012.Copyright 2012 by the author(s)/owner(s).

ing techniques differ by the way they modulate thehost signal to embed information. There are two ma-jor classes of watermark embedding schemes, namelyspread spectrum and quantization index modulation(QIM) (Chen & Wornell, 2001).

Spread spectrum watermarking (Cox et al., 2007;Hartung et al., 1999) constitutes a popular class ofwatermarking algorithms. In their simplest form, thewatermarked signal is constructed by adding the hostand watermark signals together, i.e. additive water-mark embedding. Although, in terms of additive noiseattacks they have been outperformed by the more ro-bust QIM watermarking techniques (Chen & Wornell,2001), spread-spectrum techniques have advantageousfeatures that make them preferable in some water-marking scenarios. Examples of such inherent featuresinclude their simplicity and robustness to removal at-tacks. Another advantage is that spread-spectrum wa-termarking can be applied in different forms (multi-plicative watermarking (Huang & Zhang, 2007)) thatcan further improve performance in some cases. Theycan also effectively exploit the human visual system(HVS) (Podilchuk & Zheng, 1998) to reduce percep-tual degradation of the host signal.

Many attacks have been designed to hamper theperformance of watermarking in general and spread-spectrum watermarking in particular. The attacksare usually classified with respect to the attacker’s as-sumed knowledge about the watermark scheme. Ro-bustness attacks pertain to the class of attacks underwhich the attacker has no knowledge of the watermarkscheme. Examples of such attacks include adding ran-dom noise (Chen & Wornell, 2001) to the watermarkedsignal, replacing signal blocks with perceptually simi-lar blocks computed in a certain way (Kirovski et al.,2007), applying a geometric transformation (cropping,scaling, translation, etc.) to the watermarked signal,or applying a malicious filtering operation (Su et al.,2001), to name a few. Other attacks belong to the

Page 2: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

so called worst case class of attacks, where the at-tacker has knowledge about the watermark techniqueand designs the attack such that the watermark detec-tor (decoder) performance is minimized, under suit-ably defined distortion constraints. Usually, this typeof attack is based on game theory (Cohen & Lapidoth,2002) and is mostly of theoretical importance.

A third class of attacks aims at compromisingthe watermark system security (security attacks)(Cayre et al., 2005; Freire & Gonzalez, 2009). Underthis scenario, the attacker has access only to the wa-termarked data and tries to estimate the secret keyused for embedding the watermark. Having estimatedthe secret key, he can then reconstruct the water-mark and remove it from the watermarked data (theso called removal attacks), thus creating a forgery ofthe host signal, which can then be freely copied anddistributed by pirates. Although (Cayre et al., 2005;Freire & Gonzalez, 2009) develop theoretical securityattack frameworks, the proposed algorithms do notperform well with real correlated host signals.

Another type of attack is the so called sensitiv-ity analysis attack, which constitutes a powerfulsubclass of removal attacks (Kalker et al., 1998;Linnartz & van Dijk, 1998; Choubassi & Moulin,2007). In their attempt to estimate the watermarksignal, they rely on unlimited access to the decoder.

In this paper, we consider the scenario in which awatermark signal is repeatedly embedded in specific(possibly secretly chosen) host signals. The host sig-nal can represent a patch of pixels from image orvideo frame. The host signals may be perceptu-ally similar or quite disparate, as the watermark al-gorithm may choose, for security reasons, to embedthe watermark in specific signals of the host databased on a secret message bitstream. Repetitive wa-termark embedding is of particular interest in imageand video watermarking (Voloshynovskiy et al., 2001;Lu & Hsu, 2007; Bas et al., 2002; Tang & Hang, 2003;Doerr & Dugelay, 2004; Kalker et al., 1999), wherethe watermark signal is repeatedly allocated into smallblocks to ensure robustness and resistance to geometric(desynchronization) attacks. However, the proposedattacks related to this scenario assume that the wa-termark signal is not secretly hidden but is added toevery host signal and therefore do not try to estimatean embedded message bitstream.

The attack model proposed in this paper jointly es-timates the embedded message bitstream and water-mark signal from the watermarked data, without ac-cess to the decoder. We develop a probabilistic modelbased on Bayesian statistics. The algorithm models

the host signal as having a multivariate Gaussian dis-tribution with unknown mean and full covariance ma-trix. The watermark signal itself is also modeled ashaving a multivariate Gaussian distribution, but withseparate unknown mean and full covariance matrix.The model parameters are updated sequentially fromtheir respective conjugate full conditional posteriordistributions, via Markov chain Monte Carlo (MCMC)sampling. To further increase the convergence speedof the proposed algorithm, we develop a variationalBayesian (VB) (Beal & Ghahramani, 2003) solution toit. In addition to its suitability for large scale dataanalysis, the VB solution also allows for diagnosingconvergence, via the lower bound to the log-likelihood.Both MCMC and VB solutions perform comparablywith respect to probability of bit error and relativewatermark reconstruction error, with both syntheticand real host data.

Our model borrows similar ideas from sparse factorregression formulations (sparse models) used in geneexpression data analysis (West, 2003; Carvalho et al.,2008). The objective of such sparse models is to spec-ify a prior for the elements of a highly sparse factorloadings matrix, with most elements being exactly zeroand few of them having relatively large variances. Tocontrast with our model, the role of the zero elementshere is taken by the data points (signals) that are notwatermarked, which do not necessarily constitute themajority of all data points. The watermarked datapoints have the interpretation of the non-zero elementsin the factor loadings matrix. However in our case,they are a sum of the host and the watermark signals,with the watermark signal being much weaker thanthe host signal. The problem becomes that of a jointidentification-estimation of a subtle signal.

2. Spread-Spectrum Watermarking

Throughout this paper, random variables are denotedby small letters. Random vectors and their realiza-tions are denoted by bold small letters. The notationx ∈ Rd indicates a d-dimensional random vector ofreal elements. Square random matrices and their real-izations are denoted by bold capital letters. The nota-tionX ∈ Rd×d indicates a d×d matrix of real elementsand X′ is its transpose. The probability of an eventis denoted by Pr(·). The notation x ∼ p(x) indicatesthat x has a probability density function (pdf) p(x).

In this paper we concentrate on one of the most pop-ular additive spread-spectrum watermarking systems,in which a watermark signal is repeatedly used to em-bed a message bitstream into a host data. The wa-termark encoder is shown in Fig. 1. Considering

Page 3: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

the ith data point and depending on the message bitbi ∈ {0, 1}, the encoder adds (bi = 1) the watermarksignal w to the host signal xi, or leaves the host sig-nal unchanged (bi = 0). The watermarked signal cantherefore be written as

yi =

{

xi +w if bi = 1,xi if bi = 0,

(1)

where i ∈ {1, . . . , n} and n is the number of availabledata points.

xi yi

w

Figure 1. Additive spread-spectrum watermark encoder.

The watermark decoder is shown in Fig. 2. The de-coder has access to the watermark signal w. Basedon the received (watermarked) signal yi and the wa-termark signal, the decoder computes a detection teststatistic f(yi,w) and compares it to a suitably chosen

threshold τ . The decoder then outputs an estimate biof the embedded message bit bi in the following way

bi =

{

1 if f(yi,w) > τ,0 if otherwise.

(2)

yi

f(yi,w)

w

bi

Figure 2. Watermark decoder.

Throughout the paper, the document-to-watermark(DWR) ratio is defined as DWR = 10 log10 σ

2x/σ

2w,

where σ2x is the variance of a single element in xi, for

i ∈ {1, . . . , n}, and σ2w is the variance of a single ele-

ment in w.

3. Attack Model

It is assumed that the attacker has access to the wa-termarked signal, but has no access to the watermarkdecoder. The complete form of the attack model can

be summarized as follows:

yi = xi + biw (3)

xi ∼ N (xi|µ,Σ) (4)

w ∼ N (w|m,V) (5)

bi ∼ Bernoulli(bi|π) (6)

{µ,Σ} ∼ N (µ|µ0,Σ)IW(Σ|ω0,Σ0) (7)

{m,V} ∼ N (m|m0,V)IW(V|ω0,V0) (8)

π ∼ Beta(π|aπ, bπ), (9)

where N (x|µ,Σ) is the d-variate Gaussian distribu-tion of x with mean µ and covariance matrix Σ,IW(Σ|ω0,Σ0) is the inverse Wishart distribution ofΣ with degrees of freedom ω0 and base covariance ma-trix Σ0, Bernoulli(bi|π) is the Bernoulli distribution ofbi with mean π, and Beta(π|aπ , bπ) is the Beta distri-bution of π with parameters aπ and bπ.

A graphical representation of the attack model isshown in Fig. 3. The blue circle represents theobserved variable, the white circles represent hidden(latent) variables and the squares represent hyper-parameters. Conditional dependence between vari-ables is shown via the directed edges.

π

bi

ω0 Σ0 aπ bπ ω0

V m

V0 m0

yi

w

Σ

xi

µ

µ0

Figure 3. A graphical representation of attack model.

4. Posterior Updates

In this section we derive the update equations for theattack model parameters, with respect to the MCMCand VB solutions. The update equations are based onthe full likelihood of the model, which can be writtenas

L(y) = p(y,x,w,µ,Σ,m,V,b, π)

=∏

i

p(yi|xi,w, bi)p(xi|µ,Σ)p(bi|π)

× p(w|m,V)p(µ,Σ)p(m,V)p(π) (10)

Page 4: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

As it can be seen the likelihood (10) is in an intractableform, since it is not possible to jointly estimate allmodel parameters directly from (10). That is why theMCMC and VB solutions developed below update ev-ery model parameter sequentially from its respectiveconditional posterior distribution. While the MCMCsolution is based on exact conditional posterior distri-butions, the VB solution utilizes an approximation tothe true conditional posterior distribution.

4.1. MCMC Update Equations

We derive the MCMC update equations, based on theexact full conditional posteriors of the attack modelparameters and construct a Gibbs sampler that itera-tively samples from these update equations.

The full conditional posterior distributions of themodel parameters are as follows:

• updating {µ,Σ}.

p(µ,Σ|x) ∝ p(µ,Σ)∏

i

p(xi|µ,Σ)

∝ IW(

Σ|ω0 + n,ΣΣ

)

N(

µ|µµ,Σ

n+ 1

)

, (11)

where

ΣΣ = Σ0 +n

n+ 1(x− µ0)(x− µ0)

+∑

i

(xi − x)(xi − x)′ (12)

µµ =µ0 + nx

n+ 1(13)

x =1

n

i

xi. (14)

• updating w.

p(w|y) ∝ p(w|m,V)∏

i

1(bi = 1)p(yi|µ,Σ,w)

∝ N(

w|mw,Vw

)

, (15)

where

Vw = (V−1 + n1Σ−1)−1 (16)

mw = Vw

(

V−1m

+ Σ−1∑

i

1(bi = 1)(yi − µ))

(17)

n1 =∑

i

1(bi = 1), (18)

and 1(·) is an indicator function.

• updating bi.

p(bi|πi) ∝ Bernoulli(bi|πi), (19)

where

πi =1

1 + 1−ππ

N (yi|µ,Σ)N (yi|µ+m,Σ+V)

. (20)

• updating π.

p(π|b) = Beta(π|aπ , bπ), (21)

where

aπ = aπ +∑

i

1(bi = 1), (22)

bπ = bπ +∑

i

1(bi = 0). (23)

• updating {m,V}.

p(m,V|w) ∝ p(m,V)p(w|m,V)

∝ IW(

V|ω0 + 1,Vv

)

N (m|mm,V

2), (24)

where

Vv = V0 +1

2(w −m0)(w −m0)

′ (25)

mm =m0 +w

2. (26)

• updating xi.

xi =

{

yi −w if bi = 1,yi if bi = 0

(27)

4.2. VB Update Equations

The VB approach tries to find a tractable lower boundL(q) to the logarithm of the marginal likelihood (10),which can be iteratively updated (tightened). If we de-note by θ the model parameters {w,µ,Σ,m,V,b, π}that we want to update, the optimal posterior updatethat gives the tightest bound (Beal & Ghahramani,2003) is given as

qj(θj) ∝ exp(

ln p(y, θ)⟩

−j

)

, (28)

where 〈·〉−j denotes expectation with respect to all pa-rameters except for the jth parameter that is beingupdated.

Using (28), the posterior VB updates of the modelparameters are as follows:

Page 5: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

• updating {µ,Σ}.

q(µ,Σ) ∝ exp⟨

ln(

p(µ,Σ)∏

i

p(xi|µ,Σ))

∝ IW(

Σ|ω0 + n,ΣvbΣ

)

N(

µ|µvbµ,〈Σ−1〉−1

n+ 1

)

, (29)

where

ΣvbΣ = Σ0 +

n

n+ 1(z+ µ0)(z + µ0)

+∑

i

(

〈b2i 〉〈ww′〉 − 〈bi〉2〈w〉〈w〉′

)

+∑

i

(

〈bi〉〈w〉 − yi − z)(〈bi〉〈w〉 − yi − z)′)

(30)

µvbµ

=µ0 − nz

n+ 1(31)

z =1

n

i

(

〈bi〉〈w〉 − yi

)

, (32)

and b2i = bi, following the properties of theBernoulli random variable.

• updating w.

q(w) ∝ exp⟨

ln(

p(w)∏

i

p(yi|w))

∝ N (w|mvbw ,Vvb

w ), (33)

where

Vvbw =

(

〈V−1〉+ 〈Σ−1〉

×∑

i

(

〈bi〉2 +

bi − 〈bi〉⟩2)

)−1

(34)

mvbw = Vvb

w

(

〈V−1〉〈m〉

+ 〈Σ−1〉∑

i

〈bi〉(

yi − 〈µ〉)

)

. (35)

• updating bi.

q(bi|πi) ∝ Bernoulli(bi|πi), (36)

where

πi =1

1 +exp

ln(1−π)⟩

N(

yi|〈µ〉,〈Σ−1〉−1

)

exp 〈lnπ〉N(

yi|〈µ〉+〈m〉,〈Σ−1〉−1+〈V−1〉−1

)

.

• updating π.

q(π|b) ∝ Beta(π|avbπ , bvbπ ), (37)

where

avbπ = aπ +∑

i

〈bi〉 (38)

bvbπ = bπ + n−∑

i

〈bi〉. (39)

• updating {m,V}.

q(m,V|w) ∝ exp⟨

ln(

p(m,V)p(w|m,V))

∝ IW(

V|ω0 + 1,Vvbv

)

N(

m|mvbm ,

〈V−1〉−1

2

)

, (40)

where

Vvbv = V0 +

1

2

(

m0 − 〈w〉)(

m0 − 〈w〉)′

(41)

mvbm =

m0 + 〈w〉

2. (42)

5. Experiments

We perform experiments with both synthetic and realhost signals. To quantify the performance of our al-gorithm, we compute the probability of error Pe =1n

i 1(bi 6= bi) and the relative watermark reconstruc-

tion error Rw = ‖w−w‖2

‖w‖2

, where ‖ · ‖2 is the L2 norm,

and w is the estimated watermark signal.

In all experiments, the model hyper-parameters areinitialized as aπ = bπ = 0.5n, µ0 = m0 = 1

n

i yi,ω0 = d + 1, Σ0 = 1

n

i(yi − µ0)(yi − µ0)′ and

V0 = 110DWR/10Σ0. For each iteration, we computed

95% credible intervals of the individual samples in thewatermark signal estimate w. With respect to theMCMC solution, we performed 2000 iterations of theGibbs sampler, discarding the first 1000 as burn initerations and averaging the results of the remaining1000 iterations. For the VB solution, we performed100 iterations and using the last iteration updates of〈w〉, and 〈bi〉, as the estimated watermark signal w

and message bit bi for i ∈ {1, . . . , n}, respectively.

We implemented the proposed attack model solu-tions in R, with some of the routines implementedin C/C++. It takes approximately 6 minutes for theMCMC solution to perform 2000 iterations, using 409664-dimensional data points. In contrast, the VB so-lution performs 100 iterations in less than a minute,using the same data points.

5.1. Synthetic Host Signals

In this subsection we perform experiments with syn-thetic host signals. We generated n = 4096, d = 64-dimensional host data points. Each data point was in-

dependent and identically drawn from N(

0, IW(

d +

1, IW(d + 1, I))

)

, where I is the identity matrix. In

this way, the host signal covariance matrix, althoughrandomly drawn, imposes some structure on the hostsignal. The watermark signal was drawn from a mul-tivariate Gaussian with mean N

(

0, IW(d+1, I))

and

Page 6: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

covariance IW(d + 1, I). The watermark signal waszero mean transformed and scaled so that DWR =30db. The watermark message bits were drawn fromBernoulli(0.5), and the watermarked signal was formedby additive spread-spectrum modulation. The hostimage and the difference between the watermarked andhost images are shown in Fig. 4. Each block of pix-els was formed by row-wise transformation of the datapoint into an 8 × 8 matrix. The blocks were then or-dered row-wise to form the whole image.

Figure 4. Synthetic host image (left) and watermark signalmodulated by the message bits (right). DWR = 30db.

Experimental results of the difference bi − bi for i ∈{1, . . . , n}, the watermark signal w and its estimate wfor the MCMC and VB solutions are shown in Fig. 5.The results show that the algorithm was able to obtaina good estimate of the watermark signal and messagebitstream, with only a small fraction of misidentifiedbits. Based on the experimental results, we can seethat the MCMC and VB solutions perform compara-bly in terms of Pe and Rw.

0 1000 2000 3000 4000

−1.

00.

01.

0

b−b

0 1000 2000 3000 4000

−1.

00.

01.

0

b−b

0 10 20 30 40 50 60

−0.

040.

02

w,w

0 10 20 30 40 50 60

−0.

030.

01

w,w

Figure 5. Experimental results for the MCMC (left col-umn) and VB (right column) solutions applied to the syn-thetic host signal. The plots present bi − bi, the water-mark signal w (solid line) and its estimate w (dashed line).The gray regions represent 95% credible intervals. For theMCMC solution Pe = 0.004 and Rw = 0.48. For the VBsolution Pe = 0 and Rw = 0.256. Chosen DWR = 30db.

5.2. Real Host Signals

In this subsection we perform experiments with realimage signals. We applied our algorithm on gray scaleimages. In the experiments we used image sizes of512 × 512 and 1024 × 1024 pixels. The images weresplit in 8 × 8 patches of pixels, making a total ofn = 4096 and n = 16384 patches respectively. Thepixels within each patch were concatenated row-wiseto form the d-dimensional (d = 64) data points. Thehost signal was then normalized to have zero mean.

As in the case of synthetic host signals, the water-mark signal was drawn from a multivariate Gaus-sian with mean N

(

0, IW(d + 1, I))

and covarianceIW(d + 1, I). The watermark signal was then scaledsuch that DWR = 30db and embedded by additivespread-spectrum modulation with Pr(bi = 1) = 0.5,for i ∈ {1 . . . n}. The real host images used in theexperiments are shown in Fig. 6.

Figure 6. Real host images. From top left to bottom right:Lake, Boat, Children, Fruits, Lena, Pirate.

Experimental results for the MCMC and VB solutionsapplied to the real host images from Fig. 6 are shownin Fig. 7. It can be seen that both solutions performcomparably with respect to real host signals.

0 10 20 30 40 50 60

−4

02

4

w,w

0 10 20 30 40 50 60

−4

02

4

w,w

0 10 20 30 40 50 60

−3

−1

13

w,w

0 10 20 30 40 50 60

−3

−1

13

w,w

0 10 20 30 40 50 60

−3

−1

1

w,w

0 10 20 30 40 50 60

−3

−1

1

w,w

0 10 20 30 40 50 60

−4

02

4

w,w

0 10 20 30 40 50 60

−4

02

4

w,w

0 10 20 30 40 50 60

−3

02

w,w

0 10 20 30 40 50 60

−3

02

w,w

0 10 20 30 40 50 60

−4

02

w,w

0 10 20 30 40 50 60

−4

02

w,w

Figure 7. Experimental results for the MCMC (left col-umn) and VB (right column) solutions. The solid line rep-resents the watermark signal w and the dashed line is itsestimate w. The gray regions represent 95% credible in-tervals. From top to bottom: Lake, Boat, Children, Fruits,Lena, Pirate. Chosen DWR = 30db.

Computations of the lower bound L(q) for the VB so-lution applied on the real host images in Fig. 6 areshown in Fig. 8. The results show that the VB so-

Page 7: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

lution converges in less than 20 iterations for all realhost images considered in the experiments.

0 20 40 60 80 100−69

1500

−68

9500

iteration

L(q)

0 20 40 60 80 100

−64

1000

iteration

L(q)

0 20 40 60 80 100−57

4500

iteration

L(q)

0 20 40 60 80 100−58

3000

−57

9000

iteration

L(q)

0 20 40 60 80 100

−56

1000

iteration

L(q)

0 20 40 60 80 100

−25

1400

0

iteration

L(q)

Figure 8. Experimental results for the lower bound to thelog-likelihood. Please see supplementary for a detailed ex-pression of L(q). From top left to bottom right: Lake, Boat,Children, Fruits, Lena, Pirate. Chosen DWR = 30db.

To quantify the performance of the attack model withrespect to different DWR levels, we perform experi-ments with the real host images in Fig. 6 and vary-ing the DWR ∈ {20, . . . , 40}. The interval of valuesfor the DWR was chosen so that the middle is atDWR = 30db, at which level no perceptual degrada-tion to the host image was observed. The watermarksignal was drawn from a multivariate Gaussian withmean N

(

0, IW(d+1, I))

and covariance IW(d+1, I).For each host image, the watermark signal was drawnonly once and then zero mean transformed, and scaleddown differently to achieve the different DWR levels.Experimental results of Pe and Rw as functions ofDWR are shown in Fig. 9 and Fig. 10 respectively,with both solutions performing comparably.

6. Discussion

We presented a new attack model on repetitive spread-spectrum watermarking systems, based on Bayesianstatistics. The proposed algorithm jointly estimatesthe watermark signal and message bitstream, directlyfrom the watermarked signal and without access to thewatermark decoder. We developed MCMC and VBsolutions that perform comparably in terms of prob-ability of error and relative watermark reconstructionerror, on both synthetic and real host signals. Fastconvergence is observed in both solutions, particularlyin the VB solution where the algorithm converges inless than 20 iterations, with both synthetic and realhost signals. While the MCMC solution is expected toresult in more accurate estimates for infinite numberof iterations, the VB solution is computationally moreefficient and therefore more appropriate for large datasets. We demonstrated that the attack model is able

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

20 25 30 35 40

0.0

0.3

DWR

Pe

Figure 9. Experimental results of Pe, based on the MCMC(left column) and VB (right column) solutions. From topto bottom: Lake, Boat, Children, Fruits, Lena, Pirate.

to correctly infer a large part of the message bitstreamwhile at the same time obtaining a good estimate ofthe watermark signal.

7. Acknowledgments

The authors wish to acknowledge the helpful sugges-tions of the reviewers.

References

Bas, P., Chassery, J. M., and Macq, B. Geometricallyinvariant watermarking using feature points. IEEE

Transactions on Image Processing, 11(9):1014–1028,September 2002.

Beal, M. J. and Ghahramani, Z. The variationalbayesian em algorithm for incomplete data: withapplication to scoring graphical model structures.Bayesian Statistics 7, pp. 453–464, 2003.

Carvalho, C. M., Chang, J., Lucas, J. E., Nevins, J. R.,Wang, Q., and West, M. High-dimensional sparsefactor modeling: Applications in gene expression ge-nomics. Journal of the American Statistical Associ-

ation, 103(484):1438–1456, March 2008.

Cayre, F., Fontaine, C., and Furon, T. Watermark-ing security: Theory and practice. IEEE Transac-

tions on Signal Processing, 53(10):3976–3987, Octo-ber 2005.

Page 8: Bayesian Watermark Attacks · Ivo D. Shterev i.shterev@duke.edu Duke University, Durham NC 27705, USA David B. Dunson dunson@stat.duke.edu Duke University, Durham NC 27708, USA Abstract

Bayesian Watermark Attacks

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWRR

w

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

20 25 30 35 40

0.0

1.0

DWR

Rw

Figure 10. Experimental results of Rw, based on MCMC(left column) and VB (right column) solutions. From topto bottom: Lake, Boat, Children, Fruits, Lena, Pirate.

Chen, B. and Wornell, G. W. Quantization indexmodulation: A class of provably good methods fordigital watermarking and information embedding.IEEE Transactions on Information Theory, 47(4):1423–1443, May 2001.

Choubassi, M. El and Moulin, P. Noniterative algo-rithms for sensitivity analysis attacks. IEEE Trans-

actions on Information Forensics and Security, 2(2):113–126, June 2007.

Cohen, A. S. and Lapidoth, A. The gaussian water-marking game. IEEE Transactions on Information

Theory, 48(6):1639–1667, June 2002.

Cox, I. J., Miller, M. L., Bloom, J. A., Fridrich, J.,and Kalker, T. Digital Watermarking and Steganog-

raphy, 2nd Ed. Morgan Kaufmann, 2007.

Doerr, G. and Dugelay, J. L. Security pitfalls offrame-by-frame approaches to video watermarking.IEEE Transactions on Signal Processing, 52(10):2955–2964, October 2004.

Freire, L. P. and Gonzalez, F. P. Spread-spectrumwatermarking security. IEEE Transactions on In-

formation Forensics and Security, 4(1):2–24, March2009.

Hartung, F., Su, J. K., and Girod, B. Spread spec-trum watermarking: Malicious attacks and counter-

attacks. In Proc. SPIE Security and Watermarking

of Multimedia Contents, San Jose, CA, 1999.

Huang, X. and Zhang, B. Statistically robust detec-tion of multiplicative spread-spectrum watermarks.IEEE Transactions on Information Forensics and

Security, 2(1):1–13, March 2007.

Kalker, T., Linnartz, J. P., and van Dijk, M. Wa-termark estimation through detection analysis. InProc. International Conference on Image Process-

ing, Chicago, IL, 1998.

Kalker, T., Depovere, G., Haitsma, J., and Maes, M.A video watermarking system for broadcast moni-toring. In Proc. SPIE Security and Watermarking

of Multimedia Contents, San Jose, CA 1999.

Kirovski, D., Petitcolas, F. A. P., and Landau, Z. Thereplacement attack. IEEE Transactions on Audio,

Speech, and Language Processing, 15(6):1922–1931,August 2007.

Linnartz, J. P. and van Dijk, M. Analysis of the sen-sitivity attack against electronic watermarks in im-ages. In Proc. Workshop on Information Hiding,Portland, OR, 1998.

Lu, C. S. and Hsu, C. Y. Near-optimal watermark esti-mation and its countermeasure: Antidisclosure wa-termark for multiple watermark embedding. IEEE

Transactions on Circuits and Systems for Video

Technology, 17(4):454–467, April 2007.

Podilchuk, C. I. and Zheng, W. Image-adaptive water-marking using visual models. IEEE Transactions on

Selected Areas in Communications, 16(4):525–539,May 1998.

Su, J. K., Eggers, J. J., and Girod, B. Analysis ofdigital watermarks subjected to optimum linear fil-tering and additive noise. Signal Processing, 81(6):1141–1175, June 2001.

Tang, C. W. and Hang, H. M. A feature-based robustdigital image watermarking scheme. IEEE Trans-

actions on Signal Processing, 51(4):950–959, April2003.

Voloshynovskiy, S., Deguillaume, F., and Pun, T.Multibit digital watermarking robust against localnonlinear geometrical distortions. In Proc. Inter-

national Conference on Image Processing, Thessa-loniki, Greece, 2001.

West, M. Bayesian factor regression models in thelarge p, small n paradigm. Bayesian Statistics 7,pp. 723–732, 2003.


Recommended