Date post: | 18-Nov-2023 |
Category: |
Documents |
Upload: | nottingham |
View: | 0 times |
Download: | 0 times |
Self-Validating Binary Document Images
By
Thesis Submitted to The University of Nottingham
for the Degree of Doctor of Philosophy
School of Computer Science
September 2009
Ammar Albakaa
Abstract
Binary printed documents are used in a wide variety of applications, for example
contracts, wills, recommendation letters, etc. Such documents are intrinsically
valuable but as a results may be subject to attacks by forgers who would make
deliberate changes to the documents, for example changing a name or a date in a
contract. However, during transmission (e.g. printing, in the mail, when being
scanned), they will be subject to noise, for example spurious marks or errors in
printing may occur. This thesis presents a methodology for use in the verification of
such documents. The methods used seek to reject deliberately forged documents
while at the same time not rejecting those which have only been altered by the
addition of noise. The thesis demonstrates that current techniques used for other
types of document are either unacceptably expensive or intrinsically unsuited to
binary documents. The methods developed have been tested on a range of
deliberately forged documents as well as on valid documents and are shown to be
robust. Limitations in the methods have been identified and solutions to the
problems are proposed.
Contents
Table of Contents
List of Tables ...................................................................................... 1
List of Figures ..................................................................................... 3
List of Abbreviations ............................................................................ 5
Acknowledgements .............................................................................. 7
Chapter One: Introduction .................................................................... 8
1.1 Overview ...................................................................................... 8
1.2 The Project Objective ..................................................................... 9
1.2.1 The Significance of Alternations .............................................. 10
1.2.2 The Location of Changes ........................................................ 10
1.2.3 The Representation of Preservative Data .................................. 11
1.2.4 The Insertion of Preservative Data .......................................... 11
1.2.5 The Sensitivity of the Verifier ................................................. 12
1.3 Overview of Chapters .................................................................... 12
Chapter Two: Literature Review ........................................................... 14
2.1 Introduction ................................................................................. 14
2.2 Methods to Protect Documents ....................................................... 16
2.3 Data Hiding Techniques ................................................................. 16
2.3.1 Steganography ..................................................................... 17
2.3.2 Digital Watermarking ............................................................ 17
2.3.3 Data Hiding Applications ........................................................ 18
2.4 Digital Watermarks ....................................................................... 20
2.4.1 Watermarking Applications ..................................................... 20
2.4.2 Watermark Visibility .............................................................. 21
2.4.3. Watermarking System .......................................................... 21
2.4.4. Watermarks Robustness ....................................................... 22
2.4.4.1. Fragile Watermarks .......................................................... 23
2.4.4.2. Semi-Fragile Watermarks .................................................. 23
2.4.4.3. Robust Watermarks .......................................................... 24
2.4.5 Imperceptibility vs Robustness ............................................... 24
2.4.6 Attacks on Watermarks ......................................................... 25
2.4.7 Image Watermarks Domains .................................................. 26
Contents
2.4.8 Watermark Properties ............................................................ 27
2.4.9 Watermarks for Document Images .......................................... 27
2.4.10 Comparison between Binary Image Watermarking Techniques ... 29
2.5 Methods in Literature .................................................................... 31
2.5.1 Colour and Greyscale Image Data Hiding Methods ................... 31
2.5.1.1 Spatial Domain Data Hiding Techniques ........................... 31
2.5.1.2 Frequency Domain Data Hiding Techniques ...................... 36
2.5.1.3 Hybrid Data Hiding Techniques ....................................... 43
2.5.1.4 Other Data Hiding Techniques ........................................ 44
2.5.2 Binary and Half-tone Image Data Hiding Methods ................... 46
2.5.3 Other Techniques.................................................................. 57
2.6 Conclusion ................................................................................... 60
Chapter Three: Fundamental Methodologies and Principles ....................... 61
3.1 Introduction ................................................................................. 61
3.2 Print and Scan Model ..................................................................... 61
3.2.1 The Print Process .................................................................. 62
3.2.2 The Scan Process .................................................................. 64
3.3 Centre of Gravity .......................................................................... 65
3.4 The Method of Moments ................................................................. 67
3.5 Quad Tree ................................................................................... 68
3.6 The 2D Data Matrix Barcode ........................................................... 70
3.7 Type I and Type II Statistical Errors ................................................ 72
Chapter Four: Methodology ................................................................. 74
4.1 Introduction ................................................................................. 74
4.2 Method 1 ..................................................................................... 74
4.2.1 The Creation of Self-validating Documents ............................. 74
4.2.2 The Verification of Self-validating Documents ......................... 77
4.3 Method 2 ..................................................................................... 82
4.3.1 The Creation of Self-Validating Documents ............................. 82
4.3.2 The Verification of Self-validating Documents ......................... 88
4.4 The Extended Version of Method 2 .................................................. 94
4.5 A comparison between Method 1 and Method 2 ................................. 97
4.5.1 Similarities of Method 1 and Method 2 ..................................... 97
Contents
4.5.2 Differences of Method 1 and Method 2 ..................................... 98
4.6 Limitations of using Method 1 ......................................................... 98
4.7 Modifications to Method 1 ............................................................ 102
Chapter Five: Experimental Results and Analysis .................................. 103
5.1 Overview ................................................................................... 103
5.2 The Experimental Results of Method 1 ........................................... 103
5.2.1 Character Image Analysis .................................................... 104
5.2.2 Results of the Verification System of Method 1 ........................ 115
5.3 The Experimental Results of Method 2 ........................................... 118
5.4 The Experimental Results of the Extended Version of Method 2 ......... 121
5.5 A Comparison between the Results of the Proposed Methods ............. 123
5.6 The Influence of Rotation on the Verification Process ....................... 124
5.7 Invisible Data Hiding ................................................................... 125
Chapter Six: Conclusions and Future Research ..................................... 128
6.1 Recommendations to Improve the System ...................................... 128
6.1.1 Noise Removal ................................................................. 128
6.1.2 De-skewing ..................................................................... 129
6.1.3 Optical Character Recognition (OCR) ................................... 130
6.1.4 Data Compression ............................................................ 131
6.1.5 Data Encryption ............................................................... 131
6.1.6 Auto Selection of Data Embedding ...................................... 132
6.2 Conclusions ................................................................................ 133
Appendices ...................................................................................... 135
Appendix A: Two Sample Document Images Divided by Method 1 ...... 136
Appendix B: Two Sample Document Images Divided by Method 2 ...... 137
Appendix C: Two Sample Document Images Divided by The Extended Version of Method 2 .......................................................... 138
Appendix D: A Sample RYANAIR Flight Boarding Pass....................... 139
List of References ............................................................................. 140
1
List of Tables Table 2.1: Comparison between Steganography and Digital Watermarking ............ 18
Table 2.2 : Applications of Digital Watermarking ................................................ 21
Table 2.3 : A Comparison between Several Watermarking Techniques for Document
Images ........................................................................................ 30
Table 3.1: Hypothesis Test in Statistics ............................................................ 73
Table 4.1 : Possibilities and Results of the 2-bit Representation Comparison .......... 92
Table 5.1 : The Influence of Print-Scan Operation on the Centroids and the Rates of
Black Pixels of 71 Characters Written in 7 Different Font Styles .......... 107
Table 5.2 : The Average Influence of Print-Scan Operation on Characters Written in 7
Different Font Styles Sorted from the Least to The Most Affected Fonts 108
Table 5.3 : The Maximum and Minimum Change in the X, Y, and R Values and their
Most and Least Influenced Characters ............................................. 109
Table 5.4 : The 10 Most Affected Characters by Print-Scan Operation and the
Average Difference in Their Centroid Positions and the Rate of Black Pixels
................................................................................................ 109
Table 5.5 : The 10 Least Affected Characters by Print-Scan Operation and the
Average Difference in their Centroid Positions and the Rate of Black Pixels
................................................................................................ 110
Table 5.6 : The Frequency Percentage of the 10 Most Affected Characters by Print-
Scan Operation ........................................................................... 111
Table 5.7 : The Influence of Print-Scan Operation on the Centroids and the Rates of
Black Pixels of 71 Characters Written in 9 Different Font Sizes ........... 113
Table 5.8 : The Average Influence of Print-Scan Operation on the Centroids and
Rates of Black Pixels of Characters Written in 9 Different Font Styles .. 114
Table 5.9: The Experimental Results of Method 1 Verificatoin System on 55 Unaltered
and 100 Forged Scanned Documents .............................................. 116
2
Table 5.10: The Statistical Error of the Verification System of Method 1 .............. 118
Table 5.11 : The Experimental Results of the Verification System of Method 2 on
Scanned and Forged Images ......................................................... 119
Table 5.12: The Statistical Error of the Test Results of Methd 2 .......................... 120
Table 5.13: The experimental Results of the Verification System of the Extended
Version of Method 2 on Scanned and Forged Images ........................ 122
Table 5.14: The Statistical Error in the Verification System of the Extended Version of
Method 2 .................................................................................... 123
Table 5.15: The Average Rates of Successfully Verified and Rejected Documents by
the Three Proposed Methods After Print-Scan Operation .................... 124
Table 5.16: The Rotation Influence on Document Verification ............................ 125
Table 5.17 : The Percentages of Frequencies of (i & j) Letters per 1000 Words and
Per Page in a Sample (2,188,153 words) Text Book .......................... 127
3
List of Figures
Figure 2.1 : The Block Diagram of Watermark Embedding Scheme ....................... 44
Figure 2.2 : The Block Diagram of Watermark Detection Scheme ......................... 44
Figure 3.1 : The Print-Scan Model .................................................................... 62
Figure 3.2 : The Structure of Laser Printer ........................................................ 63
Figure 3.3: The Structure of Scanner ............................................................... 65
Figure 3.4: The Centre of Mass of (a) Symmetric Shapes and (b) Asymmetric Shapes
.................................................................................................... 65
Figure 3.5: Closed Polygon with 6 Vertices ........................................................ 66
Figure 3.6: Self Intersecting Polygons .............................................................. 67
Figure 3.7: An Example of Quad Tree of an Image (a) before and (b) after
partitioning ..................................................................................... 69
Figure 3.8 : The Depth Levels of Quad Tree Partitioning of an Image .................... 69
Figure 3.9: The 4 Components of the 2D Data Matrix Barcode ............................. 71
Figure 4.1 : An Example of Computed Centroid Point (Xc, Yc) & the Rate of Black
Pixels of a Binary Image Contains the Letter “O”. ................................. 76
Figure 4.2 : A Block Diagram of Creating a Self-Validating Document ................... 77
Figure 4.3: A Sample Scanned Image (a) before & (b) after Noise Removal Process 78
Figure 4.4 : An Example of a Character (a) before and (b) after Forgery, its Centroid,
and Black Rate ................................................................................ 80
Figure 4.5 : The Verification Process of a Scanned Document .............................. 81
Figure 4.6 : An Example of a (10×10) Image where Tmargin = 10 in (a) & Tmargin
= 20 in (b) ..................................................................................... 83
Figure 4.7 : An Example of the Data Compression Technique Used to Convert the 8-
bits/value into 2-bits/value data stream .............................................. 86
Figure 4.8 : A Flowchart of the Creation Process of Self-validating Documents ....... 87
Figure 4.9: An Example of Selected Area of Interest .......................................... 88
4
Figure 4.10: An Example of Cropping an Image With No Registration Marks Using
Method 2 (A) Before Cropping (B) After Cropping ................................. 89
Figure 4.11: A flowchart of the verification process of self-validating documents .... 93
Figure 4.12 : An Example of Two Different Sub-images and Their Vertical &
Horizontal Distributions .................................................................... 96
Figure 4.13: A Sample of a Strikethrough Text .................................................. 99
Figure 4.14 : A Sample of a Strikethrough Text ................................................. 99
Figure 4.15: Texts Written in (a) Arabic, (b) Urdu, and (c) Farsi languages ......... 100
Figure 4.16: Samples of Texts Written in English Script Fonts............................ 100
Figure 4.17: A Sample of Text Image Written in Italic Style and How it is Divided into
Sub-images Using Method 1 ............................................................ 101
Figure 4.18: A Sample of Text (a) Neighbouring a Vertical Line (b) in a Table With
Solid Lines .................................................................................... 101
Figure 5.1: Samples of Connected Characters Due to the Additional Noise of Print-
Scan Operation ............................................................................. 104
Figure 5.2: Samples of Calibri, Arial, Latha, Verdana, Times New Roman, Implact and
Georgia fonts ................................................................................ 105
Figure 5.3: The Average Influence of Print-Scan Operation on the Centroids and
Rates of Black Pixels of Characters Written in 7 Different Font Styles .... 108
Figure 5.4: The Average Influence of Print-Scan Operation on the Centroids and
Rates of Black Pixels of Characters Written in 9 Different Font Styles .... 114
Figure 5.5: The Rotation Influence on Document Verification ............................. 125
Figure 5.6 : An Example of Dot Shifting Technique to Embed 3 Bits of Data by
Shifting the Dot in the Letter (i) ....................................................... 126
Figure 6.1: A Flowchart for the Auto-Selection of Data Embedding Technique ...... 133
5
List of Abbreviations
2D ................. Two Dimensional
AFID ................. Anti-Forgery Identification Document
ASCII ................. American Standard Code for Information Interchange
AWST ................. Authentication Watermarking by Self Toggling
BCH ................. Bose-Chaudhuri-Hocquenghem
BMP ................. BitMaP
CAD ................. Computer Aided Design
CCD ................. Charge Coupled Device
CCW ................. Counter ClockWise
CW ................. ClockWise
CWDD ................. Curvature-Weighted Distance Difference
D⁄A ................. Digital-to-Analogue
DCT ................. Discrete Cosine Transform
DES ................. Data Encryption Standard
DHC ................. Data Hiding Capacity
DHST ................. Data Hiding by Self Toggling
DPI ................. Dots per Inch
DRDM ................. Distance Reciprocal Distortion Measure
DRM ................. Digital Rights Management
DS ................. Digital Signature
DWT ................. Discrete Wavelet Transform
ECC ................. Error Correction Code
FFT ................. Fast Fourier Transform
GIF ................. Graphics Interchange Format
HMAC ................. Hashed Message Authentication Code
IDCT ................. Inverse Discrete Cosine Transform
IDWT ................. Inverse Discrete Wavelet Transform
LSB ................. Least Significant Bits
MAC ................. Message Authentication Code
6
MD5 ................. Message-Digest algorithm-5
MSB ................. Most Significant Bits
MTF ................. Modulation Transfer Function
OCR ................. Optical Character Recognition
PC ................. Personal Computer
PDF ................. Portable Document Format
PSNR ................. Peak Signal-to-Noise Ratio
Q-tree ................. Quad Tree
RGB ................. Red-Green-Blue
RLE ................. Run Length Encoding
RSA ................. Rivest, Shamir, & Adleman
SARI ................. Self-Authentication-and–Recovery Image Watermarking System
SPIHT ................. Set Partitioning in Hierarchical Trees
SysCoP ................. System for Copyright Protection
TIFF ................. Tagged Image File Format
TNR ................. Times New Roman
UDWT ................. Undecimated Discrete Wavelet Transform
WPC ................. Wet Paper Coding
ZMM ................. Zernike Moments Magnitude
7
Acknowledgements
It is a pleasure to thank those who made this thesis possible. First of all, I would
like to thank my father Tahir Albakaa and my mother Koulod Albadri for their unlimited
support. They were always beside me when I needed financial help and incorporeal
support. Dr. Dave Elliman was my supervisor when I started my PhD who helped me
to choose the research topic. Dr. Peter Blenchfield became my supervisor after Dave’s
retirement. I would like to thank both Dave and Peter for sharing thoughts with me,
for their time, encouragement, guidance, advice, and patience. This thesis would not
have been possible without them. Also, I would like to show my gratitude to my friends
Nawfal A. Mehdy, Belal Al-khateeb, Musamer Shamil, and Sanar Shamil for their help
and support. At last but not least, I would like to acknowledge the staff members of
the School of Computer Science and The International Office at the University of
Nottingham.
Chapter 1: Introduction
8
Chapter One: Introduction
1.1 Overview
Many documents are created or converted and then stored in digital format
such as Portable Document Format (PDF) or Microsoft Word Document (DOC) files or
any other digital format. Digital documents are often converted to hardcopy
documents when required. Printed documents with significant information such as
birth certificates, recommendation letters, prescriptions, contracts, and sale receipts
are subject to forgery.
However, it is not likely to validate hardcopy versions of documents unless they
are already printed as machine readable documents. The technology to create a
machine readable document is expensive and used only to protect documents of high
importance or financial value such as bank notes and passports. The price of using
high cost technology to protect ordinary documents such as a letter of
recommendation would not be recovered (Sun et al, 2001).
The simplest way to check the integrity of a printed document is to compare it
with the original copy. However, the requirement for an original copy of each
document during the verification process is unreasonable. Therefore, it is better to
embed an information digest about the whole document, or parts of the document
which may be the subject of counterfeiting, into the document itself. A data hiding
method would be useful in making the document carry this information digest.
Most documents are printed in black and white rather than colour due to the
wide use and availability of black and white printers in offices and the work place
(Zhu, 2003). Hiding information in binary document images is a very challenging
operation and most of the data hiding techniques available are designed for
Chapter 1: Introduction
9
greyscale and colour images because of the fact that any alternation in the binary
image grid can easily cause perceptual artefacts in the modified image. In addition,
the data hiding capacity in binary images is low compared to that in colour or
greyscale images. Furthermore, data hiding methods are very sensitive to D⁄A
conversion which can easily remove or corrupt the embedded data. (Furht et al,
2005; Yang and Kot, 2004; Lu et al, 2003; Mei et al, 2001; Low et al, 1998; Arnold,
2003; Kim and Oh, 2004).
1.2 The Project Objective
The aim of this work is to provide a mechanism to determine the validity of black
and white printed documents. This mechanism must be cost effective compared to
the value of the documents. The degree to which the validity can be verified will
depend on an intrinsic/extrinsic value of the binary documents. This validation will be
of documents which cannot already be guaranteed by security methods.
To create a verification system which is able to check the integrity of document
images after the print-scan operation, it is essential to take the following points into
consideration in order to detect tampering in document images.
1- The significance of alternations in the document to be considered as a
forgery.
2- The locations of the altered parts.
3- The preservative data extracted from the original document that represent
the protected parts of the document. The preservative data are selected
information in the document which counterfieters are interested to alter for
malicious purposes. Examples of the preservative data are someone’s name
or singnature.
Chapter 1: Introduction
10
4- The information hiding or insertion method to embed the preservative data
into the document.
5- The sensitivity measure of the method against minor changes which may
occur in the document before and during the print-scan operation.
The above-mentioned points will be answered and discussed in detail in this
chapter.
1.2.1 The Significance of Alternations
If a hardcopy document has been altered, it is important to measure how
significant the change is to determine whether this document has been forged or not.
The consequences of the failure of detecting significant modifications could be severe
if the forged document were a financial paper such as a contract or were an official
document such as a university certificate or a recommendation letter. This wrong
approval can cost a substantial loss of money in banks and official organizations. In
statistics, this failure is called a Type II error, and is also known as a false-negative
error (Onwuegbuzie and Daniel, 2003). In the case of document authentication, it
means the endorsement of a fake document as a genuine one. The probability of
failure has to be very low in order to achieve a satisfactory rate of reliability.
Otherwise, the authentication system will be defective.
1.2.2 The Location of Changes
If a scanned document has been modified, it is important to locate the position
of parts that have been significantly altered to distinguish between forged and
genuine areas in the document. It makes it easier for the person who verifies the
scanned document to locate where exactly the changes are.
Chapter 1: Introduction
11
1.2.3 The Representation of Preservative Data
Most documents which need to be protected contain vital information which
counterfeiters have an interest in changing. Examples of text that can be a forger’s
target might be an amount of money, a name of person, or a signature in a contract
or any legal document. It is unlikely that every single word in the document is in the
area of the counterfeiters’ interest. Only the important information needs to be
encoded and verified and failure to verify these data leads to considering the
document as a forgery.
This significant part of the document must be represented in different ways
depending on many elements such as the data type, data size, the data host, and
the level of robustness required for the verification system. Examples of this data
representation can be a character’s ASCII code, a hash function, or any other
computed values representing details about the data such as the centre of gravity or
the black/white colour ratio of objects or characters in the scanned document. To
create a self-validating document, this preservative data must be inserted in the
document itself.
1.2.4 The Insertion of Preservative Data
The preservative data needs to be stored in the same document where this data
was extracted from, in order to produce a self-validating document. The inserted
data must be extractable after a print/scan operation is applied on the host. Various
methods can be used to embed this data into the document. Data hiding techniques
can be used if the capacity of the host image is sufficient to hold the preservative
data. Visible barcodes attached to the document have also been used in some
Chapter 1: Introduction
12
applications to carry some information about the host. In general, the more
preservative information is stored in the document the more reliable the verification
system will be.
1.2.5 The Sensitivity of the Verifier
When a document is printed and scanned, various levels of noise appear on the
document after the operation. The additive noise can be caused during the print or
scan process even if the glass of the flatbed scanner was completely clean. Also,
printed details in hardcopy documents may fade if kept and exposed to light for a
long time. The mistreatment and negligence by the holders of paper documents will
lead to paper documents sometimes being bent or torn.
Despite the fact that those documents are still genuine, the verification system
may consider them as forged copies because of the unintentional added noise. The
system should be able to distinguish between malicious and innocent changes.
Otherwise, authentic papers may be rejected whilst they are supposed to be
accepted and this kind of false verification is called a Type I Error (false-positive).
Therefore, it is necessary to compromise on the rates of noise to come to a decision
which document is sufficiently corrupted to be considered as a forgery and which one
does not to reach an acceptable level of accuracy in verification.
1.3 Overview of Chapters
This thesis contains six chapters. An introduction about the research area, the
problem of validating binary document images after the print-scan operation, and the
main objective of this research were presented earlier in this chapter. Chapter Two
sheds light on work related to image integrity checks as well as on information hiding
techniques for digital images proposed in the literature. Chapter Three defines some
Chapter 1: Introduction
13
general terminologies and procedures used in the proposed validation system in this
research. The methods to validate black and white hardcopy documents, their
limitations, and modifications are proposed in Chapter Four. The experimental results
of the method proposed in Chapter Four are presented and discussed in Chapter
Five. Furthermore, practical measurements of the print-scan operation impact on
characters in the English language are also discussed in Chapter Five. Chapter Six
concludes the proposed methods and their results and provides recommendations
and suggestions to modify the proposed document validation methods for future
research.
Chapter 2: Literature Review
14
Chapter Two: Literature Review
2.1 Introduction
Document authentication has been an interesting topic for researchers because
of the extensive use of documents in business and society in general. Important
documents, for instance currency notes and passports, can be protected by either
physical means such as added metal security threads or by chemical means such as
special inks, subtle colours, and holograms using expensive high resolution printers.
However, it is impracticable to apply those techniques on some other types of
documents such as certificates, deeds, and contracts because of the high cost of the
materials used in these techniques (Sun et al, 2001).
The widespread availability of low-priced scanners, printers, photocopiers, and
other office equipment facilitates creation and distribution of illicit documents.
Furthermore, paper documents can be converted to digital form easily by using any
Optical Character Recognition (OCR) software without the need for retyping texts
(Arnold, 2003). In addition, data forgery has been made effortless by using one of
the widely-available image editors such as Photoshop. Images can be easily faked by
cropping some regions and substituting them for other parts of the image. This
facilitates undesirable results such as changing names in passports, university
certificates, provincial health cards, or driving licences (Fridrich and Goljan, 1999).
Consequently, the need has increased enormously for protecting important digital
document images used in government institutions and organizations. Many
techniques such as fingerprinting, and encryption have been used to protect sensitive
digital deeds (Alattar et al, 2005). Encryption can prevent illegal access to digital
data. However, the probable collusion of some authorized users and supposed-to-be
Chapter 2: Literature Review
15
trusted employees who work in these organizations creates a serious issue as
decrypted data might be copied and distributed illegally as long as these employees
have full access to original documents and they know the decryption key (Furht et al,
2005).
This type of collusion can be made more difficult by one of the aforementioned
methods. Fingerprinting is one proposed solution for finding the source of a leak that
distributes sensitive documents and it should be applied alongside the above
mentioned method to create an efficient security system for digital papers. The main
objective of fingerprinting is to identify any employee who circulates a secret
document outside the workplace. This identification is done by inserting information
into the document itself such as an identification number referring to a particular
employee who was given a document (Alattar et al, 2005).
Moreover, digital multimedia such as images, text, audio, and video files have
been widely used and disseminated via the Internet. Therefore, people around the
world have been increasingly relying on using digital data in communication instead
of traditional paper. However, the extensive use of the Internet has facilitated the
distribution of digital data and, as a result, digital data forgery has become a
lucrative industry (Lan and Tawfik, 1999; Yang et al, 2004).
Indeed, intellectual property protection and authentication of digital data are
considered as essential. Digital watermarking techniques have been used for the
purpose of copyright protection. More recently, counterfeiting detection has become
one of watermarking’s important applications (Lan and Tawfik, 1999). Nevertheless,
more research is needed to detect forgery in document images.
Chapter 2: Literature Review
16
2.2 Methods to Protect Documents
According to Zhu et al (2003), there are four different main methods to generate
authentic documents:
1- Using either special physical material such as high-resolution printers or
chemical material such as special inks and papers. This is used for special
applications with high protection requirements for instance currency notes
and bank cheques.
2- Fingerprinting: this is used to trace illegal copies and identify the source of
collusion by inserting information about the legitimate user in each copy given
to this client. However, this method requires special high-priced detectors to
extract the hidden labels. Consequently, it can only be used in a limited range
of application for example tickets and cheques.
3- Digital methods: the aim of these techniques is to embed digital signatures,
generated by authorized users using a private key, into paper documents.
Examples include the use of barcodes and watermarks in bills and ID cards.
4- Visual Cryptography and Optical watermarks: Visual cryptography is based on
dividing an image outline into several parts such that it can be constructed
again by merging all the parts. Optical watermarks offer higher visual quality
and embed several layers of data. Both techniques are used for manual
authentication such as in brand protection applications. However, they are not
able to verify printed and scanned documents.
2.3 Data Hiding Techniques
There has been a substantial increase in production and development of data
hiding techniques to protect the copyright and intellectual property of digital
multimedia such as images, music, video, and texts. The set of these technologies is
Chapter 2: Literature Review
17
called Digital Rights Management (DRM). In multimedia DRM, the encryption and
watermarking techniques can be used for data authentication, secret
communications, and copy control (Furht et al, 2005).
Both Digital Watermarking and Steganography are data hiding techniques used to
embed digital information into a cover object. However, their characteristics and
application requirements are different.
2.3.1 Steganography
It is a Greek word which means “Secret Writing” (Brown, 1993). The main
objective of steganography is to embed a secret message into the cover object
without causing any visual artefacts to appear in the cover data. Therefore, the
secret message is the most important element in this technique while the cover is a
tool used to hide and protect this hidden message. The security level of the
embedding and extractions as well as the imperceptibility levels must be high in
steganography. However, the level of robustness does not necessarily have to be
high.
2.3.2 Digital Watermarking
By contrast, Digital Watermarking techniques embed digital data in a cover in
order to protect the host data. Unlike steganography, the imperceptibility of the
inserted watermark is not a priority in some watermarking applications. Therefore,
the watermark can be visible or invisible. Watermarks are usually supposed to be
robust and not easy to eliminate (Katzenbeisser, 2000).
Table (2.1) shows the similarities and differences between Steganography and
digital watermarking in terms of visibility, robustness, security of keys, applications,
and the importance of data.
Chapter 2: Literature Review
18
Steganography Watermarking
Visibility Strictly Invisible Visible or Invisible
Robustness Fragile Robust or Fragile
Security of keys High level Low/high level
Applications Point to point secret
communications
Copyright protection, Copy control,
Authentication and tracking.
Importance of Data The hidden data is vital The inserted data is used to protect the
host data
Table 2.1: Comparison between Steganography and Digital Watermarking (Katzenbeisser, 2000)
2.3.3 Data Hiding Applications
Most newspapers, books, pictures, music and videos are being converted to
digital form because digital data are easy to publish, faster to transmit, cheaper to
produce, and smaller in size than analogue data or hardcopies. Therefore, digital
data have been increasingly used and published on the Internet. However, anyone
who has access to the Internet can download and may misuse the published digital
materials. Accordingly, it has been essential to find ways to protect digital
multimedia and printed materials, trace illegal copies, and discover the source of
leaks. Data hiding techniques aim to embed a piece of data imperceptibly into
document images to solve many issues. Chen et al (2004) have indicated five
different applications of data hiding techniques:
1- Copyright Protection: can be obtained by inserting a robust watermark,
such as owner ID, signature, or logo, into a digital document and this
Chapter 2: Literature Review
19
embedded mark should not be eliminated by any malicious attacks applied on
the document unless the protected media becomes corrupted and worthless.
2- Fingerprinting: aims to embed imperceptible and robust distinctive
signatures (e.g. the recipient’s name) into each copy of the document to allow
digital data owners or distributors to discover any collusion that may happen
after distribution.
3- Copy prevention and copy control: Copiers such as photocopy machines
and CD writers can be designed to prevent illegal copies or limit the number of
copies if they are designed to detect the presence of a watermark already
embedded in digital or paper documents and, then, either to forbid or allow
copying depending on the system requirement.
4- Authentication: is the detection of malicious manipulations that may possibly
be applied on the document (e.g. changing the holder’s name on a passport or
a number in a cheque). Authentication can be achieved by embedding
information about the original document into the document itself, without
affecting the quality of the host. In the verification process, the document can
be considered as genuine if and only if the extracted watermark and the
document information are alike. The research described in this thesis is aimed
at the authentication of digital binary document images.
5- Metadata binding: used to embed metadata information into an image for
purposes such as product advertisement in digital cameras.
Digital images can be verified by cryptographic authentication methods using
hash functions to determine whether they are authentic or not. However, the
protection of image integrity cannot be guaranteed following image decryption. Also,
hash functions are incapable of locating or quantizing changes that might have been
applied to these digital images. Therefore, all digital data such as document images,
Chapter 2: Literature Review
20
with malicious or desirable changes, are not verified during the authentication
process in spite of the magnitude of the visual impact on the altered data (Fridrich,
2000). One of the suggested solutions, for recognising the regions that have been
tampered with maliciously, is the use of digital watermarking techniques (Fridrich,
1999a). Consequently, watermarking is used to protect data copyrights, to trace
illegitimate copies, and to find out sources of leaked information.
2.4 Digital Watermarks
Digital watermarking can be defined as an information hiding technique that
inserts signatures into digital content for the purpose of data ownership protection
and authentication. The inserted watermark must be imperceptible, secure, and
robust (Furht et al, 2005). In document communication, digital watermarking can be
used to assist a document recipient to recognize document owners and also to verify
the genuineness of documents (Furht et al, 2005).
2.4.1 Watermarking Applications
Digital watermarks can be embedded into multimedia files for different purposes
and applications. Furht et al (2005) classify the different applications of digital
watermarking as shown in Table (2.2).
Chapter 2: Literature Review
21
Application Class Watermark purpose Application Scenarios
Protection of intellectual property rights
Conveys information about content ownership and intellectual property rights
• Copyright protection • Copy protection • Fingerprinting
Content verification
Ensure that the original multimedia content has not been altered and/or helps to determine the type and location of alternation
• Authentication • Integrity checking
Side-channel Information Represent side-channel used to carry additional information
• Broadcast monitoring • System Enhancement • Covert communications
Table 2.2 : Applications of Digital Watermarking (Furht et al, 2005)
2.4.2 Watermark Visibility
There are two different types of watermarks in terms of perceptibility. Visible
watermarks are usually presented as a corporate logo or visual text and this kind of
mark is used to protect the copyrights of digital media. Invisible watermarks cannot
be seen with the naked eye and thus extraction or detection methods are required to
take out or to check the existence of such watermarks (Kim & Afif, 2003; Kim & Afif,
2004).
The main aim of using visible watermarking is to warn users of digital media
against copying or misusing the copyrighted materials. These kinds of watermarks
are easy to implement and also easy to remove (Sun et al, 2001).
2.4.3. Watermarking System
Lan and Tawfik (1999) classify watermarking systems into three categories:
1- Private watermarking system: there are two types of this system and both
types require the original image; Type I is used to extract the embedded
watermark while Type II is to check the existence of the watermark.
Chapter 2: Literature Review
22
2- Semi-Private watermarking system: detects the presence of the
watermark, as does Type II; however, it does not use the original image.
3- Public watermarking system: extracts the embedded watermark without
using the original image or the key. It is also called Blind Detection System.
2.4.4. Watermarks Robustness
Watermarks can be classified as robust or fragile according to their resistance
to attacks and each class is used for different purposes. Copyright and ownership
protection applications require using a robust watermark while fragile watermarks
are used for authentication systems because they are very sensitive and can be
removed with ease; consequently, when watermarked media are attacked,
watermark detectors may fail to detect the embedded signature and, then, fail to
validate the tampered files (Kim and Queiroz, 2004b).
Robust watermarks are used for copyright protection and this type of watermark
must not be removable by ordinary image processing tools such as image resizing
and compression. Fragile watermarks are sensitive to any manipulation applied on
the watermarked data and this kind of watermark is used to determine whether the
watermarked data have been tampered with or not (Kim & Afif, 2003; Kim & Afif,
2004).
Kim and Afif (2003) believe that only cryptography-based watermarking
techniques are reliable for authentication purposes while other watermarking
techniques cannot be robust.
The main difference between copyright protection watermarks and authentication
watermarks is that copyright protection watermarks are designed to be robust and
must survive not only all unintentional alternations but also any intentional attacks
that may be applied on the cover media. However, authentication watermarks should
Chapter 2: Literature Review
23
be more sensitive in order to detect any alternations applied on the host media (Li et
al, 2003).
In Fridrich (1999a), authentication watermarking techniques are classified into
three main categories according to their robustness: fragile watermarks, semi-
fragile watermarks, and robust watermarks where each class has been designed
for particular purposes. Also, it is possible to combine two or more types of
watermark together to generate a hybrid watermark provided that the second
chosen watermark to be inserted (usually fragile) does not exterminate the
previously embedded watermark.
2.4.4.1. Fragile Watermarks
They provide a high level of vulnerability and can be easily corrupted by any
possible modification that might be applied on a watermarked image such as
changing a pixel value. One of the most well-known primitive fragile watermarking
techniques is based on setting the Least Significant Bits (LSB) of the image pixels
(Fridrich, 1999a).
2.4.4.2. Semi-Fragile Watermarks
The second category of watermark is semi-fragile. It is less sensitive than the
first type and can be used to gauge the corruption in the attacked images. It is
necessary for it to be robust against ordinary image processing operations such as
JPEG compression and brightness adjustment; simultaneously, it must be vulnerable
to malicious perceptible attacks such as cropping or adding additional features to the
watermarked images. Semi-fragile watermarking techniques use either the frequency
or spatial domain depending on the requirements of the system to be designed. A
threshold can be set to adjust the fragility of the designed technique (Fridrich,
1999a).
Chapter 2: Literature Review
24
2.4.4.3. Robust Watermarks
They must survive regardless of applied attacks, whether they be malicious or
conventional, unless the watermarked image is visibly damaged to the extent it has
been made worthless. Therefore, this kind of watermark is mainly proposed for
ownership and copyright protection. The majority of robust watermarking techniques
are based on frequency transforms such as Discrete Cosine and Wavelet transforms
(Fridrich, 1999a).
2.4.5 Imperceptibility vs Robustness
Kwok et al. (2000) point out that all visible watermarking techniques proposed in
the literature are easy to remove or corrupt while all invisible watermarks are robust.
They also believe that all text watermarks have to be invisible and fragile, audio
watermarks must be inaudible and robust, video watermarks must be invisible and
robust, and image watermarks have to be either visible-fragile or invisible-robust.
However, these assumptions and categorizations cannot be strictly true because
different watermarking systems have different requirements. Embedding a
watermark in an image’s LSBs is a simple example of an invisible-fragile image
watermark; therefore, it is not necessary for it to be visible-fragile or invisible-
robust.
Nowadays, a contemporary category of image authentication techniques is being
investigated called self-embedding watermarking which aims to identify and
restore corrupted parts in attacked images as it stores the image content into the
image itself. This type requires high data embedding capacity which makes it difficult
to conserve the quality of the watermarked image. Furthermore, it seems to be
unlikely to provide a robust self-embedding scheme capable of surviving all available
image processing applications such as lossy compression (Fridrich, 1999a).
Chapter 2: Literature Review
25
In content-based data hiding techniques, a message authentication code
(MAC) or a digital signature (DS) must be generated from the original image and,
then, embedded into the same image to produce the watermarked image. Inserting
the MAC/DS to the original image causes some changes to the host image; therefore
a subsequent derived MAC/DS from the watermarked image will not be identical to
the inserted MAC/DS. As a result, the verification process will not be successful (Kim
& Afif, 2003; Kim & Afif, 2004).
For that reason, least significant bits (LSB), in colour and greyscale images,
must be cleared before generating the original MAC/DS and, then, used to hold the
code or the signature. However, this cannot be applied on binary images because
each pixel is represented by one bit and any modification to the binary grid will cause
salt-and-pepper noise and the watermark will be invalid (Kim & Afif, 2003; Kim &
Afif, 2004).
2.4.6 Attacks on Watermarks
In authentication and forgery exploration, there are three main causes of
extraction failure during the detection process. Firstly, it may be caused by
intentional alteration to the meaning of a sentence or value of a number and, in this
case, the detector must discover this kind of forgery and inform the recipient that
the document has been counterfeited. The second possible attack could occur by
image processing operations such as rotation, skewing, and compression, which do
not change the content of documents, and detectors need to ignore this sort of
regular change. Also, detectors may fail to validate document images if they are
basically not watermarked and belong to another owner.
Therefore, it is preferable to embed yet another watermark which is robust to
malicious attacks for the purpose of identifying legitimate owners. A digital
Chapter 2: Literature Review
26
watermarking technique that inserts two watermarks is called two-level
watermarking. Watermarks must not be integrated into file headers because the
conversion of a file format, (for example from .BMP to .GIF), and D/A conversion
such as a print/scan attack eradicate watermarks embedded in headers (Furht et al,
2005). Imperceptible data embedded in digital images for the purpose of detecting
potential changes are called authentication watermarks (Kim & Afif, 2003; Kim & Afif,
2004).
2.4.7 Image Watermarks Domains
Watermarks can be embedded in two different domains into digital images
(Hsieh et al, 2004):
1- Spatial domain: a simple approach where the watermark stream is directly
embedded into the host by modifying the pixel values of the image. The most
popular approach is altering LSBs in colour and greyscale images or flipping
pixels in binary images. Watermarks embedded in the spatial domain are
fragile and it is easy to eliminate inserted data by applying one of the image
processing operations such as JPEG compression or smoothing.
2- Frequency domain: Watermarks are embedded in the most significant
components of the host image after applying one of the frequency-domain
transforms such as Discrete Cosine Transform (DCT) and Fast Fourier
Transform (FFT) on the image. Hsieh, M. & Tseng, D. (2004) believe that
frequency domain watermarking techniques are not only more robust to
common attacks, especially image compression, than spatial domain
techniques but also provide larger data hiding capacity.
Chapter 2: Literature Review
27
2.4.8 Watermark Properties
A reliable image watermarking technique must have the following features
(Hsieh et al, 2004):
1. Imperceptibility: The inserted watermark should be invisible to the naked
eye and the watermarked image must appear to be an identical copy of the
original.
2. Robustness: Watermarks must survive image processing operations and
resist other deliberate attacks that could be applied to the watermark image
in order to eliminate the inserted watermarks.
3. Unambiguousness: Embedded marks must contain sufficient information
about the lawful owner of the protected image to prove the ownership.
4. Security: A strong cryptography method can be used to achieve a high level
of security to prevent illegal users from accessing hidden information. Only
authorized private/public key holders are permitted to decrypt watermarks.
2.4.9 Watermarks for Document Images
In Kim and Oh (2004), watermarks can be inserted into document images at
three levels. The first one is the character-level which adjusts characters, words or
lines by shifting them upward, downward, left or right. The second is Feature-level
which alters stroke width or serif shape of letters or modifies the font formatting of
the text such as colours, sizes, or form. Pixel-level is the third which is not robust
even against simple attacks but is suitable for embedding large amounts of data.
In addition, replacing some words with synonyms, provided that the meaning is
equivalent to the original, is another way to watermark plain text but this method is
not applicable on some texts, for instance, poems (Arnold, 2003).
Chapter 2: Literature Review
28
According to Kim and Queiroz (2004b), there are 3 approaches to embedding data
in digital binary or halftone images:
1- Pixel-wise: This directly alters the pixels values. Consequently, it may cause
salt-and-pepper noise if applied on binary document images but is more
applicable for halftone images.
2- Component-wise: This modifies certain features of the image content and
cannot be applied to all images as its reliability depends on the details of the
image to be watermarked.
3- Block-wise: This splits the images into small blocks (e.g. 8x8) and then embeds
watermark bits into each block depending on its features.
Other data hiding techniques for document images have been proposed. Chen et
al (2005) have classified these techniques with regard to the embedding process as
follows:
1- Text line, word and character shifting: can be applied on formatted text
documents by shifting lines, groups of words, or groups of characters slightly
to the left/right or up/down to encode a 0 or 1. Line shifting techniques
conspicuously show better robustness against D/A attacks such as print/scan
than other methods while character shifting techniques have a higher data
hiding capacity. However, different languages and fonts play a significant role
in robustness and capacity of word and character shifting methods.
2- Fixed partitioning into blocks: divides an image into small blocks with
fixed size and flips one pixel or more within selected blocks and without
creating artefacts depending on the characteristics of those chosen blocks.
The capacity of data hiding using this technique depends on the contents of
the document and it can be increased at the expense of the possible
perceptibility of the added data.
Chapter 2: Literature Review
29
3- Boundary modifications: these can embed 5.69 bits of data in the 8-
connected boundary of each character or connected components as shown by
Mei et al (2001). A blind decoding method without using the original
document can be used to extract the inserted data.
4- Modifications of character features: this approach embeds data by
changing pre-extracted character features of the document. It offers
considerable resistance to D/A conversion but still needs more enhancements
in order to be robust against photocopying. Its data embedding capacity relies
on the number of characters in the document.
5- Modifications of run-length patterns: data can be embedded using this
method by shortening or lengthening the run-lengths of the black pixels
within a facsimile image.
6- Modifications of Half-tone images: different methods have been
implemented to watermark halftone images but those techniques cannot be
applied on document images.
2.4.10 Comparison between Binary Image Watermarking Techniques
Chen et al (2004) have compared the robustness against the aforementioned
attacks, capacity, benefits/drawbacks, and limitations of several watermarking
techniques for document images as shown in Table (2.3).
Chapter 2: Literature Review
30
Techniques Robustness Benefits(+)/ Drawbacks(-) Capacity Limitations
Line Shifting high Low Formatted text only
Word Shifting medium Low/Medium Formatted text only
Bounding Box
Expansion medium - sensitive to document skewing Low/Medium Formatted text only
Character Spacing medium
+ can be applied to languages
with no clear cut word
boundaries
Low/Medium Formatted text only
Fixed Partitioning
Odd/Even pixels none
+ can be applied to binary
images in general High
Fixed Partitioning
percentage of
white/black pixels
Low/ medium
+ can be applied to binary
images in general
-image quality may be reduced
High
Fixed Partitioning
logical invariant none
+ embeds multiple bits within
each block
+ use of a secret key
High
Boundary modification none
+ can be applied to binary
images in general
+ direct control on image quality
High
Modification of
horizontal stroke width medium Low/Medium
Languages rich in
horizontal strokes only
Intensity modulations
of sub-character
regions
medium Medium Greyscale images of
scanned documents only
Run-length modification none -image quality may be reduced
Use two-dithering
matrices none High Half-tone image only
Embed data at pseudo-
random locations none High Half-tone image only
Modified ordered
dithering none High Half-tone image only
Modified error diffusion none High Half-tone image only
Table 2.3 : A Comparison between Several Watermarking Techniques for Document Images (Chen et al, 2004)
Chapter 2: Literature Review
31
2.5 Methods in Literature
A number of researchers have shown interest in data hiding techniques for digital
images. Their ideas and experiments are reviewed in this section.
Several data hiding techniques have been proposed in the literature for colour,
greyscale, half-tone, and binary images. Date hiding methods for images will be
discussed in the next two sections. Colour and greyscale data hiding methods will be
presented in the first section. The second section presents data hiding methods for
binary and half-tone images. Other data hiding techniques are discussed in section
2.5.3.
2.5.1 Colour and Greyscale Image Data Hiding Methods
Data hiding techniques for colour and greyscale images, as proposed by different
researchers, are discussed in this section. Each pixel in a colour image is represented
by up to 3 bytes while a single byte is needed to represent a pixel in a greyscale
image. Spatial domain data hiding techniques are discussed and evaluated first.
Next, frequency domain watermarking techniques are shown. Hybrid data hiding
techniques are presented in the third part. In the final section, data hiding
techniques using different embedding strategies are discussed.
2.5.1.1 Spatial Domain Data Hiding Techniques
Yeung and Mintzer (1997) proposed an invisible watermarking technique, for 24-
bit and greyscale image verification, based on embedding a small size image (logo)
repeatedly into the LSBs of the host image (to be protected) which must have larger
coordinates to hold the watermark. In the verification process, the method uses a
verification key, which is previously created during the embedding operation, to
Chapter 2: Literature Review
32
extract the embedded watermark and to recognise the parts of the image which have
been tampered with. Obviously, this invisible watermark is fragile and unable to
survive regular image processing operations as well as the D/A conversion because it
embeds the watermark in the LSBs of the host image and those bits can easily be
flipped from 0 to 1 or vice versa when a pixel value is modified. Also, this method
cannot be used for binary images because each pixel is represented in a single bit in
black & white images.
Fridrich et al (2000) proposed a fragile authentication watermarking technique for
images. The method aims to avoid the weaknesses of the scheme proposed by
Yeung-Mintzer (1997) by substituting the binary look-up tables of individual pixels
for an encryption map of local neighbours to prevent counterfeiters from estimating
both the key binary function and the logo used repeatedly in different images. Also,
it determines whether an attacker has pasted together sub-images from different
source images while retaining their locations in the attacked image (collage attack).
Fridrich and Coljan (1999) proposed a self embedding image watermarking
technique to retrieve the original contents after tampering. The technique is based
on hiding the DCT coefficients of an image inside the image itself.
In the watermark embedding process, the original image is divided into 8x8
blocks and a DCT transform and quantization are applied to each block. Then, the
first 11 low-frequency values, in zigzag order, of the quantized blocks will be
encoded into 64 bits and embedded in the LSBs of blocks in the original image
provided that the host block does not hold its quantized values.
In this case, if an important part of an image has been distorted, such as the
name of a licensed person in a certificate, it is possible to recover these items by
extracting them from an unaltered part of the same image. Two LSBs are also used
in order to enhance the data hiding capacity and the quality of hidden images.
Chapter 2: Literature Review
33
On the other hand, it is not possible to reconstruct the tampered part if the
embedded code area has also been attacked. In addition, this technique is fragile
and any image processing operation which alters the LSBs of the watermarked image
can remove the watermark.
Wong (1998) introduced a public key watermarking method to verify the
authenticity of greyscale and colour images in the spatial domain. The method
basically inserts the watermark into the LSBs of greyscale images. In colour images,
the watermark can be embedded into the Red-Green-Blue (RGB) palette or into the
Luminance-Bandwidth-Chrominance (YUV) colour components. A binary image
having the same dimensions as the host image is used in this method as a
watermark.
In the embedding process, the LSBs of the original image X are flipped to zeros
to form X’. After that, both X’ and the binary watermark image B are divided into
blocks fixed in size. Each block belonging to X’ is then hashed with a cryptographic
hash function H. The Exclusive-OR (XOR) operation is then applied between each
output block of the hash function and its equivalent watermark block. The output
blocks are then encrypted with a private key K’ and the resulting binary grid is then
embedded in the LSBs of X’ to generate the watermarked image Y.
To extract the watermark, the LSBs of the watermarked image Z, which could
have been tampered with, are extracted first to form the binary grid G which is then
divided into blocks. Every block is then decrypted by the public key K to produce the
binary matrix U. The LSBs of the watermarked image are then set to zero to
generate Z’ which is also divided into blocks. A hash function is then applied on each
block to produce the 2D-matrix Q. The watermark O is then extracted by applying
the XOR operation between the hashed matrix Q and the decrypted matrix U.
Chapter 2: Literature Review
34
If the extracted watermark is the same binary watermark inserted earlier, the
test image has not been altered. The inserted binary watermark image should be the
logo, name, or signature of the legal owner or originator of the image. The Message-
Digest algorithm-5 (MD5) was used as the hash function in this method while Rivest,
Shamir, & Adleman (RSA) algorithm was the chosen public key encryption algorithm.
This method embeds invisible fragile watermarks in greyscale and colour images
for the purpose of verifying image authenticity. It can detect any modifications
occurring in the watermarked image such as image resizing or changing pixel values.
However, this method is certainly not applicable on binary images and has a limited
range of applications because it is based on hiding in the LSBs.
Mohanty et al (1999) proposed a dual watermarking technique to protect the
copyright and ownership of colour and greyscale images. The method embeds two
watermarks (visible and invisible respectively) into the host image to create the dual
watermarked image. The visible watermark should be inserted first because it may
destroy the invisible watermark if inserted later.
To embed the visible watermark, both the original image (I) and the
watermarked image (W) are divided into blocks of the same size. The mean and
variance of each block are then computed and used to adjust the scaling and the
embedding factors for inserting the watermark image into the original.
The embedding process is applied in a block-wise manner and there are some
necessary restrictions in order to avoid serious distortion to the content of the
original image and degrading of its quality. In order to keep the original details of the
image, edge blocks must not be altered significantly. Blocks with high variance of
colours are subject to more modification than those with uniform intensity because
they are less sensitive to modifications.
Chapter 2: Literature Review
35
Afterwards, the invisible watermark is randomly inserted into the visible
watermarked image in the spatial domain. The most significant bits (MSBs) of the
original image pixels are chosen first to hold the watermark binary sequence.
However, if this would cause a significant distortion that increases the SNR
significantly (a preset threshold is used), less significant bits are used instead to
carry the watermark. The extraction of the invisible watermark is required only if the
visible watermark was removed or to detect unintentional modifications to the dual
watermarked image.
This technique is applicable only on colour and greyscale images and cannot be
used to protect black and white images because it is based on hiding the watermarks
in the colour intensities or the bit-plane of the original.
Li et al. (2003) proposed a content-based watermarking scheme to verify the
integrity and authenticity of images. The method simply divides the original image
into fixed-size blocks, extracts features from the 7 most significant bits (MSBs) of
each block, and represents each pixel in a single bit watermark. The watermark
blocks are then blended by replacing the right half of each block with the right half of
its neighbour in a zigzag path in order to make the scheme able to locate any
modification to the watermarked image that may occur later.
Afterwards, the blended blocks are encrypted with the private/public RSA
algorithm to increase the security of the method. Therefore, if an attacker uses the
same watermarking method, with a different RSA key, to create a false watermark
and then resends this image to the intended recipient, the received image will fail to
be unauthenticated by the verification system when the RSA public key is used. The
encrypted binary blocks are then embedded in the least significant bits (LSBs) of the
image to form the watermarked image.
Chapter 2: Literature Review
36
In the verification process, the LSBs of the watermarked image are extracted,
divided into blocks, and encrypted with the public RSA key. Next, the watermarked
blocks are recombined. The extraction of image features from the MSBs as done
during embedding is also required. Finally, the two extracted watermarks are
compared against each other. If they are identical, the image is authentic otherwise
the image has been tampered with.
The method is simple and there is no need for sending prior information about
the original, such as image size, to the verifier. Only the public key is required during
the verification process. Adding objects, using a wrong RSA key, equalizing the
histogram, scaling, and cropping are five different attacks applied on an image in Li’s
experiment and the method has not confirmed the authenticity of the altered images.
Also unwatermarked images have been tested and rejected by the verifier. Only the
intact watermarked image was verified.
However, this method has some disadvantages as it is applicable only on colour
and greyscale images, so, binary images cannot be watermarked by this method.
Also, it is very sensitive, in the sense that it cannot verify the integrity of images if
ordinary image processing tools, such as compression and brightness/contrast
adjustment, are applied to the watermarked image despite the fact that the visual
details of the image are not affected. Therefore, it certainly fails to verify images
after D/A conversion.
2.5.1.2 Frequency Domain Data Hiding Techniques The Anti-Forgery Identification Document (AFID) is a technique proposed by
Chow et al (1993) and designed to protect IDs against forgery. AFID uses a
private/public key encryption method and barcodes to generate and verify
identification cards.
Chapter 2: Literature Review
37
The construction process of an identification card consists of a number of stages.
First of all, an asymmetric RSA encryption algorithm is used to generate two 592-bit
encryption keys, a private key used in creation and a public key used during
verification. Some information such as the card holder’s name and the expiry date of
each card produced must be stored in a card database file to be used for
authentication.
The “Picture Descriptor” is extracted from the card holder personal photograph
which usually appears in one of the corners of the card. The picture descriptor is
generated by dividing the photograph into parts and calculating the gray level of
each part. Next, the gray level of each part is encoded into three different numbers
using two pre-determined thresholds. Any value less than the first threshold is
considered as black, any gray level higher than the second threshold is set to be
white, while other values between these two thresholds are treated as gray. The
vector is then encrypted with the private key to produce the picture descriptor.
Afterwards, the picture descriptor is represented in a two-dimensional barcode
containing 80 bytes of data (20 bits of height x 32 bits of width); the barcode is then
attached to the generated ID card.
For verification, a tested card must be scanned and the same descriptor vector
should be extracted from the photo. Also, the encrypted vector previously embedded
in the barcode must be extracted and decrypted using the public key given to the
verification station. The Euclidian distance between the two vector descriptors is then
calculated. The scanned ID card is determined to be a forgery only if the distance is
greater than a preselected threshold; otherwise, the authenticity of the ID is
certified. Therefore, attempts to replace an original photo with a forged one will be
unsuccessful. The biographical information printed on the surface of the card can
also be a counterfeiter’s target; though, any inconsistency between the information
Chapter 2: Literature Review
38
written on the card and the data extracted from the security seal will result in
rejecting the authenticity of the card.
The AFID technique depends on the security level of the public key; therefore, a
secure cryptography method must be carefully chosen to safeguard the secret
information in the seal. Also, the limited capacity of the barcode, which holds the
secret information, may decrease the reliability level of this technique.
Fridrich (1998a, 1998b) proposed an invisible robust watermarking method to
detect malicious changes in digital colour images. It divides the original image into
medium size blocks (e.g. 64x64) and inserts spread spectrum signals of the image
contents into the middle frequencies of the DCT coefficients of those blocks. The
embedding procedure of this method can be used as copyright marks for images in
digital cameras as it requires only small memory size and a short computational
time. It can detect some undesirable alterations such as cropping and replacements
while regular image processing operations, for instance a slight increase/decrease in
image sharpness/brightness, are undetected.
Nevertheless, it cannot be used to protect binary images because it uses the
frequency domain to hide watermarks. Also, its data hiding capacity and its
robustness are dependent on the characteristic of the host blocks where blocks with
a plain area do not have as large as embedding capacity to retain the watermark as
do textured blocks.
Lan and Tawfik (1999) have proposed self-embedding watermarking technique in
order to detect tampering and to retrieve the corrupted areas in colour images. It
embeds some important parts, expected to be attacked, of the image into the DCT
coefficients of the image itself. The data hiding capacity of their algorithm is about 1
watermark bit per 167 bits of the host image. However, the embedding process
cannot determine the visual distortion; therefore, users of this technique are
Chapter 2: Literature Review
39
supposed to stop embedding more data once there is perceptible artefacts. Also, this
technique cannot be applied on binary and halftone images because it is based on
the DCT.
Self-Authentication-and–Recovery Image Watermarking System (SARI) is a semi-
fragile technique designed by Lin and Chang (2001) and used to authenticate
images. SARI can identify malicious attacks, discover ruined areas and almost
recover original regions by extracting hidden information about the original image
from the attacked watermarked image itself.
In the embedding process, this system inserts imperceptible watermark in digital
images, while in authentication and recovery it detects many malicious attacks
excluding some desired operations such as JPEG lossy compression. Also, SARI can
locate the modified parts of the watermarked image and may be able to reconstruct
the affected areas. The free trial version of SARI software and its authenticator are
available online at: http://www.ee.columbia.edu/sari/auth.html. The watermark in
this technique seems to be designed to survive JPEG compression attack provided
that the quality of the compressed image is not visually corrupted.
Kim and Oh (2004) proposed a digital watermarking algorithm for text-based
images that embeds the watermark in edge-direction histograms of the greyscale
document image. Their method initially divides the image into sub-images of words,
lines or fixed-size blocks and afterwards constructs the edge direction histograms
quantized into 16 levels for each block. The method has been applied to 3 text-
images written in English, Korean and Chinese languages. Subsequently, it
designates one block or more as “mother block” while other blocks are considered as
“child blocks”. One bit of the watermark is inserted into each child block by enlarging
or shortening the histogram directions of the chosen block depending on the binary
Chapter 2: Literature Review
40
value of the watermark bit. However, the mother block must not be modified. The
peak signal-to-noise ratio PSNR of this method approximately is 13.5 after insertion.
The watermarked images were tested by applying various attacks such as
rotation, noise addition, column deletion, blurring, binarization and sharpening and
showed plausible robustness. However, it may fail to extract the inserted watermark
after a print/scan operation. Kim and Oh (2004) believe that their algorithm is
functional in some practical areas, for instance, digital libraries.
Hsieh and Tseng (2004) proposed a Discrete Wavelet Transform (DWT) based
watermarking technique to embed a greyscale image into another image
proportionally larger in size.
In embedding, a 3-level DWT sub-band decomposition is applied on both the
small greyscale image (watermark) and the large image (host). Then, the DWT sub-
band coefficients of the watermark are inserted into the high context energy DWT
coefficients of the host to achieve a high level of robustness as well as
imperceptibility. The Inverse Discrete Wavelet Transform (IDWT) is then applied on
the modified host sub-bands to reconstruct the watermarked image.
For detection, a similar 3-level DWT is applied on the watermarked image to
extract the watermark sub-band coefficients using an extraction key. Afterwards, the
IDWT of the extracted sub-band must be computed to reshape the extracted
watermark. The existence of the original watermark is required to assess the quality
of the extracted image by using the peak signal-to-noise ratio (PSNR).
Their technique has been tested and has shown a considerable level of
robustness to JPEG compression, sharpening, blurring, and smoothing operations.
However, this technique is not applicable on binary images and is not robust to print-
and-scan attacks.
Chapter 2: Literature Review
41
Sharkas et al (2005) also proposed a DWT-based watermarking technique to
protect the copyright and to prove the ownership of digital colour images. To insert a
watermark, a 2-level DWT is applied on the original image. Then, the watermark
image is embedded in the horizontal, vertical, and diagonal wavelet coefficients of a
host image which must be larger in size. Inverse DWT is required to produce the
watermarked image.
In detection, a 2-level DWT is applied, as in the embedding, and a key is needed
to extract the watermark from the modified parts of the watermarked image is DWT
coefficients.
Compression, low pass filtering, salt and pepper noise, and luminance changes
were the four different image processing operations applied on the watermarked
image in order to test the robustness of this method and the watermark was
successfully extracted from the attacked images. However, this method cannot
protect binary images as it is based on hiding large amount of data in the frequency
domain of colour images. Also, it could fail to prove the ownership of colour images
after a print/scan attack.
Han et al (2005) proposed a new watermarking algorithm for image copyright
protection based on Wavelet Transform. The technique embeds a 2D Data-Matrix
barcode as a watermark into the high and middle frequency subbands of the wavelet
coefficients to conserve the quality of the image and to avoid any visible artefacts
caused during embedding. IDWT is applied after the insertion process to construct
the watermarked image.
The method shows satisfactory robustness against JPEG and The Set Partitioning
in Hierarchical Trees (SPIHT) where 90% of the watermarks embedded in test
images were successfully extracted after compression. The 2D Data-Matrix barcode
is chosen because of the fact that the embedded barcode watermark can be
Chapter 2: Literature Review
42
reconstructed even if up to 60% of the barcode is damaged provided that a Reed-
Solomon error correction code is used. More details about the 2D Data-Matrix
barcode are given in chapter 3.
However, this algorithm is unsuitable for protecting binary images because it is
based on the Wavelet transform. More investigation is required to make this method
applicable to moving photos and audio files. Also, it has not been tested for other
kind of attacks.
Sutcu et al (2007) proposed a wavelet transform based method to detect
tampering in colour and greyscale images. The method can identify
sharpness/brightness adjustments and region replacements applied to images by
counterfeiters. The regularity of the image must be estimated by applying an edge
detection algorithm first to find out the locations of edges in the image. After that, a
4-level undecimated discrete wavelet transform (UDWT) of each row and column of
the image is computed separately. The UDWT is the discretized version of the
continuous wavelet transform. Then, the row-based and the column-based linear
curve fittings of the highest amplitude values of the UDWT coefficients are measured.
The average of these two fittings represents the final sharpness/blurriness measure.
In detection, if this value increases or decreases significantly, the test image will be
considered a forgery.
However, this method works only on colour and greyscale images and cannot be
used for binary images. Also it can detect sharpness/brightness adjustments and
copy/paste attacks but it may fail to detect counterfeiting if other image processing
tools are used.
Chapter 2: Literature Review
43
2.5.1.3 Hybrid Data Hiding Techniques
Fridrich (1999b) presented a hybrid watermarking technique that embeds two
watermarks (robust and fragile) to detect tampering in colour images. A robust
watermark is firstly inserted in the DCT coefficients of the transformed image and it
is used to detect obvious unwanted modifications such as part replacement of the
watermarked images with fractions of other images. Then, a fragile watermark is
embedded in the LSBs of the image grid without causing any significant distortion to
the previously added watermark. The second embedded watermark discovers all
simple image processing operations that could be applied to the watermarked image.
In detection, the fragile watermark can be extracted directly from the LSBs while
IDCT is required to extract the robust watermark from the DCT coefficients.
However, this technique needs more improvements to be robust against geometrical
operations such as cropping and resizing. Also, it cannot be applied on binary images
because it is based on utilizing the LSBs and the DCT coefficients of the image.
Tsai et al (2007) proposed another hybrid feature-based invisible image
watermarking technique for copyright protection and integrity verification. The
Hessian-Affine feature detector is used to locate the robust feature spots in the
image. Next, the image regions are divided into two groups, one for the copyright
signature and another for the verification code. Due to their resistance capability to
image processing operations, the detected robust regions are chosen to embrace the
copyright signature while the other parts of the image are used to carry a verification
fragile watermark in their LSBs. The two groups are then combined to form the
watermarked image. A secret key and a hash function are used during embedding
and extraction. The embedding and detection schemes are illustrated in figures (2.1)
and (2.2) respectively.
Chapter 2: Literature Review
44
Original Image
Hessian-Affine
Detector
Region Selection
Copyright insertion
Authentic insertion
Watermarked image
Watermarked Image
Hessian-Affine
Detector
Region Selection
Copyright detection
Authentic detection
Ownership verification
Content Authentication
Figure 2.1 : The Block Diagram of Watermark Embedding Scheme (Tsai et al, 2007)
Figure 2.2 : The Block Diagram of Watermark Detection Scheme (Tsai et al, 2007)
Once again, this algorithm is not suitable for binary images because it utilizes
the LSBs of the host image to hide the watermarks. Moreover, the watermark can be
easily corrupted by the D/A conversion even if the watermarked image is not
tampered with. The time complexity of this algorithm is high due to the repetitive
process of the Hessian-Affine detector.
2.5.1.4 Other Data Hiding Techniques
A content-based digital signature method for image authentication was proposed
by Schneider et al. (1996). Continuous authenticity measures are used in the
verification process of this method to create more flexibility to confirm the
authenticity of images that have been manipulated with JPEG compression or noise
reduction tools without affecting the visual objects of the watermarked image.
Therefore, the tested image could be completely authentic, partially authentic, or
unauthentic. A threshold needs to be set in order to obtain optimal verification
results and this should be variable depending on the requirements of the system.
Chapter 2: Literature Review
45
To generate the signature, the original image is firstly divided into blocks and
the intensity histogram of each block is then computed. After that, the Euclidean
distances between intensity histograms are calculated. Subsequently, the data size is
then reduced by hashing the histograms with a hash function fh. Afterwards, the
hashed data is encrypted with a secret cryptographic key Kpr to produce the
signature S.
In verification, the same steps done previously in the signature creation stage,
excluding the encryption, must be applied on the test image. The signature S is then
decrypted by the public key Kpu and compared with the hashed data of the test
image using a threshold tau. The test image is authentic only if the difference
between the values of the feature vectors is smaller than tau.
The threshold can be used to reduce the sensitivity of the verifier and
experiments are needed to set the optimal value of tau. For example, to make this
method robust to JPEG compression, the highest difference between vector values of
both the original and compressed image must be computed. The threshold must
have a value equal to or higher than the maximum difference. However, the
threshold must be set to zero if the hash function is used.
The main drawback of this method is that the extracted signature S is not
embedded into the image itself and it must be sent separately to the verifier which
reduces the security level of this method. Also, it requires a large database to save
the encrypted histograms of images. In addition, this method cannot be used to
verify binary images because it is based on computing the image histogram and this
would just be a count of the frequency of 1s and 0s.
Chapter 2: Literature Review
46
2.5.2 Binary and Half-tone Image Data Hiding Methods
Despite the digital revolution, paper documents are still required in business and
offices and these documents are normally printed in black and white (Zhu et al,
2003). Also, binary document images have been widely-used in daily life as
immigration papers, birth certificate, bank cheques, books, deeds, music scores, or
other official documents. However, counterfeiters can exploit cheap scanners,
printers, photocopiers and other office devices to create forged copies to attain their
malicious objectives (Chen et al, 2004).
There is more flexibility and capacity to embed watermarks in colour and
greyscale images because each pixel in those types of images is represented by
more than one bit, for instance each pixel in true colour images is represented by 24
bits. Therefore, the majority of data hiding techniques proposed in the literature are
suitable for colour and greyscale images but those techniques are not applicable to
binary images that have an obvious division between foreground (texts) and
background regions. This lack of binary image data hiding methods is a result of the
dual nature of binary images where each pixel is represented by a single bit, 0 for
black and 1 for white. Consequently, any alternation in binary images typically
causes perceptible artefacts and visual irregularities such as salt and pepper noise.
(Furht et al, 2005; Yang and Kot, 2004; Lu et al, 2003; Mei et al, 2001; Low et al,
1998; Arnold, 2003; Kim and Oh, 2004).
Half-tone images are black and white images which give the impression of being
greyscale images if seen from a distance. Some of the images shown in newspapers,
fax documents, and books are half-tone images. Dispersed-dots and Clustered-dots
are two different types of half-tone images. However, they differ in visual quality
where clustered-dots half-tone images have poorer quality than dispersed-dot
images. The process of converting greyscale images to binary images is called half-
Chapter 2: Literature Review
47
toning (Kim & Afif, 2003; Kim & Afif, 2004). The techniques proposed to hide data in
binary and half-tone images will be reviewed in this part of the thesis.
The major difficulty in implementing a reliable authentication technique for
printed documents is the trade-off between the robustness of the inserted
authentication label and the ability to avoid false verification due to the print/scan
operation causing noise in the scanned document which results in false-negative
detection (Sun et al, 2001).
Low et al (1998) proposed a method to watermark document images by shifting
words in the original document slightly left or right or by shifting a whole line up or
down insignificantly as well as to identify the embedded watermark in those images
by using centroid detection. The watermark in this method can be easily removed by
retyping the text or using scanners with Optical Pattern Recognition (OCR).
In their detection process, the horizontal profile of lines must be compiled to
detect line shifts, or alternatively, it is essential to compute the vertical profile in
word shifts, and afterwards a comparison between the centroids of profiles of the
original and the watermarked text images should be made to obtain watermarks.
The experimental results in the presence of noise caused by printing,
photocopying and rescanning show that the performance of centroid detection for
line shifts is considerably better than that for word shifts.
Mei et al (2001) proposed a method to watermark binary document images by
inserting data in the 8-connected boundary of each letter. Data can be embedded, in
this technique, by matching and replacing a selected set of connected boundaries
with a predefined pattern stored in a lookup table to indicate 0 or 1. The data can be
easily extracted by using the same method without referring to the original
document image. The substantial level of imperceptibility has been achieved as data
Chapter 2: Literature Review
48
are inserted in non-smooth border edges. The capacity of data hiding depends on the
resolution of the image and it can be increased by using the inner boundaries in
addition to the outer lines. It can be useful for adding short labels to document
images.
Lu et al (2003) proposed a watermarking technique using the Distance Reciprocal
Distortion Measure (DRDM) which is used to select the pixels to hold a watermark
with the lowest visual distortion in the watermarked image.
The method also uses a 2D shifting technique and odd-even scheme in the
embedding process. The 2D shifting is essential to make tampering detectable in
extraction. The experiments show a high rate of imperceptibility and easy extraction
of the watermark.
Zhu et al (2003) proposed a novel print signature technique to embed a unique
signature in each printed copy alongside embedding information about document
contents. The method includes registration and authentication procedures.
In registration, the digital signature is extracted from certain features in the
document as well as a unique document identifier which is derived from critical
information about the document to be protected. Then, both signature and identifier
are added to a blank area of the document as a barcode; In addition, some
supplementary landmarks must also be printed for the purpose of alignment. Only
documents which contain inserted signature and identifier can be circulated.
The authentication procedure is used to determine whether the printed document
has been tampered with or not. It extracts the signature and the identifier from the
scanned document as done in the registration process and then compares them with
the information attached to the same scanned document as a barcode. The
document is accepted as genuine if the extracted features are alike while it is
considered as a forgery if not.
Chapter 2: Literature Review
49
This technique might be functional in online billing, voting ballot, and e-ticket
systems and it has been tested by using two HP LaserJet printers (8100 and 4600) to
print out documents on plain, translucent, card, and coloured papers. It has
successfully rejected all the 50 forged documents. On the other hand, some original
documents have not been accepted. Another weakness is that photocopied
documents have not been tested due to the poor quality of those documents. In
addition, more investigation is required to increase the robustness of the method
against photocopying and other acceptable modifications.
Data Hiding by Self Toggling (DHST) is a simple data hiding technique suitable for
half-tone images. It is based on creating an unduplicated set of pseudo-random
numbers representing random locations in the image (I), and then replacing the
value of this location in the image I with a bit of the watermark. The probability of
changing a bit from 0 to 1 and vice versa is 50%. In extraction, the pseudo random
numbers are needed to extract the embedded message. Salt-and-pepper noise will
result when DHST is used (Kim & Afif, 2003; Kim & Afif, 2004).
Authentication Watermarking by Self Toggling (AWST) is a cryptography-based
authentication watermarking technique for dispersed-dot halftone images proposed
by Kim and Afif (2003). The technique randomly clears a number of bits in a binary
image B using a pseudo-random number generator L to create B*; it then computes
the signature S using the equation below:
S= E (Ĥ) where Ĥ= H (B*) A, where H is a cryptographically secure hash
function, A is the logo image, and E is an encryption method. E could be either a
secret (symmetric) or public key (asymmetric) ciphering method. The obtained
signature S is then embedded into the set of pixels L to generate the watermarked
image B’.
Chapter 2: Literature Review
50
The verification process uses the pseudo random number generator L to extract
the previously embedded signature S from the watermarked image X’ which could be
tampered with. Then, it clears the pixels in L to produce X*, and to compute a new
signature C using the equation C= H (X*) D(S). Where D(S) is the decryption of S
and H is the same secure hash function used in the embedding process. After that, C
is compared with the logo A. If they are equivalent, the image is authentic;
otherwise, it is forged.
However, the size of the host image has to be much larger than the size of the
logo image; otherwise, the watermarked image would be corrupted. Also, salt and
pepper noise could occur if this technique is applied on pure binary images, such as
document images with plain large black or white areas. In addition, AWST uses a
fragile watermarking system; therefore, it is very sensitive to any change
(deliberately or mistakenly) applied to the watermarked image. As a result, it is
wholly inappropriate for authenticating images after D/A conversion as the
embedded watermark is eliminated after such an attack. In addition, AWST can
detect changes in images but it is incapable of locating these alterations.
Kim and Queiroz (2004a) have proposed a new secure authentication
watermarking techniques for binary images. It uses secret-key pr public/private key
cryptography where only generators of document images can embed watermarks
while recipients can verify the image integrity by using the public key. It also uses a
fragile watermarking technique with the intention of detecting any perceptible
malicious alternations.
The embedding process is based on dividing a binary image into two parts with
different sizes; the authentication signature is computed from the first part (large in
size) which while the second (small) is used to carry this calculated signature. A
Chapter 2: Literature Review
51
signature S is derived by applying a hash function on the first part, then encrypting
the results of the hash using a private key.
Undesired alternations to the first part can be identified by this method but it is
incapable of detecting possible changes in the second parts. Practically, it can be
used in document image communication via the Internet and further investigation is
required to adapt this technique to D/A conversion.
Kim and Queiroz (2004b) proposed a cryptographic authentication watermarking
technique for binary images where the original image Z is shuffled randomly with a
seed which is used as a secret key in private-key scheme or overtly distributed in
public/private-key scheme. Then, the shuffled image Ź is divided into two parts Ź1
which holds the signature and Ź2 where the signature comes from. After that, a
secure hash function H should be applied on Ź2 and the result encrypted to generate
the signature S=Ek(H(Ź2)). Finally S should be inserted in Ź1 by chopping Ź1 into
blocks (e.g. 8x8) and finding the lowest visually significant pixel in each block to be
altered to encode a single bit depending on the number of white pixels within each
block, where even numbers refer to 0 and odds are 1’s.
To verify whether the document image is authentic or not , first of all , the
secret/public key should be used as the seed of the random number generator to
shuffle the watermarked images Z̀ . Then, the shuffled image Ź̀ is divided into two
parts Ź̀1 and Ź̀2 as done in the embedding stage. Afterwards, a comparison between
the computed signature from Ź̀ 2 where S2=Ek(H(Ź̀2)) and the extracted signature
S1 from Ź̀1 should be made. The document image is genuine provided that S1 and
S2 are identical; otherwise, the image has been counterfeited or the key is incorrect.
Flipping a single bit in the whole image can be detected by this method.
However, it fails if two pixels within the same block have swapped their values.
Another point is that the data hiding capacity of this technique depends on the
Chapter 2: Literature Review
52
content of the image. Further investigation is still required to adapt this technique to
D/A conversion.
Yang and Kot (2004) proposed a watermarking method to attest the authenticity
of binary image documents. The watermark in this algorithm is generated by a Gold-
like sequence whilst the embedding process takes the smoothness and topography of
the document image into consideration to eliminate the noise and keep the
connectivity in the watermarked document. The blocks of 5x5 neighbourhood pixels
to be flipped are chosen according to pre-defined patterns. In the authentication
process, the original image and the secret key are required to verify the legitimacy of
the text image.
This technique preserves the quality of image after the embedding process and it
could be useful for real time verification as it has short computational time. On the
other hand, the quantity of inserted data is limited and depends on the nature of
host documents. Also it is a non-blind watermarking technique and needs further
investigation to make it blind.
Alattar (2004) proposed a method of fingerprinting text documents for the
purpose of discovering any collusion done by an authorized employee by adding a
unique ID, which refers to each employee, into each given copy. This technique is
appropriate for left, right or centre aligned texts, justified paragraphs, and irregular
line spacing. The embedding process in this technique is based on shifting words or
lines of the texts slightly. It uses a spread spectrum technique to solve the issue of
uneven distances and BCH (Bose-Chaudhuri-Hocquenghem) error correction code to
reduce the noise caused by Digital to Analogue conversion. The original image is not
required during detection where watermarks can be extracted by computing the
distances between lines and words in the text. This algorithm presents a significant
rate of robustness but further research is required to protect inserted data against a
print-and-scan attack.
Chapter 2: Literature Review
53
Xiaoyi et al. (2004) proposed a steganalytic technique to determine whether or
not a binary image contains a secret message. The technique is simple and tests
document images by calculating statistics U of an image I; then, embeds a test
message into the image I to obtain a stego Image I’. After that, the same statistics
U’ should be computed from I’. The block coarseness statistical analysis is used in
this technique to calculate U and U’. Initially, the binary image should be divided into
blocks; then, the edge irregularity of each block can be computed by using the block
coarseness. The Rate R = U’/U is then computed. The image I has already been
embedded with a hidden message if the value of R is approximately 1; but if R is
relatively greater than 1 by a predetermined threshold T, the Image I is plain and
does not contain a secret message. The technique has been tested by checking the
existence of secret messages embedded in binary photos using only two known data
hiding techniques. However, it may fail to detect the availability of embedded data if
other hiding techniques were used. Also, the detection might not succeed after print
and scan attack because objects in binary images could become thicker or thinner
and significant changes may occur to the coarseness of edges.
Puhan and Ho (2005) suggested a watermarking algorithm to authenticate
binary document images. To embed a watermark, the method initially computes the
curvature-weighted distance difference (CWDD) measure for contour pixels in the
original document image and then picks the pixels with estimated distortion values
less than a pre-defined threshold T. Those selected pixels are known as suitable
pixels while the rest of the pixels are non-suitable to hold the watermark. After that,
the suitable pixels are scanned in a sequential order left-to-right and top-to-bottom
to select the reversible pixels. There must be only a single reversible pixel in the
centre of an MxM block in the original image and this pixel should not affect the
condition of the other pixels in this block after inserting a watermark bit in it. These
Chapter 2: Literature Review
54
criteria must be available in those chosen pixels to guarantee a similar scan order
during detection. Therefore, only the reversible pixels are chosen from the entire set
of suitable pixels to carry the watermark. Flipping the reversible pixels from 0 to 1 or
vice versa reduces the visual noise in the watermarked images to maintain the
imperceptibility of the method. The N irreversible pixels are then separated into two
sets SR and SM. SM is then hashed by an MD5 hash function to generate the Hashed
Message Authentication Code (HMAC). SR is used to carry the authentication
message extracted from SM. A secret key is used to embed SM pixels into SR.
In verification, the same procedures as in the embedding process must be done
to the test image to extract 2 subsets SR’ and SM’ and a comparison between pixel
values of those two set is then made. The test image is authentic only if SR’ and SM’
are equal. Otherwise, the document image is determined to be forged.
Document images transmitted or modified digitally can be verified by Puhan and
Ho’s method. However, this method is sensitive and it cannot verify hardcopy
documents after converting them to digital form. The method has limited embedding
capacity because it selects a particular set of pixels from the original image to hold.
Fridrich et al (2005) proposed the wet paper coding (WPC) method which is used
for secret communication and to solve the “Writing on Wet Paper” scenario in
steganography. The above-mentioned scenario can be summarized as follows:
“Imagine that the cover object x is an image that was exposed to rain and
the embedder can only slightly modify the dry spots of x but not the wet spots.
During transmission, the marked image y dries out and thus the detector does
not know which pixels were used by the embedder for data hiding. The task of
wet paper coding is to enable both parties to exchange secret messages under
the above scenario. The problem of data embedding in binary images fits this
Chapter 2: Literature Review
55
“writing on wet paper” paradigm quite well, with flippables viewed as dry spots
and non-flippables as wet spots” (Gou and Wu, 2007, pp. 279).
The modification of flippable bits from 0 to 1 or vice versa during encoding may
change the flippability status of the neighbouring pixels which causes wrong
extraction at the recipient’s side. k flippable pixels must be found in the original
image X which has n pixels to embed the q bits of the secret message M into X to
form the marked image Y. A pseudo random binary matrix D with the size (n x q) is
then generated by using a secret key where M=DY. The matrix D is required during
extraction as the cover image must be multiplied by D to obtain the secret message.
The number of pixels used to hold the secret message is nearly the total number of
flippable pixels.
Gou and Wu (2007) suggested the “super-pixels” concept to improve the
embedding payload of the WPC method proposed by Fridrich et al (2005). The super-
pixels are a set of non-flippable pixels adjacent to a flippable pixel where their binary
values can be altered together without causing any visible distortion to the marked
image. Each neighbouring pixel of the k flippable pixels that matches one of pre-set
patterns can be chosen as a super-pixel. A continuous vertical or horizontal line of
zeros or ones is a simple example of the allowed patterns. The total number of pixels
to hold the secret image increases by t super-pixels. Gou and Wu have tested a
binary document image with the wet paper coding and the super-pixel techniques.
The data hiding capacity increased 6.7% by using the super-pixels approach without
any visual influence on the host binary image. Most of the selected super-pixels were
grouped in vertical or horizontal lines located on the strokes of the characters or
objects.
Chapter 2: Literature Review
56
Both techniques can encode a secret message or a watermark in a binary image.
However, these techniques are sensitive and the embedded data cannot be extracted
if the marked image has been printed and scanned again.
Liu et al (2007) suggested a content-based watermarking algorithm for binary
image authentication. It divides the image into 2 parts P1 and P2 using a random
mask K. The Zernike Moments Magnitude (ZMM) is used to extract a feature vector
Z1 from the first part P1 of the divided image and Z1 is then quantized and embedded
in the other part as a watermark. The time complexity of the ZMM is very high, for
instance, the computational time of a ZMM vector for a 256x256 greyscale image is
approximately 5 minutes using MATLAB on a 1.5 GHz Pentium machine (Liu et al,
2007). In authentication, the same mask K1 is also used to split the watermarked
image to two parts P1 and P2. The feature vector Z2 is then generated from P1 using
the ZMM and quantized. On the other hand, Z1 is extracted from P2. The test image
is authentic only if Z1 and Z2 are equal; if not, it is considered as forgery.
The pixels of P1 and P2 are distributed randomly and close to each other.
Therefore, both will be altered if the watermarked image is modified. As a result, the
extracted vectors Z1 and Z2 cannot be identical and the authenticity of the image will
be rejected. However, the sensitivity of the watermarks generated by this algorithm
makes it unable to distinguish between malicious and non-malicious attacks.
Consequently, it fails to authenticate images after print and scan. Also, it cannot
determine the locations of the modified parts of images.
Chapter 2: Literature Review
57
2.5.3 Other Techniques
Kwok et al. (2000) proposed watermark design patterns that help watermark
technique developers to choose an efficient technique for electronic commerce
copyright protection among available watermarking techniques depending on the
requirements of system and its data type.
Online distribution of multimedia content, broadcast services, document
verification, ownership identification, electronic publishing and advertisement, real-
time information delivery, product ordering, transaction processing, photograph
galleries, digital libraries, and digital newspapers and magazines are some examples
of electronic commerce applications (Kwok et al., 2000). The data types used in
these applications and websites could be text, images, audio, or video.
System for Copyright Protection (SysCoP) is a data hiding algorithm proposed by
Zhao and Koch (1995) for colour, greyscale, and binary images. It divides the
cover image into 8x8 contiguous or distributed blocks and encodes a single bit in
each block provided that the embedded bit does not cause any visual noise to the
host block. A contiguous block is an 8x8 square existing as a unit and taken directly
from the original image grid while the pixels in a distributed block are selected
randomly from the original image without repetition. Distributed blocks make
detecting locations of the embedded labels harder for attackers even if they compare
different labelled images.
In colour and greyscale images, the method applies a DCT transform to each
contiguous block and then each transformed block is quantized. The middle
frequencies of the quantized DCT coefficients are then modified to hold the label bits.
Afterwards, de-quantization and the Inverse Discrete Cosine Transform (DCT) are
applied on the modified blocks to construct the labelled image. The label can be
Chapter 2: Literature Review
58
extracted by applying quantization and DCT transform on the 8x8 blocks of the test
image. This method is supposed to be robust against JPEG compression because it
uses the DCT transform, which is also used in this lossy compression technique, but
it may fail to prove the ownership of the labelled images after print/scan attack.
In binary images, the rate of black pixels in each 8x8 block is computed. Then, a
bit 1 of the label is inserted into blocks which have a black rate greater than a given
threshold and a bit 0 is embedded if the rate of white pixels is less than another
predetermined threshold. The data embedding capacity of this method depends on
the content of the host image. In dithered binary images, the black and white pixels
are distributed all over the image and interweaved. Therefore, dithered binary
images can carry more data than other binary images with distinctive large flat
black/white areas where labels can only be inserted at the borders between black
and white regions. Thresholds must be chosen carefully to increase the robustness of
the inserted label and to guarantee accurate extraction. This method is sensitive to
simple attacks, for instance removing a line from the label image could cause wrong
label extraction. Also it fails to prove the copyrights after D/A conversion or image
resizing attacks.
Sun et al (2001) proposed a content-based optical watermarking technique to
authenticate document images without the need for converting the test document to
digital format. In this technique, the watermark is embedded by modulating the high
frequencies of the document. Visual cryptography with n keys is used in this method
to increase the level of security. The watermarked document can be visually verified
using a photocopier provided that the printer used to produce this document has a
higher resolution than the photocopy machine. The cryptographic keys are overlaid
during verification to distinguish between forged and original documents. The
Chapter 2: Literature Review
59
embedded watermark image will appear if the photocopied document is original.
Otherwise, a noisy image will be the output from the photocopier.
This method could be useful for verifying printed documents. However, it
requires high-priced equipment such as printers with high resolution. Therefore, the
cost could be inconvenient for some organizations which find this technology
unaffordable. In addition, the number of verification sites is limited as only those
who have the cryptographic key, which could be a transparent slide, can verify
documents. Also, photocopiers with higher resolution than printers used in
embedding can cause false document verification.
In Digimarc (2009), watermarking solutions to verify the integrity of document
images, detect unauthorized alternations, and discover the source of leak are
offered. No detail about the proposed technology is given in the Digmarc website.
Signum Technologies, a partner of Digimarc, designed Signum's VeriData
software to check the integrity and validity of imaged documents (Signum
Technologies, 2009). VeriData is said to be commercial low-priced software with a
high level of security to detect unauthorized modification in (Tagged Image File
Format) TIFF and JPG document images. The software is also flexible enough to be
compatible with different systems.
Chapter 2: Literature Review
60
2.6 Conclusion
In this chapter, many watermarking techniques proposed in the literature to
protect copyrights or to authenticate digital images were reviewed. The vast majority
of these techniques are suitable for colour and greyscale images due to the high data
capacity of those types of digital data. Only a minority of data hiding techniques are
proposed for binary images. However, most of them have not shown enough
robustness against print/scan attack. Therefore, those techniques can be used only
to protect digitally transmitted and manipulated binary images.
The lack of techniques to protect binary images against print/scan attacks is due
to two main reasons. First, binary images have low data hiding capacity and altering
the values in a binary image grid can cause visual noise. Secondly, digital-to-
analogue conversion can distort objects in digital images.
Chapter 3: Fundamental Methodologies and Principles
61
Chapter Three: Fundamental Methodologies and Principles
3.1 Introduction A number of essential concepts and basic techniques have been in the proposed
method of this research. Therefore, it is important to give background information
about those principles and techniques. This chapter sheds light on those techniques
and discusses them in details.
3.2 Print and Scan Model The vast majority of image watermarking techniques are fragile to the print/scan
operation. According to Yu et al (2005), the three main difficulties behind the lack of
watermarking techniques proposed to survive digital/analogue conversion are:
1- Randomness: The watermarked image after a print/scan operation looks
exactly like the original watermarked copy when seen by the naked eye but in
fact there are enough differences to damage the embedded watermark. The
variation is due to the inaccuracy of the mechanisms of printers and scanners.
2- Human-dependency: The printed/scanned document could be adjusted one
way or another by the user, printer, or scanner which may eliminate the
inserted watermark. Examples of those desirable or undesirable adjusments
are contrast enhancement, scaling, gamma correction, or orientation of
papers which may occur in the tray of a printer or the flatbed glass of a
scanner.
3- Indistinguishability: The distortion of both printing and scanning processes on
the printed/scanned materials is combined and cannot be separated because
when a paper is printed, it has to be scanned to be analyzed or processed by
Chapter 3: Fundamental Methodologies and Principles
62
computer. Therefore, the changes occurring in both print and scan will be
merged.
A robust watermarking technique to protect hardcopy documents is still needed in
order to change the future of copyright protection and verification technology for
paper media such as books and newspapers. However, the print-and-scan issue is
still not resolved. The Print-Scan model is shown in Figure (3.1) in Yu et al (2005).
Low-pass filter and geometrical distortion are introduced to the document in both
print and scan phases. But, there is more geometrical distortion during the scan
process than in printing.
Figure 3.1 : The Print-Scan Model (Yu et al, 2005)
Print and scan operations, their structure, and influences on the original media
are discussed in detail, separately, in the next two sections.
3.2.1 The Print Process
The Structure of a laser printer is shown in Figure (3.2) in Yu et al (2005). When
a document or an image is printed, distortion is very likely to happen to the printed
media and there are different sources of distortion during the print process:
Halftone
Grayscale Conversion
Attack
Geometric Distortion
Blur
Geometric Distortion
Geometric Attack
Lowpass filter attack
Blur
Print Scan
+ + Output Original Image
Chapter 3: Fundamental Methodologies and Principles
63
1- In the optical system of the laser printer, the lenses are supposed to work
always in linearity. Otherwise, noise could appear on printed materials if
any non-linearity is introduced.
2- The luminance bias and movement inconsistency of the laser source can
also cause noise in the printed materials.
3- Blur of the hardcopy could be caused by dot gain which results from the
optical system and from colorant spreading.
4- Continuous greyscale to binary pattern conversion causes significant
differences between the original and the printed copy.
5- Geometrical image displacement could happen if there is non-uniform
distribution of toner. This displacement will not happen if the Organic
Photo Conductor drum and all the rollers in the Electrophotographic system
are perfectly cylindrical with a smooth surface.
Figure 3.2 : The Structure of Laser Printer (Yu et al, 2005)
Raster Image Processor
Buffer
Actuator
Laser Source
Parabolic lens
F-θ lens
Organic Photo Conductor drum
Developing Roller
Charging Roller
Transferring Roller
Fusing Roller
Pressing Roller
Laser Printer
Original Image
Hardcopy
Control System
Optical System
Electrophotographic System
Latent image
Chapter 3: Fundamental Methodologies and Principles
64
3.2.2 The Scan Process
The scan process also has a major impact on the details of the hardcopy. The
structure of the scanner is shown in Figure (3.3) taken from (Yu et al, 2005). The
mechanism of a scanner can be summarised as follows, the reflector sheds the light,
produced from the light source, onto the whole glass of the flatbed where the
document is placed. The luminance of the reflected light is then converted to signals
by the Charge Coupled Device (CCD) and sent to the ADC system and then to the
PC. Output digital images differ from hardcopies after converting them to a digital
form via a scanner. The variation can happen at different stages:
1- The Modulation Transfer Function (MTF) in the optical system of the
scanner causes a Gaussian low-pass filter attack which blurs the scanned
document.
2- A Charge Coupled Device sensors causes thermal noise and dark
current noise which affects the quality of the scanned image.
3- The stepped motion jitter of the carriage also causes Gaussian random
noise.
4- The scanning process also causes additive and multiplicative noise and
its degree varies from a scanner to another depending on the resolution of
the scanner.
5- Users cannot avoid changes of orientation when a document is placed
on the glass of a flatbed scanner. Most likely, all documents are subject to
rotation in some degree.
Chapter 3: Fundamental Methodologies and Principles
65
Figure 3.3: The Structure of Scanner (Yu et al, 2005)
3.3 Centre of Gravity
The definition of the centre of gravity (also known as centre of mass, or
centroid) is a unique point for a system (one object or more) that represents the
average position of the mass of this system (Pratap & Ruina 2009; Bourke, 1988).
The centre of mass can be computed for standard geometrical shapes such as circle,
triangle, square, rectangular, oval, pentagon, and hexagon as well as for any
asymmetrical polygons. Examples of symmetrical objects are shown in Figure (3.4-a)
and asymmetrical objects are shown in Figure (3.4-b). The centroid of each object is
marked by a dot (•). The centre of gravity of symmetrical objects is located at the
axis of symmetry in the middle of the object.
Figure 3.4: The Centre of Mass of (a) Symmetric Shapes and (b) Asymmetric Shapes
Light Source
Reflector
Charger Coupled Device sensor
ADC System
Drive System
Scanner
Hardcopy Digital image
Optical system
(a) (b)
Chapter 3: Fundamental Methodologies and Principles
66
In engineering, it is essential to calculate the centroid of any rotating piece of a
machine such as a train, car, aeroplane, or boat accurately to stabilize the engine
and reduce or avoid vibration to the minimum level. Otherwise, shaky revolving
objects in a machine may lead to a catastrophe such as an engine explosion.
Nowadays, Computer Aided Design (CAD) software makes it easier for engineers to
calculate the centre of mass of a machine (Pratap & Ruina, 2009). The centroid is
fixed for rigid objects and it can be calculated for any 1D, 2D objects, or higher D.
To calculate the centre of mass of a closed polygon consists of N vertices (xi,yi)
where (0 ≤ i < N) and the vertex (x0,y0) and (xn,yn) are the same point. An
example of a closed polygon is shown in Figure (3.5) in (Bourke, 1988).
Figure 3.5: Closed Polygon with 6 Vertices
The area (A) of the polygon must be computed first by equation (3.1):
The centroid point (Cx, Cy) can be calculated by using equations (3.2) and (3.3)
respectively.
…. Equation 3.1 (Bourke, 1988)
Chapter 3: Fundamental Methodologies and Principles
67
If an object has one or more inner holes, its centre of mass can also be
computed using the above equations, provided that the areas of the inner holes are
calculated and subtracted from the total area of the object.
According to (Bourke, 1988) these equations cannot be used to calculate the
centre of mass of self-intersecting polygons such as those shown in Figure (3.6).
Figure 3.6: Self Intersecting Polygons
This limitation therefore makes this technique inapplicable to some objects which
may exist in the system. In addition, if this method is applied on a digital image
containing many objects with a large number of vertices, the computational time to
detect the vertices and calculate the cenroid of each object in that image will be
long. Therefore, it is better to use a simpler and faster technique to compute the
centre of gravity of objects within an image.
3.4 The Method of Moments
The method of moments is a simpler and faster technique to compute the centre
of gravity of objects within an image. It is also preferable for use as an alternative
way to calculate the centroid of a binary image. Consider all the black areas in the
…. Equation 3.2 (Bourke, 1988)
…. Equation 3.3 (Bourke, 1988)
Chapter 3: Fundamental Methodologies and Principles
68
binary image I as objects (forground) and the white areas as the space
(background). If the mass of each black dot in the binary grid equals 1, the moment
of each dot (x,y) in the grid is the multiplication of the mass m and the directed
distance d from both the x-axis & y-axis as (moment = m × d). The average of the
moments of each dot is then computed to calculate the centroid (Xc, Yc) of the whole
M×N image as shown in Equations 3.4 and 3.5 below (Dai et al, 1992; Li, 1993;
Flusser, 1998; Chung & Chen, 2005):
Where T is the number of black pixels in the image I, f(x,y) is the mass of the
black point (x,y) and it always equals 1, M is the height, and N is the width of the
image I. This technique can be used to work out the location of the centroid of the
whole image or of a particular object in an image if it was processed individually as a
separate image. Its computational time is relatively short.
3.5 Quad Tree
A Quad tree, also known as Q-tree, is a data structure named by Finkel &
Bentley (1974) and defined as “the Simplest and most powerful geometric data
structure” by Har-Peled (2008). The Q-tree is preferred over other data structures
because it is easy to implement and fast for data search. In two-dimensional space,
the root region is partitioned by vertical and horizontal cuts into 4 square or
rectangular quadrants called children. The generated children are then recursively
... Equation 3.5
... Equation 3.4
Chapter 3: Fundamental Methodologies and Principles
69
sub-divided until certain conditions apply. Each node has either 0 or 4 children. A
node that has zero children and cannot be split any further is called a leaf. The depth
of the quad tree is the number of partitioning iterations which generated the very
last leaves (the root node has depth zero) (Har-Peled, 2008; Berg et al, 2008). The
resultant grid does not have to look like a mesh because the quadrants are not
necessarily equal in size or similar in shape. An example of Quad-tree of a 2D image
before and after partitioning is shown in Figure (3.7). The depth levels of Quad tree
partitioning are shown in Figure (3.8).
Figure 3.7: An Example of Quad Tree of an Image (a) before and (b) after partitioning
Figure 3.8 : The Depth Levels of Quad Tree Partitioning of an Image
(a) (b)
Chapter 3: Fundamental Methodologies and Principles
70
Quad trees can be used in a variety of applications such as (QuadTree, 2009):
1. Spatial indexing.
2. Image representation.
3. Efficient collision detection in two dimensions
4. Viewing frustum culling of terrain data
5. Storing sparse data, such as a formatting information for a spreadsheet or
for some matrix calculations
6. Solution of multidimensional fields (computational fluid dynamics,
electromagnetism).
3.6 The 2D Data Matrix Barcode
The 2D barcodes such as DataMatrix ECC200 are more desirable than 1D barcodes
not only because of their data capacity but also due to their robustness. The encoded
data can be recognised even if the 2D barcode is up to 60% corrupted because a
Reed Solomon Error Correction Code is used (Hen et al, 2005).
Applications with a large data capacity labelling requirement such as Direct Part
Marketing and Product Marketing use 2D barcodes due to their large data capacity.
The 2D barcode is constructed of 4 main components as shown in Figure (3.9):
1- Solid-Line Locator: this is an L-Shape outline that shows the proper
direction of the code and the boundaries of the data area
2- Patterned-Line Locator: two black and white lines making a corner at the
side opposite the solid-line locator lines. They specify the number of rows
and columns in the barcode as well as the boundaries of the data area.
Chapter 3: Fundamental Methodologies and Principles
71
3- Data Area: the binary data is repeatedly encoded in this area to guarantee
successful decoding even if some parts of this area have been corrupted.
4- Quiet Zone: is a vacant region with at least 1.5 inches breadth surrounds
the entire barcode; therefore, no data are included in this area.
Figure 3.9: The 4 Components of the 2D Data Matrix Barcode
The 2D data matrix barcode is preferred for several reasons (The IET, 2005):
1- High data capacity: can hold an abundance of encoded data (e.g. 3116 digits
or 1556 American Standard Code for Information Interchange (ASCII)
characters in a small area.
2- Flexible encoding: all kinds of data such as numbers, texts or ASCII code can
be encoded into the data matrix.
3- Adaptable: as barcodes with only 20% contrast can be correctly read, it can
be printed on any type of material.
4- Robust: Solomom Error Correction Code is used in the most widely used ECC-
200 2D barcode; thus, encoded data even in the most of damaged barcodes
can be properly recognized by barcode readers.
5- Orientation-independent: barcode readers can read encoded data in rotated
barcodes at arbitrary rotation angles.
6- Multiple-asset automatic identification: hundreds of unique codes can be
encoded and detected from a single image. Furthermore, barcode readers
have high speed in recognizing encoded data.
1- Solid-Line Locator 2- Patterned-Line Locator 3- Data Area 4- Quiet Zone
Chapter 3: Fundamental Methodologies and Principles
72
In this research, the 2D date matrix barcode has been chosen for the following
reasons:
1- The limited data hiding capacity in binary images as each pixel is represented
by a single bit.
2- The large amount of noise which could occur in the host binary image after
embedding. This noise may spoil the content of the original document.
3- The sensitivity of data hiding techniques to a print⁄scan operation which can
significantly change the grid contents of the binary image.
4- The visible attached barcode may deter the counterfeiters from altering the
protected document.
3.7 Type I and Type II Statistical Errors
Type I and Type II statistical errors were used in this thesis to describe the
inaccuracy of the verification system of the proposed methods in this thesis. These
types of errors show how the verifier can accept or reject a scanned document. The
statistical glossary by Easton & McColl (1997) shows that the two possible errors can
be either:
1- Type I error (Also known as Type-α or False-Positive error) when a null
hypothesis is rejected when it is actually true. In this research, false positive
errors happen when an authentic scanned document is rejected and considered
as a forgery by the verifier.
2- Type II error (also called Type-β or False-Negative error) when a null
hypothesis is not rejected when it is actually false. In the methods proposed in
this thesis, a false negative error happens when a forged document is accepted
and treated as a genuine copy by the verifier.
Chapter 3: Fundamental Methodologies and Principles
73
Table (3.1) in Rogers & Link (1998) shows the statistical errors for true and
false hypothesis test.
Null Hypothesis True
Null Hypothesis False
Reject Null Hypothesis Type I error Correct
Fail to Reject Null Hypothesis Correct Type II error
Table 3.1: Hypothesis Test in Statistics
The methods proposed to validate document images are discussed in detail in
the next chapter.
Chapter 4: Methodologies
74
Chapter Four: Methodology
4.1 Introduction
In this chapter, two different methods to create and verify self-validating
documents are proposed and the theory of each method is discussed in detail. The
advantages and disadvantages of each method and the similarities and differences
between these two methods are also given.
4.2 Method 1
In this method, document images need to pass through the creation and
verification stages. To create a self-validating document, the preservative data are
derived from the area of interest in the document, encoded to a barcode, and then
the barcode is attached to the document itself. In the verification stage, the
preservative data is extracted from both the document contents and the attached
barcode and a comparison between those extracted preservatives will decide
whether the scanned document has been significantly altered or not.
4.2.1 The Creation of Self-validating Documents
To generate a self-validating document, a digital document has to pass through
several stages. First, the area that counterfeiters may aim to alter, in order to
achieve illegitimate benefits, needs to be manually chosen. The selection of this area
has to be done by the creator of the document. This area may contain the signature
of a person, the stamp of an organisation, or any numbers, text, or a combination of
Chapter 4: Methodologies
75
both that symbolize a name, a date, an amount of money, or any valuable
information in a document.
The selected area of the document image is then converted to a binary format if
the creator has accidently chosen an image that is not already in black and white
format. The binarization can be simply applied on a greyscale image I by setting any
pixel with a value less than a predetermined threshold (Tb) to a black pixel (set to
zero) in the binary image Ibinary. Otherwise, the pixel is converted to a white pixel in
the new binary grid. As shown in equation 4.1.
In true colour images (24-bits), the luminance of the colour image must be
calculated first using equation (4.2) to convert the RGB palettes to a grey-level
image Y (Koschan & Abidi, 2008; Pratt, 2007):
Y= (0.299×R) + (0.587×G) + (0.114×B) … equation (4.2).
To convert the values of the Y grid ranging between 0 and 255 to a binary-level
image, Equation 4.3 is used. The threshold can be set by the software designer
depending on the system requirements:
The binary text image is then scanned horizontally in order to detect the number
of lines in the image. If there is more than one line, the image is then split into
separate lines. The detection of lines is based on finding at least a white line (i.e.
… Equation (4.3). Ibinary(x,y) =
0 if 0 ≤ Y(x,y) < Tb
1 if Tb ≤ Y(x,y) ≤ 255
… Equation (4.1). Ibinary (x,y) =
0 if I(x,y)< Tb
1 if I (x,y)≥ Tb
Chapter 4: Methodologies
76
without any single black pixel) between any two sets of rows containing black pixels
and separating those two sets of rows into individual images. Afterwards, each line is
divided into separate characters or objects by scanning it vertically. The principle of
an objects/characters scanner is similar to the lines separator. If there is a single
white vertical line, or more, between a character and another neighbouring
character, these two characters are split into 2 separate sub-images to be processed
individually later. Unlike the connected component analysis in Tan et al (2002) which
only divides only lines into separate parts if they are connected, this method divides
lines into objects if there is at least one white line between characters.
The centroid (Xc,Yc) and the black pixel rate (Rb) of each character/object is then
computed, the values of Xc, Yc, and Rb range between 1 and 100%. An example of
computed Xc, Yc, and Rb values for the letter ‘O’ is shown in Figure (4.1).
Figure 4 : An Example of Computed Centroid Point (Xc, Yc) & the Rate of Black Pixels of a Binary Image Containing the Letter “O”.
The value Rb has a range between 0 and 100%. If the image is totally white (i.e.
there is no object in the image), Rb=0. But if it is totally black, Rb=100. Otherwise,
Rb can be any number ranging from 1 to 99%.
As Xc & Yc represent the location of the centroid of an object, or set of objects, in
an image, the values of Xc & Yc must be in the range 1 to 100% because the centroid
point has to be inside the image itself. If the centroid is in the top right corner, then
(Xc=1% and Yc =1%). If the centroid is the last pixel in the right bottom corner, then
(Xc=100%, Yc=100%) and those are the maximum values of Xc & Yc.
The Centroid is in the middle of the object. Therefore the values of Xc= 50%, Yc =50%. The rate of black pixels Rb = 34%
Chapter 4: Methodologies
77
Therefore, it requires at least 7 bits to represent each value of Xc, Yc, and Rb in
a binary stream. In this method one byte is used to represent each value.
This binary stream representing Xc, Yc, & Rb values is the preservative data of the
area of interest in the document. Each object/character requires 24 bits of data to be
represented in that stream. This data stream is then encoded into a 2D barcode. The
barcode that carries the preservative data is then attached to the document image in
a blank white area to generate a self-validating document and only documents with
barcodes can be printed and distributed. Figure (4.2) shows the process of creating
self-validating documents.
Figure 4.2 : A Block Diagram of Creating a Self-Validating Document
4.2.2 The Verification of Self-validating Documents
The verification and creation procedures have some stages in common. In
verification, the distributed hard copy document needs to be converted to the digital
format first by using any optical scanner. After that, the scanned document image
(I‘) is converted to the binary format in the same way during the creation stage
because the verifying user may scan the document as a colour or a greyscale image.
The area of interest which might be forged is then chosen by the user. Afterwards,
Chapter 4: Methodologies
78
the selected part is scanned horizontally line by line in order to detect undesirable
minor noise such as dots caused during printing, transmission, or scanning.
A noise threshold Tn for the noise must be pre-determined and any rate of noise
lower than this threshold will be eliminated from the document. The value threshold
can be set by measuring the maximum possible number of lines in a document to
determine the average height of the line. Tn has to be significantly less than the
average height of lines. In the horizontal scan, if the height of a line is lower than Tn,
it will be considered as noise and consequently removed from the document. If there
is a single dot or a small number of added dots in the scanned document, possibly
caused by some dirt on the glass of the flatbed scanner, they will be erased by this
noise removal process. An example of a scanned image containing noise caused by
a scanner is shown in Figure (4.3a) and the same image after the noise removal
process is shown in Figure (4.3b)
Figure 4.3: A Sample Scanned Image (a) before & (b) after Noise Removal Process
Afterwards, the same line splitter applied in the creation stage is used to divide
the selected parts into separate lines. In addition, the vertical scanner used during
creation is also applied to divide each line into separate characters and to save each
character as an individual binary sub image. The centroid point (X‘c, Y‘c) of each sub-
image and the rate of black pixels (R‘b) are then computed.
Afterwards, the data stream previously encoded in the 2D barcode attached to
the scanned document is extracted using a 2D Data Matrix barcode reader.
Chapter 4: Methodologies
79
If the size of the generated stream (S‘) from the scanned document and the size
of the other stream (S) extracted from the barcode are different, the document
image is considered as a forgery. There are two main reasons for this variation in a
stream size:
1- One or more undesirable characters has been added to the document. In this
case, the size of (S‘) is larger than the size of (S).
2- Some details have been eliminated from the document. The size of (S‘) in this
case must be smaller than the size of (S).
The malicious add/remove operation could be done by a counterfeiter after the
document was printed out and distributed and before scanning it back to the
computer for verification.
Once the difference in size between (S) and (S‘) is detected, the verifier will
reject the scanned document and there is no need to take further action in the
verification process. However, a sequential byte-by-byte comparison between the
two data streams can be applied to locate the position of the added/removed data in
the scanned document.
If (S) and (S‘) are equal in size, the extracted values of Xc, Yc, & Rb of each
character/object in the data stream stored in the barcode (S) are consecutively
compared with the values X‘c, Y‘c, & R‘b calculated earlier from the scanned document
image I‘.
Two pre-determined thresholds (Tcentroid & Tblack) are needed during the
comparison to verify the authenticity of the scanned document. Tcentroid is used to
measure the difference between centroid values while Tblack is used to compare the
rates of black pixels.
Chapter 4: Methodologies
80
If (|Xc - X‘c| ≥ Tcentroid) or (|Yc - Y‘c| ≥ Tcentroid) of a particular sub image, it means
the content of this sub image in the scanned document has been tampered with and
its centroid point has been significantly shifted in a vertical, horizontal, or diagonal
way.
In addition to the centroid points comparison, the difference between each rate of
black pixels (R‘b) in S‘ & (Rb) in S needs to be computed to measure how significant
the change in the number of black dots in the sub image is. The comparison of black
rate is essential in case the centroid comparison fails to detect any changes in the
image and vice versa. The Tdiff threshold is used to decide whether the content of the
scanned sub-image has been significantly altered or not.
If |Rb - R‘b| ≥ Tdiff, it means that there are many black pixels have been added
to/removed from the scanned sub-image. As a result this part of the document will
be considered as a forgery and consequently the whole document will be rejected.
An example of a shifted centroid position and a significant change in the rate of
black pixels of a character before and after forgery is shown in Figure (4.4). The
white dot in the character represents the centroid points.
Figure 4.4 : An Example of a Character (a) before and (b) after Forgery, its Centroid, and Black Rate
(b) Forged Black Rate = 67.6% Centroid X = 49% Centroid Y = 49%
(a) Original Black Rate = 57% Centroid X = 54% Centroid Y = 49%
Chapter 4: Methodologies
81
The differences between the values of those characters in the example above are
computed. The difference in black rate = 10.6%, centroid X = 5%, and centroid Y =
0%.
There must be at least one or more significant differences exceeding the
predetermined thresholds between the values of S and S‘ in order to consider the
scanned document as a forgery. Otherwise, the content of the scanned document is
considered as authentic. The verification process is shown in Figure (4.5).
Figure 4.5 : The Verification Process of a Scanned Document
The creation and verification processes of self-validating documents using
method 1 are discussed previously in this section. The next section will discuss
another method also used to create and validate document images.
Chapter 4: Methodologies
82
4.3 Method 2
In this section, the second method to generate and authenticate self-validating
documents is presented. Different procedures are used in this method. A quad tree
technique is used to divide the parts to be protected recursively and to generate the
preservative data during both the creation and verification processes. This
preservative data is then embedded in a barcode and attached to the document
itself. A comparison between the data in the barcode and the data extracted from the
scanned image is required to verify the authenticity of the document. The creation of
self-validating documents will be explained in detail in section (4.3.1) while the
verification process is discussed in the subsequent section.
4.3.1 The Creation of Self-Validating Documents
The process of creating self validating documents consists of several steps. First
of all, the area of interest needs to be manually selected by the creator. Two
bookmarks are then added to the top-left and bottom-right corners of the selected
area. Those bookmarks are essential to specify the exact area of interest and will be
needed later during the verification process. The selected part may need to be
converted to binary format if it is initially in colour or grey scale format.
Initially, a scan is made to detect whether there is an object/character or not in
the chosen black and white image. If the image is totally white (i.e. it does not
contain any black pixels), it will be ignored and no further process is needed because
there is no point in protecting an empty image. Otherwise, if the image has at least
one object, it will pass through the rest of the creation procedures. The area of
interest also needs to be checked to see if it is totally black. It is unlikely to be totally
white or totally black but if this happened, the image would not be processed.
Chapter 4: Methodologies
83
The centroid point (Xc, Yc) of the entire binary image is calculated using the
technique discussed earlier in (section 3.4). If the centroid location is close to the
margins of the image, no further processing of the image is required to avoid
obtaining malformed images with large height and insignificant width or vice-versa.
This could happen only when the object location is close to one of the corners or the
side edges of the image. A predetermined threshold Tmargin is required to measure
how close the centroid is to the image border. Two thresholds (Tx & Ty) are derived
from Tmargin to be compared with the values of the centroid point of Xc & Yc. The
values of these thresholds are variable depending on the dimension of the image (I)
being processed. Tx can be calculated by multiplying the width of (I) with Tmargin and
dividing the result by 100 while the value of Ty is derived from the multiplication of
Tmargin with the height of the image (I) and the total must also be divided by 100. The
calculation of Tx & Ty is shown in equation (4.4) and (4.5) respectively.
Tx = Width (I) × Tmargin ÷ 100 … Equation (4.4)
Ty = Height (I) × Tmargin ÷ 100 … Equation (4.5)
An example of a (10×10) image is shown in Figure (4.6) where Tmargin is set to 10
in (a) and 20 in (b).
Figure 4.6 : An Example of a (10×10) Image where Tmargin = 10 in (a) & Tmargin = 20 in (b)
(a) (b)
Chapter 4: Methodologies
84
In Figure (4.6a), the centroid is within the accepted region (the white area)
because (Tx < Xc < Width (I)- Tx) and (Ty < Yc < Height(I)- Ty). Therefore, the image
(I) can be passed to the next procedure (the quad tree), which will be discussed later
in this section. Otherwise, the centroid point is in the grey area of the grid as in
Figure (4.6b) and no further process is applied on the image (I).
If all the conditions above are applied on the image (I), it will be divided
vertically and horizontally into 4 parts (P1, P2, P3, P4) from its centroid as a quad
tree. The quad tree partition is discussed in Chapter 3.
After that, the height and width of each divided sub-image the 4 new parts are
compared with the height and width of the original area of interest selected earlier. If
the size of either the height or the width of any part is relatively small, no further
processing is needed to this particular part. A threshold TS is used to compare the
dimensions of (P1, P2, P3, P4) with the size of the area of interest. Let P and A denote
a sub-image and the area of interest respectively. The sub-image P is undersized if
one of the two following conditions applies:
- The width of P < (The width of A × TS /100)
- The height of P < (The height of A × TS /100)
The resulting 4 parts are then processed individually in a recursive loop until one
of the conditions above does not apply for any sub-image undergoing processing.
The recursion of the document creation process is shown in Figure (4.8).
The rates of black pixels (Rb) of all divided sub-images are then computed and
represented sequentially in a data stream. Each value of the black rate is stored in
one byte. Therefore, the size of the preservative data of n sub-images equals n
bytes. This preservative data is then encoded in a 2D barcode and attached to the
image itself.
Chapter 4: Methodologies
85
If the size of the preservative data is too large to be encoded in the barcode, the
values of the black rates of each sub-image can be represented in 2-bits instead of
8-bits in order to diminish the size of the preservative data to the extent that
barcode capacity will be sufficient for holding this data.
To reduce the size of the preservative data (D), two thresholds T1 and T2 are
needed to encode the 8-bit values of D into 2 bits each. The values of D are ranging
between (00000000) and (01100100) in binary when the one byte-per-value
representation is used. When the 2-bit representation is applied, each value (v) in D‘
will be equal to one of the following binary values (00, 10, or 11). Equation (4.6) is
used to convert the 8-bit values of D into 2-bits values.
The converted values of (D‘) have 3 different values and the meaning of each
value is given below:
i. White (00): means the colour of the image is mainly white because it contains
more white than black pixels. The rate of black pixels in these images must be
between 0 and T1.
ii. Uncertain (10): means the image has a critical/delicate mixture of black and
white pixels. The rate of black pixels must be in the range T1 to T2.
iii. Black (11): means the black is dominant in the image because the majority of
the pixels are black. The rate of black pixels must be between T2 and 100.
00: if 0≤ D‘<T1
10: if T1≤ D‘ ≤T2
11: if T2< D‘ ≤100
D‘ =
…. Equation (4.6)
Chapter 4: Methodologies
86
Once the preservative data is converted from the 8-bit to 2-bit format, it will be
impossible to recover the previous format because this conversion works as a lossy
data compression method.
Assume D is a data stream contains n values (v1, v2, v3, v4, ...,vn) where each
value in this stream is in the range 0 and 100. The size of this data stream will equal
n×8 bits. This data size can be compressed to the size n×2 bits. Figure (4.7) is an
example that shows how the values in data stream D are compressed where T1=35
and T2=50 in this example. The size of the output data stream (D‘) is a quarter of
the size of the input stream (D).
Figure 4.7 : An Example of the Data Compression Technique Used to Convert the 8-bits/value into 2-bits/value data stream
The values (v1, v2, v3, v4, ...,vn) in D represent the rate of black pixels in the
images divided earlier. Therefore, a higher value means there are more black pixels
in this image.
The compressed binary stream D2 will be used as a data representation of the
area of interest previously chosen. This short data stream will be encoded in a 2D
barcode instead of the larger uncompressed stream. The barcode is then attached to
a blank area in the document to be used later in verification. Only documents with
0
T1= 35
T2=50
100
8-bit data stream D v1 = 39 (00100111) v2 = 12 (00001100) v3 = 83 (01010011) v4 = 55 (00110111)
.
.
. vn = 41 (00101001)
8-bit data stream D‘ v1 = 10 v2 = 00
v3 = 11
v4 = 11
.
.
.
vn = 10
Input Output
00
10
11
Chapter 4: Methodologies
87
attached barcodes can be distributed. The whole creation process of self-validating
documents is shown in Figure (4.8).
Figure 4.8 : A Flowchart of the Creation Process of Self-validating Documents
Attach the barcode to the document (I)
Self-Validating Document (Ì)
Hide D in a 2D barcode
Compress D
Yes
No
Yes
Yes
No
Compute Centroid C
Yes
Is the image totally white or totally black?
Document Image (I)
No Is A in binary format?
Select area of interest A
Binarize A
No
Is C close to the
margins?
Divide the image into 4 parts (Quad Tree)
Is the image too
small? Yes
Compute the rate of black pixels of the sub-images & generate the
preservative data D
Is D too
large to be saved in a barcode?
No
Chapter 4: Methodologies
88
4.3.2 The Verification of Self-validating Documents
In verification, the hardcopy self-validating document initially needs to be
scanned to the computer as a black and white image. If the image is scanned in
colour or greyscale format, it must be converted to binary format. After that, the
area of interest located between, and including, the registration marks must be
manually selected by the user by using a computer mouse.
The first process is to locate the two registration marks by scanning the top-left
and bottom-right corners of the selected sub-image. The image is then cropped to
eliminate the bookmarks and retain the precise area of interest which was selected in
the creation process.
The main reason for using registration marks is to locate the area of interest
accurately and to avoid any unintentional selection of additional pixels while selecting
the area of interest. The additional pixels would cause an increase/decrease in the
rate of black pixels of the selected part which would lead to failure in verification. An
example of an image before and after the registration mark removal is shown in
Figure (4.9).
Figure 4.9: An Example of Selected Area of Interest (a) with Registration Marks (before cropping)
(b) without Registration Marks (after cropping)
In the case where no registration marks are detected, the system will make
four scans to detect the border of the image as described below:
(a) (b)
Chapter 4: Methodologies
89
1- A top-to-bottom scan from the left side of the image to search for the first
black pixel (x1, y1).
2- A right-to-left scan from the top of the image to search for the first black
pixel (x2, y2).
3- A bottom-to-top scan from the right side of the image to search for the first
black pixel (x3, y3).
4- A left-to-right scan from the bottom of the image to search for the first
black pixel (x4, y4).
Afterwards, the system will consider the point (x2, y1) as the top-left corner
point in the image and consider the point (x4, y3) as the bottom-right corner point of
the image. Then, the image is cropped from those points to form another image
which will be taken to the next stage of process. Figure (4.10) shows an example of
how Method 2 crops an image with no registration marks.
(A)
(B)
Figure 4.10: An Example of Cropping an Image With No Registration Marks Using
Method 2 (A) Before Cropping (B) After Cropping
The same procedures, previously discussed in the creation process, are then
applied on the cropped part of the image. First, the content of the image is checked
and no further processing is required if it is totally black or totally white. Otherwise,
Chapter 4: Methodologies
90
the centroid point (Xc , Yc) of the image is computed and checked to detect if it was
close to the margin of the image using Tmargin. If not, the quad tree process is made
to divide the image into 4 parts as discussed in (section 4.2.1).
The size of each divided part is then compared against the size of the cropped
area of interest. If the width or the height of any of these parts is relatively small
comparing to the original document image, no more processing is needed to this
particular part. Other parts are then processed recursively as shown in Figure (4.11).
When the above mentioned recursive loop is done and there are no more sub-
images to split, the rate of black pixels of each sub-image is then calculated. The
consequent stream of black-rate values represents the preservative data D2 of the
scanned document. The size of the preservative data is then checked to determine
whether it is too large to be saved in a barcode or not. If so, the same data
compression used during the creation process in section (4.2.1) is used to diminish
the size of the preservative data.
The preservative data D previously embedded in the attached barcode is then
extracted. A comparison between this extracted data D and the newly generated
data stream D2 of the scanned document is then made in order to detect any
significant modifications in the scanned document. The values in D & D2 can be
represented by either 2 bits or 8 bits each.
In the 8-bit representation, each value in both D & D2 is represented in one byte.
Initially, the size of D is compared first to the size of D2. There are 2 different results
from the size comparison. D > D2 or D < D2 this means the scanned documents has
been significantly altered and the verification system will reject it and consider it as a
forgery. However, if D & D2 are equal in size, a byte-to-byte comparison between D &
D2 is required. A threshold Tdiff needs to be set to measure how significant the
change in the document is.
Chapter 4: Methodologies
91
Assume that D contains n bytes (B1, B2, B3, B4, …, Bn) and D2 contains similar
number of bytes (B‘1, B‘2, B‘3, B‘4, …, B‘n). The comparison to detect significant
modifications in the scanned document will be as follows:
The verifier can detect the number and locations of considerably modified parts
of the scanned document by using the algorithm above.
The sensitivity of the verifier to detecting forgery depends on the value of the
threshold Tdiff. Increasing the value of Tdiff makes the verifier less sensitive to
detecting modified parts. The threshold Tdiff can be adjusted to reach the optimal
degree of sensitivity which leads to an efficient verification tool.
On the other hand, the comparison in the 2-bit per sub-image representation is
different. The values in D and D2 can be one of the following: white (00), uncertain
(10), or black (11). A comparison of a pair of bits from D with another pair of bits
from D2 is needed in order to detect which part has been significantly changed. Table
(4.1) shows all the possibilities of the comparison and how the verifier decides
whether a certain part has been altered or not.
Only if a pair of bits with the value of (white 00) is switched to (black 11) or
vice-versa, the change in this part of the document image will count as a significant
alteration and, as a result, the document will be rejected. Otherwise, this part will be
accepted as an unaltered genuine image.
for i = 1, 2, 3, …, n
if |Bi - B‘i| > Tdiff : the ith part of the scanned document has been significantly altered.
otherwise: No noticeable changes are made to the ith part of the scanned document.
Chapter 4: Methodologies
92
D D2 Comparison result
White 00 White 00 Genuine
White 00 Uncertain 10 Genuine
White 00 Black 11 Forged
Uncertain 10 White 00 Genuine
Uncertain 10 Uncertain 10 Genuine
Uncertain 10 Black 11 Genuine
Black 11 White 00 Forged
Black 11 Uncertain 10 Genuine
Black 11 Black 11 Genuine
Table 4.1 : Possibilities and Results of the 2-bit Representation Comparison
The sensitivity of the verifier depends on the range of the uncertainly scale which
can be adjusted by changing the values of the T1 & T2 thresholds previously
discussed in section (4.2.1). The larger the scale is, the less sensitive the verifier is.
If the verifier is very sensitive, genuine scanned documents could be considered as a
forgery and rejected by the system.
On the other hand, the verifier will accept forged scanned documents as
authentic copies if the range of the uncertain scale is large. Therefore, the sensitivity
level of the verification system must be optimal in order to achieve a high level of
reliability in document authentication.
Chapter 4: Methodologies
93
Figure 4.11: A flowchart of the verification process of self-validating documents
No
Yes
Yes
No
No Yes
Yes
Yes
Compare the size of D & D2
Compute Centroid C
Is the image totally white or totally black?
No
Is C close to the
margins?
Divide the image into 4 parts (Quad Tree)
Is the image too
small?
Crop the image and remove the bookmarks
Hardcopy Document
Scan to computer
Yes
No Is A in binary format?
Binarize A
Select area of interest A
Extract D from the attached barcode
Compress D2
Yes
Compute the rate of black pixels of the sub-images & generate the preservative data D2
Is D2 too
large to be saved in a barcode?
Are D with D2
equal in size?
Are the contents of
D & D2 equal?
No
No
The scanned document is genuine
The scanned document is forged
Chapter 4: Methodologies
94
4.4 The Extended Version of Method 2
Method 2 can be slightly modified to obtain a different level of accuracy in
authentication. The only difference between the extended version and the method
discussed above is that the images are sliced either horizontally or vertically into 2
parts instead of 4 sub-images as done in the quad tree stage. The main reason for
this amendment is to avoid obtaining sub-images with very small dimensions. This
can be caused if the width of document, or the area of interest, is very much greater
than its height or vice versa, for instance, a single line document. Therefore, it is
better to divide this kind of document recursively in a vertical way. The verification
system can be developed to choose between method 2 or its extended version
depending on the height and width of the selected document.
The decision of whether the image should be divided in a vertical or horizontal
way can be made depending on the contents of the image. The distribution of black
pixels in the image needs to be calculated in order to choose in which way the image
should be divided. This can be done by scanning each line in the image vertically and
horizontally and calculating the average number of black pixels in each line. The
summation of the calculated averages is then divided by the number of
vertical/horizontal lines that contain at least one black pixel. If the value of
horizontal distribution is higher than the vertical value, the image is divided in a
vertical way. Otherwise, the image is sliced into 2 parts horizontally.
The horizontal distribution of the image I can be calculated by using equation
(4.7).
𝐷𝐷ℎ = 1𝑁𝑁𝑁𝑁𝑁𝑁
∑ 𝐴𝐴𝐴𝐴𝐴𝐴(𝑖𝑖)𝑁𝑁𝑖𝑖=1 … Equation (4.7)
Chapter 4: Methodologies
95
𝐴𝐴𝐴𝐴𝐴𝐴 is the average of black pixels in each row. The average of the jth row can be
computed by equation (4.8)
𝐴𝐴𝐴𝐴𝐴𝐴(𝑗𝑗) = 1𝑀𝑀∑ 𝑐𝑐𝑀𝑀𝑖𝑖=1 … Equation (4.8)
Where C = 1 if I[i,j] =0 (i.e. a black pixel)
𝑁𝑁𝑁𝑁𝑁𝑁 is the number of the rows that contains at least one black pixel in each image
and it can be computed by equation (4.9).
𝑁𝑁𝑁𝑁𝑁𝑁 = ∑ 𝑁𝑁𝑖𝑖=1 ∑ 𝑐𝑐𝑀𝑀
𝑗𝑗=1 ... Equation (4.9)
Where C = 1 if I[i,j] = 0
The vertical distribution can be calculated by using equation (4.10).
𝐷𝐷𝐴𝐴 = 1𝑁𝑁𝑁𝑁𝑁𝑁
∑ 𝐴𝐴𝐴𝐴𝐴𝐴(𝑗𝑗)𝑀𝑀𝑗𝑗=1 … Equation (4.10)
𝐴𝐴𝐴𝐴𝐴𝐴 is the average of black pixels in each column. The average of the ith column of
the image can be computed by equation (4.11)
𝐴𝐴𝐴𝐴𝐴𝐴(𝑖𝑖) = 1𝑁𝑁∑ 𝑐𝑐𝑁𝑁𝑗𝑗=1 … Equation (4.11)
Where C = 1 if I[i,j] =0
𝑁𝑁𝑁𝑁𝑁𝑁 is the number of the columns that contains at least one black pixel in each
image and it can be computed by equation (4.12).
𝑁𝑁𝑁𝑁𝑁𝑁 = ∑ 𝑀𝑀𝑗𝑗=1 ∑ 𝑐𝑐𝑁𝑁
𝑖𝑖=1 ... Equation (4.12)
Where C = 1 if I[i,j] = 0
An example of two different images containing different characters is shown in
Figure (4.12).
Chapter 4: Methodologies
96
Figure 4.12 : An Example of Two Different Sub-images and Their Vertical & Horizontal Distributions
The sub-image in Figure (4.12a) should be sliced horizontally because its vertical
distribution is higher than the horizontal distribution. The calculation of the vertical
and horizontal distribution of this sub-image is shown below:
DV = (1/8 + 1/8 + 7/8 + 1/8 + 1/8) ÷ 5
= 0.275
DH = (5/9 + 1/9 + 1/9 + 1/9 + 1/9 + 1/9 + 1/9) ÷ 7
= 0.1746
The image is divided in a horizontal way because DV > DH. The second image in
Figure (4.12b) is sliced vertically because its horizontal distribution higher than its
vertical as calculated below:
DV = (2/8 + 2/8 + 2/8 + 2/8 + 2/8 + 2/8 + 2/8 + 2/8) ÷ 8
= 0.25
DH = (4/9 + 4/9 + 4/9 + 4/9) ÷ 4
= 0.4444
0 0
0
1/8 7/8 1/8 1/8 0 0 1/8
5/9
1/9
1/9
1/9
1/9
1/9
1/9
5
7
2/8 0 2/8 2/8 2/8 2/8 2/8 2/8 2/8
0
0
4/9
4/9
4/9
4/9
0
0
8
4
(a) Vertical distribution = 0.275
Horizontal distribution = 0.1746
(b) Vertical distribution = 0.25
Horizontal distribution = 0.444
Chapter 4: Methodologies
97
4.5 A comparison between Method 1 and Method 2
The similarities and differences of Method 1 and Method 2 previously discussed
in this chapter are specified in this section.
4.5.1 Similarities of Method 1 and Method 2
The two methods have the following characteristics in common:
1- Binarization: Both methods convert the document image from the true colour
format to the binary format if the scanned image was scanned as a
monochrome image.
2- The 2D Barcode is used to carry the preservative data in both methods. This
barcode is attached to the original document to generate a validating
document. In verification, the embedded data is extracted from the attached
barcode and compared with the information of the scanned image in order to
detect any alternation.
3- Centre of Gravity: this is computed in both methods but the use mode of the
centroid is different. In Method 1, after dividing the image into characters or
smaller sub-images, the centroid is computed for each sub-image and the
location of the centroid is used as preservative data. In Method 2, the
centroid is computed for the entire protected area and the image is then
divided into parts from the centroid point. Also the centroid point values are
not used as preservative data in Method 2.
4- Rate of black pixels: Both Method 1 and Method 2 use the rate of black pixels
of sub-images as preservative data to be stored in the barcode. The stored
black rates are then compared with the black rates of the scanned document
in order to detect forgery.
Chapter 4: Methodologies
98
4.5.2 Differences of Method 1 and Method 2
Method 1 and Method 2 are also different in the following respects:
1- Method 1 scans a document horizontally and divides it into separate lines.
Then, it scans each line vertically to divide the lines into characters.
However, Method 2 does not use the same division procedure.
2- In Method 2, the sub images are not divided if the centroid is close to a
margin, if the size of the sub-image is relatively small compared to the
original document, or if the image is totally black or white. None of those
conditions apply in Method 1.
3- Method 2 uses The Quad Tree algorithm to chop the image into 4 parts. This
is not used in Method 1.
4- The preservative data in Method 1 is the location of the centroid and the rate
of black pixels of each divided character. The data stored in Method 2
consists of only the black rate of the sub-images.
5- Method 1 employs a noise removal technique when it scans documents
horizontally to detect lines. Method 2 does not detect any added noise.
4.6 Limitations of using Method 1
Method 1 scans the lines separately from top-to-bottom to find spaces between
characters in order to divide those lines into individual characters. Therefore, it fails
to divide characters in some texts such as:
a) Strikethrough texts: because the line drawn in the middle of the text
connects all the letters of that text as shown in Figure (4.13). The method
will take the whole strikethrough text to the next process without dividing it
into characters.
Chapter 4: Methodologies
99
Figure 4.13: A Sample of a Strikethrough Text
b) Underlined texts written in special fonts such CENA, Kozuka, Microsoft Yi
Baiti, Segoe Print, ZWAdobeF and many other font styles with connected
characters: Therefore, Method1 will take the whole underlined text line as a
single sub-image and will not be able to divide the underlined text into
individual characters. Figure (4.14) shows samples of underlined text written
in some of the abovementioned fonts. However, it is unusual to use one of
those fonts to write an official document.
Figure 4.14 : A Sample of a Strikethrough Text
c) Texts written in languages that have words with connected characters such
as Arabic, Urdu, and Farsi as shown in Figure (4.15): Method 1 will process
each word individually instead of splitting those words into characters.
Chapter 4: Methodologies
100
Figure 4.15: Texts Written in (a) Arabic, (b) Urdu, and (c) Farsi languages
d) Texts written in special fonts in English language also have connected
characters like those in the previously mentioned languages. Script fonts
such as Palace Script MT, Brush Script, Edwardian Script MT, French Script,
and Segoe script MT are examples of fonts with connected letters.
Consequently, each word or group of connected characters will be taken
together as a sub-image without dividing them into separate characters.
Figure (4.16) shows some samples of English script fonts with connected
characters.
Figure 4.16: Samples of Texts Written in English Script Fonts
e) Text written in italic style is not always divided into separate characters by
Method 1 if no single white line between characters is detected. Therefore, the
Chapter 4: Methodologies
101
sub-images may contain a combination of characters. Figure (4.17) shows an
example of a text line written in italic style and how it is divided using the
vertical scan for a white line of Method 1.
Figure 4.17: A Sample of Text Image Written in Italic Style and How it is Divided into Sub-images Using Method 1
f) In addition to the above-mentioned types of texts, Method1 considers a group of
any text lines as a single line if those lines are written in a text box or in a table
with solid lines. It takes the whole text box or table to the next process without
splitting the contents into characters. Also, if there was a vertical line drawn on
the side of the text lines, those lines will be taken as a single line in the line
separation stage because the method is based on finding at least a single totally
white row between text lines in order to split them. Figure (4.18) shows samples
of texts in a lined table and a text neighbouring a vertical line.
Figure 4.18: A Sample of Text (a) Neighbouring a Vertical Line (b) in a Table With Solid Lines
As a result of the failure of Method 1, Method 2 is preferred for verification of
documents written with those kinds of texts because it does not process individual
characters but deals with the whole document image.
(a) (b)
Chapter 4: Methodologies
102
4.7 Modifications to Method 1
Method 1 can be slightly modified to be able to deal with documents with
connected and other non-English texts and to avoid the problem of character linkage
caused by low resolution scanners. The impact of the print-scan operation on
separate characters will be discussed in Chapter 5. To improve the verifier, each line
should be divided vertically into a number of sub-images instead of separate
characters. The number of sub-images per line can be decided by equation (4.13).
No. of parts⁄line = 𝑁𝑁𝑁𝑁𝑅𝑅𝑅𝑅𝑅𝑅( 𝑊𝑊𝑖𝑖𝑅𝑅𝑊𝑊 ℎ (𝑙𝑙𝑖𝑖𝑅𝑅𝑙𝑙 )𝐻𝐻𝑙𝑙𝑖𝑖𝐻𝐻ℎ𝑊𝑊 (𝑙𝑙𝑖𝑖𝑅𝑅𝑙𝑙 )
) ........ (4.13)
Using equation (4.13) converts the method from character-based to content-
based and divides each line into sub-images larger than character images. Also, it
reduces the size of the preservative data because the number of resulting sub-
images is certainly smaller than the number of characters.
The experimental results for use of both Method 1 and Method 2, their reliability
in detecting forgery and validating scanned documents, the statistical error in each
method, the analysis of the results and a comparison between the results of both
methods are given in chapter 5.
Chapter 5: Experimental Results and Analysis
103
Chapter Five: Experimental Results and Analysis
5.1 Overview
The methods proposed in chapter 4 have been implemented by using Borland
Delphi Enterprise Version 6.0. The verification system of each method was tested to
measure the efficiency and the reliability of those methods in verifying unaltered
scanned documents and detecting forgery in modified images. In experiments, 55
test document images with four different typefaces (Verdana, Times New Roman,
Georgia, and Arial) and several font sizes have been printed out using an HP LaserJet
8150 printer. Copies of all printed documents have been forged and rescanned to the
computer. Unaltered copies have also been scanned as monochrome bitmap images
using an HP ScanJet 2200c scanner with 300dpi (Dots per Inches) resolution. The
verification systems of Method 1 and Method 2 are then applied on the scanned
documents to check whether they were maliciously modified or not. The
experimental results, the analysis, and a comparison between the efficiency of the
methods are all shown in this chapter.
5.2 The Experimental Results of Method 1
As Method 1 divides text images into lines and divides those lines afterwards into
character objects, it is essential to analyse the effect of print-scan operation on
characters. The character image analysis is discussed in the next section and the
experimental results of Method 1 are given in the subsequent section.
Chapter 5: Experimental Results and Analysis
104
5.2.1 Character Image Analysis
Method 1 divides text images into character images. The rate of black pixels and
the centroid location of each character are then computed and used as preservative
data. In experiments, character objects in the printed document are subject to
alternations. Additional black pixels around the edges of characters may appear due
to the low-resolution scanners. The extra pixels added to the characters may link two
characters or more with each other. Also, the added noise may fill the white line gap
between two characters. Examples of scanned characters with additional noise are
shown in Figure (5.1).
Therefore, the method will consider those connected characters as one object
which causes inconsistencies during the verification process and, as a result, the
document will be rejected by the verifier.
Figure 5.1: Samples of Connected Characters Due to the Additional Noise of Print-Scan Operation
Consequently, it is important to measure the additional noise caused by scanners
and to analyze the amount of noise in each character separately. The impact of print-
scan operation on all letters (A..Z, a..z) in the English language, digits (0..9), and 9
Chapter 5: Experimental Results and Analysis
105
other symbols has been examined. The characters have been written in 7 different
typefaces and 9 different font sizes. The typefaces used in experiments are Calibri,
Arial, Times New Roman (TNR), Verdana, Georgia, Impact, and Latha. While the
sizes of fonts chosen for the test characters were 8, 10, 12, 14, 16, 18, 20, 22, and
24. Samples of the above listed fonts are shown in Figure (5.2).
Figure 5.2: Samples of Calibri, Arial, Latha, Verdana, Times New Roman, Implact and
Georgia fonts
The main reason behind choosing these typefaces and sizes is that they are
commonly used styles and sizes in official documents. Also, it is more reasonable to
select a font size between 8 and 24 in documents like these rather than other extra
small⁄large font sizes.
Seventy one characters includes letters (A..Z), (a..z), numbers (0..9), and some
special characters have been printed and scanned back into the computer. The
influence of the digital-to-analogue operation on characters has been assessed by
making a comparison between the centroid locations and the rates of black pixels of
the original and the scanned characters. If R is the rate of black pixels in a character
image and (X,Y) is the centroid point of this image, the difference between original
and scanned character image can be computed by:
R = | R(original character) – R(scanned character)|
X = | X(original character) – X(scanned character)|
Y = | Y(original character) – Y(scanned character)|
Chapter 5: Experimental Results and Analysis
106
Each character has been affected to different levels depending on its font size
and style. Table (5.1) shows the average difference between the original and
scanned characters written in the 7 different font styles.
Font Style
Character
Calibri Arial Times
Verdana Georgia Impact Latha
X Y R X Y R X Y R X Y R X Y R X Y R X Y R Avr.
A 1.6 0.9 5.3 0.5 0.6 2.8 0.9 2.1 4.6 0.5 0.2 3.3 0.5 1.2 4 1.1 0.3 4.1 1.2 0.7 4 1.92
B 1.3 0.7 6 1.6 0.5 2.8 2.4 0.2 2.2 1.5 0.6 1.5 1.4 0.7 2.1 0.6 0.5 2.7 1.5 0.6 3.4 1.66
C 1.1 0.6 5 0.4 0.6 4.1 0.4 0.5 3.2 0.6 0.9 3.6 0.8 1 4 0.9 1 3.4 0.9 0.9 4.7 1.83
D 1.3 0.3 5.1 1.9 0.6 2.2 1.9 0.8 2.2 1.4 0.8 2.5 1.2 0.5 1.8 0.3 0.1 2.4 1.4 0.9 3.2 1.56
E 1.8 1.8 5.8 1 1 1.5 0.6 1.2 0.6 0.9 0.4 1.3 0.3 0.3 1.2 1.1 0.5 1 1.2 0.9 2.1 1.26
F 1.6 1 6.1 1 0.6 1 0.6 0.4 1 1.1 0.7 0.7 1.2 0.9 1 1.1 0.4 1.5 0.6 0.4 1.9 1.19
G 1 1.3 6.6 0.6 0.9 3.7 2 1 2.7 1.2 0.8 3.7 1.3 0.7 3.7 0.9 0.8 2.4 1.4 0.9 5.3 2.04
H 0.9 1 6.2 1 0.5 1 1 1.3 1.2 1 0.7 0.9 0.8 0.5 0.8 0.7 0.7 1.9 1 0.5 1.3 1.19
I 6.7 0.7 17 3.9 1.1 5.4 1.8 0.9 1.4 2.1 0.9 1.2 1.8 1.3 1.9 2.4 0.4 5 3.5 0.9 8 3.25
J 2.2 1.2 6.1 1.9 1.1 2.4 1.8 2.1 3 1.9 1.3 1.8 1.4 1.4 3 1.2 0.4 1.8 2.4 2.5 2 2.04
K 1.1 0.8 6.2 1.5 0.6 3.3 2 0.9 3.5 1.3 0.8 3.1 2 1.2 4.2 0.7 0.3 2.5 1.7 0.6 4.4 2.03
L 2.1 1.8 4.4 1.5 0.6 0.7 0.7 0.7 1.2 1 0.4 1.3 1 1.5 1.5 2 0.7 1.7 1.4 0.7 0.7 1.31
M 1 0.6 5.8 0.2 0.8 3.9 0.3 1 1.9 0.7 1.3 2 0.3 0.7 2.8 0.7 0.6 1.4 1 0.7 5.6 1.59
N 2.8 0.2 6.9 1.2 0.6 2 1 1.3 0.9 1.2 0.5 2.1 0.8 0.8 1.8 0.9 0.1 2.4 1.2 0.4 4 1.58
O 0.7 0.9 5.2 0.9 0.8 4.3 0.6 0.7 4.4 0.6 0.7 4.5 0.9 0.7 4.5 1 0.4 2.7 0.5 0.6 6 1.98
P 1.1 1 6.8 1.4 1.1 2.4 1.2 1.3 3.3 1.2 1.3 2.3 1.5 0.6 2.4 0.6 0.9 1.9 1.2 0.8 2.7 1.76
Q 0.4 1.1 3.4 0.9 0.8 3.3 0.2 0.2 4.2 0.4 0.4 3.9 0.4 0.7 3.9 1.1 0.4 3 0.7 1.4 5.2 1.71
R 1.3 0.7 7.2 2.4 0.8 2.7 2.7 1.1 3.5 1.4 0.3 3.2 1.9 0.7 3.1 0.5 1.1 2.4 1.2 0.7 2.1 1.96
S 1 0.7 6.7 0.6 0.4 5.4 1.4 1 5.5 0.7 0.7 5.5 0.8 0.7 6.2 1.3 0.2 4.2 1.1 0.7 6.9 2.47
T 1.1 1.8 3.4 0.7 1.5 0.7 1.1 1.4 0.4 0.3 1.2 1.2 0.5 1 0.7 0.4 0.4 1.4 0.5 1.2 1.3 1.06
U 2.5 1.1 4.5 0.5 1.7 1.5 1.1 1.4 1.7 1.4 1.2 2 0.5 1.8 1.2 0.9 0.1 1.3 1.3 1.3 2.9 1.52
V 0.8 1 5.8 0.8 0.7 4.9 1.2 0.9 3.9 0.7 0.9 4.1 1.1 0.9 4.1 0.5 0.7 3.5 0.3 0.7 4.2 1.99
W 1 1 5.8 0.9 0.9 5.1 0.4 0.7 4.5 0.7 0.4 5.3 0.5 0.1 5 0.5 0.4 3.8 0.7 0.7 7 2.16
X 1.3 0.7 5.6 0.7 0.9 5.9 0.7 0.7 4.4 0.4 0.5 3.8 0.4 1.1 4.3 0.7 0.9 3.8 0.5 0.7 6.4 2.11
Y 0.9 2 4.2 0.5 2.7 1.8 0.4 2.9 1.9 0.8 2.2 1.5 0.4 1 2.9 1.2 1.1 2.4 1.2 2.9 2.4 1.78
Z 1 1.5 5.6 0.8 1.1 2.8 1 1.7 3.6 0.6 0.8 2.7 1.2 1.2 4.3 1.1 0.4 4.2 0.7 0.6 2.4 1.87
a 1.8 1 7.1 2.8 0.9 4.9 2.4 0.8 5.2 1.6 1.2 5.4 2.1 1.2 6.1 1.2 0.7 3.5 1.9 1.2 7.2 2.87
b 1.9 1 6.8 1.3 0.6 4.1 1.4 0.4 3.6 1.9 0.3 3.4 1 0.7 2.2 0.5 0.4 2.3 1.3 0.6 4.1 1.90
c 1 1.1 6.4 0.9 1 5.6 0.8 1.2 6.7 0.8 0.8 5.2 1 0.9 5.7 0.6 0.8 4.1 1.2 0.8 6.8 2.54
d 1.4 0.7 7 2.2 0.7 4.1 3.1 0.7 3.2 1.6 0.4 4 2.7 0.4 3.1 1 0.3 3 2.9 0.2 3.9 2.22
e 1.7 1.1 8.1 1 0.3 5.8 0.9 0.6 5.4 1.2 0.9 4.9 1 0.9 5 1.2 0.7 3.9 0.7 0.3 7.1 2.51
f 1.4 1.1 6.9 2.6 1.4 2 0.8 3.4 2 0.6 1.8 1.3 0.8 1.1 2.3 2.2 0.9 1.7 1.1 1.6 2.4 1.88
g 1.2 0.7 9.2 2.1 0.6 5.4 0.8 0.3 5.2 1.3 0.6 5 0.8 0.8 5.9 1.6 0.3 2.4 1.9 0.3 5.9 2.49
h 2.2 1 4.7 1.6 1.2 2.2 2.7 1.4 2.7 0.7 0.6 1.4 1.2 1.1 2.2 0.7 1.3 1.8 1.4 1.1 2.2 1.69
i 7.1 1.8 14 3.3 0.2 5.9 5.2 3.2 8 3.4 0.2 6.7 2 2.3 4.6 2.1 0.2 2.3 5 0.8 7.1 4.07
Chapter 5: Experimental Results and Analysis
107
Font Style
Character
Calibri Arial Times
Verdana Georgia Impact Latha
X Y R X Y R X Y R X Y R X Y R X Y R X Y R Avr.
j 3.9 0.7 6.1 2.3 1.6 3 1.7 0.4 3.8 1.9 1.1 1.9 0.9 1 2.8 2.3 0.2 3.8 2.8 1.6 3.7 2.26
k 1.3 0.6 5.1 1.7 0.2 3.4 1.8 0.9 3.2 1.6 0.2 2.6 1.3 0.8 3.9 0.7 0.1 3.2 1.5 0.7 5.4 1.91
l 5.1 0.4 15 4.9 0.8 4 6.7 1.6 9 3.3 0.8 6 2.8 0.4 3 1.6 0.6 4 4.2 0.6 4 3.75
m 1.9 2.2 5.7 0.6 2.3 1.4 1.3 3.1 3.6 0.9 2.2 2 1.2 2.2 3.4 0.3 1.1 1.4 0.7 2.1 4.9 2.12
n 2.1 1.6 6.9 1.6 2.4 2.3 2.2 3.4 5.3 1.2 1.7 2.2 1.4 1.9 3.8 0.8 1 2.4 1.4 2.8 2.4 2.42
o 1.2 1.3 6.4 0.9 0.6 5.1 0.9 1 6.6 0.4 0.7 5.2 0.7 0.6 6.7 1 0.7 4.3 1 0.8 7.8 2.57
p 1.9 1.1 7.4 1.2 1.1 3.8 0.7 1.6 3.7 1.6 1.8 3.3 1.4 0.8 2.8 0.6 0.7 1.2 1.3 1.3 4 2.06
q 1.6 1.3 6.3 2.7 1.2 2.6 2.7 1.9 3.7 2.4 1.1 3.1 1.9 1.2 3.3 1.1 0.8 2.2 2.3 1.3 5.2 2.38
r 1.7 2.3 5.9 0.6 2 2.2 0.9 2.2 2.2 0.8 1.7 2 0.6 2.4 2.6 0.3 0.9 2.3 0.9 2.3 3.4 1.91
s 1.3 1.3 8.7 0.6 0.7 7.1 1 1 7.4 0.2 0.4 6.6 0.7 0.6 7.3 1.6 0.4 5.9 0.7 0.3 8 2.94
t 1.8 1.2 6 2.1 1.2 1.8 2.3 0.9 1.7 0.9 0.9 1.8 1.7 1.1 1.8 1.1 0.4 2.8 1.2 0.4 2 1.67
u 1.7 1.2 5 1.3 1 1.4 1.2 1.3 1.6 1.3 1 1.8 1.6 0.6 2.2 1.1 0.7 2 1.7 1.4 2.9 1.62
v 1.1 0.7 8 1 1.3 5.7 1 0.7 4.4 1.1 1.2 5.6 0.8 0.8 5 0.8 1 4 1.1 0.9 5.8 2.48
w 1.3 1 6.4 0.6 1.1 4.7 0.7 0.8 4.9 0.8 0.8 6.3 0.9 0.6 5.9 0.9 0.7 3.7 1 0.6 7.3 2.43
x 1.4 0.6 7.4 1.4 1.3 5.9 1.3 1.6 6.7 0.9 0.6 4 0.9 1.2 5.6 1.1 0.7 4.4 1.1 1 8.3 2.73
y 1 1.1 4.8 0.8 0.4 4.6 1.3 1.4 4 0.8 0.7 3.6 0.6 1.1 4.7 1.8 0.8 3.8 1.4 1.3 5.9 2.19
z 1.8 2.1 7.8 0.9 0.9 2.8 1.9 1.8 4.7 0.7 1.4 3.8 1.2 1.3 5.1 0.2 0.2 3 0.9 1.7 3.9 2.29
0 1 1 3.7 0.6 0.9 2 0.7 0.6 1.2 0.7 0.8 0.3 0.9 1 1.2 1 0.6 1.7 1 0.7 4.6 1.25
1 25 9.1 4.6 1.3 1.1 1.7 9.4 2 8.5 1.1 1.4 0.8 3.3 1 2.7 1.4 0.6 1.3 1.3 0.1 3.2 3.85
2 2.2 1.4 2.8 1.2 1.3 1 1.6 1.9 1.4 1 1.3 1.1 0.4 1.6 1.1 1.1 0.9 1.7 1.1 2.4 2.4 1.47
3 1.8 0.8 2.9 1 0.6 1 1.2 1 0.9 0.4 0.6 0.8 0.9 0.6 1.1 1.1 0.6 2.6 1.4 0.9 2.7 1.19
4 2.3 2.4 1.9 1 1 1 1 1.4 1.7 1 0.8 0.9 1.1 0.9 1.4 1.4 0.6 1.6 1.6 0.8 1.2 1.23
5 1.3 1.1 2.9 1.6 0.6 1.3 1.1 1 0.9 0.9 0.6 1 1.2 0.8 0.7 1.2 0.1 0.8 2.1 1.1 3.6 1.23
6 2.9 2.2 2.4 0.9 0.8 2.4 0.9 0.7 1.9 1 0.6 0.7 0.8 0.8 0.9 1.7 0.4 1.4 1.9 0.2 4.4 1.42
7 3.8 2.4 2.8 1 0.7 0.3 0.9 0.8 1.2 0.6 0.4 0.6 0.9 0.6 0.7 0.9 0.4 2.7 5.6 1.3 2.3 1.47
8 1.3 0.6 3.9 0.9 0.7 2.8 0.7 0.7 2.4 0.4 0.9 1.2 0.9 0.7 1 1.1 0.8 1.2 0.8 0.7 3.6 1.30
9 1.1 3.4 2.4 1.1 0.7 2.1 1 0.3 1.1 0.6 0.6 1.1 1 1.1 2 1.1 0.4 1.8 0.8 1.3 4.1 1.39
$ 1 0.6 5 1 0.9 5.1 1.1 0.2 4.2 0.9 0.6 3.1 0.8 0.6 3.3 1 0.3 3.4 1 0.6 6.1 1.94
£ 1 2.3 5.1 0.8 0.4 4 0.8 0.6 4.8 0.7 2.2 3.1 1 2.6 2.1 1 1 3.3 0.7 0.6 5.4 2.07
; 3.4 1.8 5.1 3.8 2.3 5.2 3.1 2.6 6.1 2.4 2.8 1.4 1.3 1.1 6.3 2 0.3 4 1.7 1.8 5.6 3.05
. 3.3 5.8 12 5.9 5.7 9.2 6.6 6.4 25 3.6 2.3 4.2 2.6 2.1 16 2.9 2.4 2.6 3.2 5.9 8.1 6.47
, 3.6 1.4 11 5.9 3.1 9.3 2.9 1.9 12 3.2 0.7 6.6 2.4 1.4 8.3 1.6 1.2 7.4 3.1 1.4 7.9 4.59
+ 0.9 1.6 3.3 0.9 1 1.3 0.7 0.8 1.1 0.6 0.9 0.8 0.2 0.6 1.3 0.8 1.1 1.2 0.6 0.8 1.7 1.06
- 2.3 6.4 19 1.3 4.9 5 1.1 4.1 2.9 1 4.7 6.1 1.1 6.6 5.3 1.3 3.1 3.1 1.7 3.9 6.6 4.36
* 1.6 1.1 9.9 0.8 1.3 8 0.9 0.6 12 1.2 1.1 5.2 1.1 1.1 12 2 1.7 6.9 1.8 1.4 13 4.03
% 1 1.4 7.1 0.3 0.4 5.7 0.8 0.6 5.2 0.4 0.8 5.4 0.6 0.6 6.3 0.8 0.4 4.7 0.6 0.7 5.9 2.37
Average 2.1 1.5 6.4 1.5 1.1 3.5 1.6 1.3 4 1.1 1 3 1.1 1.1 3.6 1.1 0.7 2.8 1.5 1.1 4.6
Table 5.1 : The Influence of Print-Scan Operation on the Centroids and the Rates of Black Pixels of 71 Characters Written in 7 Different Font Styles
Chapter 5: Experimental Results and Analysis
108
The average effect on each character is shown in the right hand column of the
table while the average effect on font style is given at the bottom of the table. The
table above shows that the average influence on characters differs from font style to
another. The values X and Y represent the rates of shifting in characters’ centroids
occurring in the x-axis and y-axis coordinates respectively. The value R represents
the percentage difference in the black rates of each character. Table (5.2) shows the
average influence of the print-scan operation on all characters centroids and rates of
black pixels written in font styles.
Font Style Impact Verdana Georgia Arial TNR Latha Calibri
Average effect 1.53 1.7 1.93 2.03 2.3 2.4 3.33
Table 5.2 : The Average Influence of Print-Scan Operation on Characters Written in 7 Different Font Styles Sorted from the Least to The Most Affected Fonts
The characters written in Calibri font are affected more than characters written
in the other fonts while those in Impact font are the least influenced characters.
Noticeably, the print-scan operation has more influence on characters written in
thinner fonts than those with thick strokes. The difference rates of each value (X, Y,
and R) and the average difference of the three values are shown in Figure (5.3).
Figure 5.3: The Average Influence of Print-Scan Operation on the Centroids and Rates of Black Pixels of Characters Written in 7 Different Font Styles
0
1
2
3
4
5
6
7
X
Y
R
Avr.
Chapter 5: Experimental Results and Analysis
109
The rates of black pixels or the centroid positions of character images with small
width or small height can change more significantly than those with wide body.
Characters with the maximum and the minimum alternations in their X, Y, and R and
the values of this modification are shown in Table (5.3).
X Y R
Maximum alteration (1) 6.11% (-) 4.81% (.) 11%
Minimum alternation (Q) 0.59% (d) 0.49% (T) 1.3% Table 5.3 : The Maximum and Minimum Change in the X, Y, and R Values and their Most and
Least Influenced Characters
If the maximum alternation values are used as thresholds to compare characters
during the verification process, the scanned documents will be verified. However,
other modified characters in forged documents will also be verified if their alteration
values are less than the maximum alteration values listed in Table (5.3). Therefore,
the verification system will fail to detect a forgery which leads to a Type-I error (i.e.
False-positive). The average influence on each of the 71 characters has been
computed. Table (5.4) shows the 10 most affected characters and the average
alteration values.
Character . , - i * 1 l I ; s Average alternation 6.47 4.59 4.36 4.07 4.03 3.85 3.75 3.25 3.05 2.94
Table 5.4 : The 10 Most Affected Characters by Print-Scan Operation and the Average Difference in Their Centroid Positions and the Rate of Black Pixels
The reason why those characters are affected more than other characters is that
if the centroid point of a character with diminutive width is compared to its height,
for instance the characters (l, I, i), or a character with diminutive height compared to
its width such as (-) have been shifted to the neighbouring pixel in the coordinate of
the smaller side of the character, it causes a significant change in the location of this
Chapter 5: Experimental Results and Analysis
110
character when compared with its original location. The value X of the letter (l, I, i,
1) is considerably altered while the value Y is slightly affected because the width in
those characters is small. In other characters such as (-), the value Y is altered more
than the value X because of its small height. The shift in the centroid location can be
caused by one or more extra pixels being added to the side of the character.
Also, the rate of black pixels in characters with diagonal strokes and edges such
as (*) or characters with curves such as (s) can be significantly influenced by
scanners. These types of characters can gain more additional black pixels than other
characters with straight vertical and horizontal lines such as (F, E, T, and H). The 10
least influenced characters and the average difference are listed in Table (5.5).
Character 8 E 0 4 5 F H 3 T + Frequency % 1.3 1.26 1.25 1.23 1.23 1.19 1.19 1.19 1.06 1.06
Table 5.5 : The 10 Least Affected Characters by Print-Scan Operation and the Average Difference in their Centroid Positions and the Rate of Black Pixels
However, the use of high-resolution scanners can reduce the potential noise in
the scanned documents.
The significant variation between original and scanned characters increases the
probability of a false-positive error during the verification process if the threshold
used for comparison has a low value. On the other hand, if a high-value threshold is
used to compare characters, modified characters will be considered as genuine and
the forged document will be accepted by the verifier. As a result, the probability of a
false-negative error will increase. The average difference between the examined 71
characters in Table (5.1) equals 2.175. If this value was used as a threshold to
compare original and scanned characters, the 10 most influenced characters will be
rejected by the verifier.
Chapter 5: Experimental Results and Analysis
111
However, it is important to know how frequently used those characters are in
documents in order to measure the rate of error. Table (5.6) shows the frequency of
the 10 most influenced characters per 100 characters in a sample English text book
containing 9,867,698 characters.
Character . , - i * 1 l I ; S Frequency % 1.29 1.26 0.22 5.76 0.001 0.005 3.84 0.31 0.106 5.38
Table 5.6 : The Frequency Percentage of the 10 Most Affected Characters by Print-Scan Operation
The letters (i, s, and l) are the most common characters among the 10 most
influenced characters, therefore, they have more impact on the false-positive error
rates than other characters when they are altered by scanner.
It is crucial to identify what characters are of interest to counterfeiters and how
often the verification system fails to detect forgery in each character. In particular
we need to estimate the actual false-negative error but unfortunately there is no
practical estimations for the counterfeiters’ interest.
Each character is affected in different levels by digital-to-analogue conversion
when it is written at different font sizes even if the same font style is used. The
impact of the print-scan operation in the 71 characters written in 9 different font
sizes was measured by comparing the rate of black pixels (R) and the centroid (X, Y)
of the original and the scanned character images. The difference rates between the
71 tested characters are shown in Table (5.7).
Chapter 5: Experimental Results and Analysis
112
Size
Char
24 22 20 18 16 14 12 10 8
X Y R X Y R X Y R X Y R X Y R X Y R X Y R X Y R X Y R
A 0.7 0.3 2 0.6 0.6 2.8 0.6 0.4 2.6 0.6 0.8 3.1 1.1 0.8 3.2 0.6 0.8 4.6 0.8 0.7 4.8 1.1 1.7 5.7 2.3 1.5 7.4
B 1 0.3 2 1.1 0.1 3.1 0.7 0.7 2.3 0.7 0.4 2.3 0.8 0.4 2.7 1.2 0.4 3.1 2.1 1 3 2.2 1.1 3.1 3.6 0.8 5.1
C 0.6 0.4 2.4 0.4 0.7 3 0.8 0.4 3 0.7 0.6 3 0.4 0.8 3.2 0.7 0.7 4 0.6 1 4.2 1.5 1 5.4 0.8 1.6 6.8
D 0.6 0.6 2 0.7 0.6 3.3 0.6 0.6 2.1 1.1 0.1 2.8 1.4 0.4 1.8 1.3 0.7 2.6 1.8 0.6 2.1 2 0.6 3.3 2.7 1.1 5
E 0.3 0.8 2 1 0.7 2.3 1.1 0.7 1.4 1.1 0.7 1.2 0.8 0.8 1.4 0.6 0.8 1.1 0.7 1 2.4 1.8 1.1 3.3 1.6 1.1 2.4
F 1.1 0.3 0.9 0.8 0.1 2.3 1.1 0.8 1.1 0.7 0.4 1.8 0.7 0.4 1.6 1.2 0.6 1.4 1.2 1 2.3 1.4 0.7 2.3 1 1.5 3.6
G 0.4 0.3 2.7 1 1.1 2.8 0.8 0.6 3.1 1.3 0.7 3 1.1 0.8 3.8 1.4 0.7 4.1 1.6 1.4 5 1.4 0.8 5.4 1.7 1.6 6.2
H 0.6 0.6 1.3 0.7 0.7 2.6 0.3 0.7 0.8 0.6 0.6 2.1 1 0.7 1.6 0.7 0.7 2 1.1 0.7 1.7 2 1.1 1.7 1.1 1 3.3
I 1.4 0.3 3.4 2.8 0.7 5.4 1.5 0.8 3.3 3 0.8 3.7 2.4 0.3 5.1 3.4 0.8 5.9 4.6 1 5.6 4.6 1.3 8.7 4.7 2 11
J 1.1 1 2.1 0.6 1 1.4 1.4 0.9 1.7 1.4 1.7 2.6 1.7 1.1 2.6 2.1 1.1 3.4 2.7 1.5 4.1 2.4 2.3 3.7 3 2.4 4.1
K 0.3 0.3 1.7 1.1 0.4 3.1 1.1 0.4 2.8 1.3 0.4 3 1 0.4 2.9 1.8 0.8 4.6 2.7 0.7 4.3 1.5 1.3 6 2.3 1.7 6.9
L 1.1 0.6 1 1.3 0.6 1.4 1 0.6 1 1.3 0.7 1.4 1.8 0.6 1.1 0.7 0.3 0.7 1.3 1.1 1.4 1.7 1.4 3.1 2.3 2.3 3.6
M 0.7 0.6 1.7 0.7 0.3 2.8 0.4 0.6 1.6 0.4 0.7 3.3 0.5 0.8 2.4 0.3 0.6 3.2 1 1 5 0.4 1.4 3.8 1 1.1 6
N 0.7 0.1 1.6 2 0.6 2 1.4 0.7 2.4 0.7 0.3 2.3 0.7 0.4 2.4 1.1 0.4 2.7 1 0.8 2.1 1.8 0.8 5.1 2.1 0.8 5.1
O 0.4 0.3 2.6 0.6 0.7 3.6 0.8 0.7 3.3 0.7 0.6 4.1 1 1.1 4.4 0.7 0.4 5 0.6 0.7 5.4 0.7 0.7 6.4 1 0.8 6
P 0.7 0.4 1.6 0.7 1 2.6 0.7 0.8 2.3 1.4 1 2.3 1.3 1.1 3.3 1.1 0.6 3.8 1.8 1.4 2.8 1.3 1.1 4.6 1.7 1.6 4.8
Q 0.4 0.4 2.1 0.1 0.3 3 0.6 0.7 3.3 0.6 0.3 3.7 0.6 0.6 4 0.7 0.6 3.6 1 1.1 4 0.7 1.1 5.1 0.7 1.4 5.9
R 1 0.7 1.7 1 0.6 2.7 1.2 0.6 2.8 1.4 0.1 3.1 1 0.7 3.6 1.6 0.4 3.1 2.4 1.3 3.4 2.1 0.7 5.3 3 1.7 5.6
S 0.6 0.4 3.6 1.1 0.2 4.1 0.7 0.4 4.4 0.8 0.4 4.6 0.7 0.7 5.8 1 0.6 6 1 0.7 6.5 1.1 1.1 7.2 1.8 0.8 9.6
T 0.8 0.8 1.3 0.4 1.1 0.3 0.4 1 0.7 0.7 1.6 0.9 0.6 1 0.7 0.4 1.6 1.9 0.6 1.6 2.3 1.1 1.4 2.3 1 1.1 1.7
U 0.7 0.6 1.7 0.6 0.7 2.3 1.1 0.6 1.7 1.1 1.1 2.6 1 1 2.3 0.7 1.6 2.3 1.4 1.4 2 1.1 2.4 1.4 3 1.7 3.3
V 0.7 0.4 2.4 0.4 0.6 2.8 0.7 0.7 3.6 0.7 1 3.7 0.8 0.7 4 0 1 4.6 1.3 1 5.6 1.4 1 5.7 0.8 0.8 6.9
W 0.3 0.6 2.7 0.6 0.6 4.1 0.3 0.4 4.3 0.9 0.7 3.6 0.6 0.6 4.4 0.7 0.6 5.1 0.9 0.4 5.7 1.1 0.9 7.7 0.9 0.7 9.3
X 0.8 0.6 3.3 0.3 0.6 3.1 0.7 0.7 4.1 0.4 0.7 3.9 0.3 0.6 4.1 0.9 0.4 4.4 0.4 1 5 1.3 0.9 6.4 1.3 1.6 9.4
Y 0.7 0.9 2.3 0.9 1.2 1.7 0.7 1.6 1.6 0.4 1.8 1.9 0.9 2.1 1.4 1 2.1 1.6 0.6 2.6 3.4 0.4 3.1 3.9 1.6 3.6 4.4
Z 0.3 0.7 3 0.7 1.3 3.1 0.6 0.7 2.3 0.6 0.6 3.1 0.9 0.9 3.1 1.1 1 4 1.1 1.4 3.7 2 1.9 5.3 0.9 1 5.1
a 0.9 0.4 4 1.4 0.9 4.1 1 0.9 5 1.6 1 4.7 2.1 0.9 5.6 1.9 1 6.3 3 1.2 6.3 2.4 1 7.7 3.4 1.7 7.1
b 1.1 0.4 2.7 0.7 0.4 2 1.4 0.1 2.7 0.9 0.6 2.3 1.4 0.4 4.3 1 0.7 3.7 1.9 1 5 1.9 0.6 5 1.9 0.9 6.4
c 0.4 0.4 3.4 0.4 0.7 2.9 0.6 0.6 4.6 0.4 1 5.9 0.7 1.2 5.6 1.1 1.2 6.4 2 0.7 6.7 1.1 1.3 7.7 1.1 1.1 7.9
d 1.3 0.4 3 1.9 0.7 2.3 1.7 0 3.4 1.7 0.4 3.9 1.9 0.6 3.9 1.7 0.6 5.1 2.9 0.4 4.6 3.1 0.7 5.1 3.1 0.6 5.1
e 0.7 0.7 4.1 0.6 0.4 5 0.9 0.1 4.1 0.9 0.7 4.4 1 0.6 5 1.2 0.7 6.4 1.7 1.1 7.4 1.6 1 6.4 1.3 0.7 8.7
f 0.4 1 2.9 1.1 1.7 1.4 0.6 1.6 2.1 0.9 1.7 2.7 0.7 1.4 2.7 1.3 1.1 2.6 2.1 1.4 3.7 2 2.3 2.6 3 2.3 3.2
g 0.7 0.4 3.7 1.4 0.4 4.4 1.3 0.7 3.7 1 0.4 4.6 1.1 0.4 5.3 1.4 0.1 6.1 1.9 0.4 6.4 1.7 0.6 7 1.9 1 9
h 1.1 0.6 2.1 1.3 0.7 1.1 1 0.9 1.6 1.1 1 1.6 1.8 1.2 3.1 1.4 1.1 3.4 2.1 1.1 2.1 2 1.2 4.4 1.7 2 2.6
i 4 0.7 6 2.3 1.3 4.6 3.4 1.3 6.9 2.3 1.6 6.1 4.3 1 8.1 4.7 1.4 6.4 3.7 1.1 11 5.6 1.1 8.1 6 1.7 4.7
j 1.7 0.6 3.1 1.9 0.7 3.1 1.7 0.9 3.3 1.4 1.1 3.6 2.7 0.4 3.4 2.6 0.8 3.9 2.7 0.8 2.7 3.1 0.8 4.1 2.4 2.1 4.9
k 0.9 0.4 2.7 0.9 0.1 1.9 1.1 0.3 3 1.1 0.6 2.6 1.7 0.9 4.1 1.7 0.6 5.7 1.4 0.3 4.4 2.5 0.6 6.1 1.2 0.7 4
l 4.1 0.6 6.7 3.1 0.6 4.7 3.6 0.6 4.9 3.3 0.9 6.3 4.4 1.1 6.3 4.4 0.9 11 4 0.3 4.2 3.4 0.7 8.1 6.2 1 4.5
m 1 1.4 2.7 0.6 1.7 2.1 0.1 2.1 1.7 0.7 1.9 2.9 1.1 1.6 2.7 0.9 2.6 2.6 2 2.3 3.4 1.2 2.4 5.3 1.1 3.7 5.4
n 1.1 1.4 2.9 1.3 1.9 2.6 1.6 1.7 3.4 1.1 2 3.9 1.4 1.9 4.1 2.3 1.7 4.7 1.1 3.3 3.6 2.1 2.3 4.3 1.7 2.9 3.2
Chapter 5: Experimental Results and Analysis
113
Size
Char
24 22 20 18 16 14 12 10 8
X Y R X Y R X Y R X Y R X Y R X Y R X Y R X Y R X Y R
o 0.4 0.4 4 0.7 0.6 5 1 0.4 4.4 0.6 0.1 5.6 0.7 1 5.6 0.9 1.1 6.6 0.9 1 7 1.3 1.3 7.6 1.4 1.1 8.4
p 0.7 0.9 3.6 0.7 0.6 3 1.2 1.3 3.9 0.7 1.3 2.1 0.9 0.9 3.6 1 0.9 3.9 1.7 1.1 4.1 2.3 1.9 4.7 1.9 2 4.9
q 1.6 0.9 3.1 1.6 1 2 1.9 0.9 2.6 1.7 1 3.6 2 1.6 3.6 2 1 4.3 3.1 1.3 3.9 2.4 1.9 5 2.6 2 6
r 0.4 1 3 0.6 1.6 2.3 0.7 1.6 2.1 0.4 1.7 2.4 0.3 2 3.1 1.1 2.1 4 1.4 2.9 2.3 1 2.1 3.9 1.2 2.9 3.4
s 0.4 0.3 4.4 1.1 0.1 5.1 1 0.4 4.7 0.6 0.6 6.1 0.4 0.6 7 0.6 0.9 8.4 0.9 1.1 8.7 1 0.9 9.2 1.7 1.3 11
t 1.1 0.6 1.9 1 0.7 0.9 1.4 0.9 1.6 1.6 0.9 1.7 2 0.7 2.6 1 0.7 2.6 1.7 0.6 3.4 1.3 1.7 3.1 3.1 1.3 5.1
u 0.9 0.7 1.7 1.6 0.6 2.1 1 0.9 2.1 1.4 1 2.6 1.3 1.3 1.8 1.6 1.6 3.7 1.7 0.7 1.9 1.3 1.4 2.3 2 1.1 3.4
v 0.7 0.3 3.9 0.3 0.4 3.4 1 1 4.4 0.9 0.9 4.9 1 0.9 5.1 1.1 0.6 6.3 1.3 1.7 6.6 1 1.3 6.8 1.6 1.4 8.1
w 1 0.6 4.3 0.7 0.3 4.6 0.3 0.7 4.7 0.7 0.6 4.7 0.9 0.4 5 1.1 0.6 5 1.1 1.3 6.9 1.1 1.3 8.1 0.9 1.3 7.1
x 1 0.9 4.6 1.4 0.6 4.4 0.7 0.4 5.3 0.7 0.4 5.4 1.3 1.4 5.6 1.4 1.3 5.9 0.9 1.1 6.6 1.4 1.7 7.4 1.7 1 9.3
y 0.6 0.3 3.4 0.9 0.9 3.1 0.7 0.9 3.3 0.4 0.7 3.7 0.7 0.6 4.1 1.6 1.3 4.4 1.4 1.6 4.9 1.6 1.3 5.9 2 1.4 7.3
z 0.7 1.6 4 0.6 1.7 3.9 0.7 0.4 2.3 0.7 0.7 3.9 1.1 0.9 3.6 0.7 1 3.3 1.9 1.4 4.8 1.3 2.1 5.9 2 2.3 7.7
0 0.7 0.4 2.1 0.4 0.6 0.9 0.3 0.9 1.6 0.9 0.7 1.7 0.9 0.9 2.4 1.3 0.9 1 0.9 0.6 2 1 0.9 3.6 1.1 1.3 3.6
1 5.4 1.6 2.4 5.7 1.6 3.1 5.7 2 2.1 4.6 1.9 3.1 4.9 2.1 3 6.9 2.3 3.9 6.4 2.7 4 8.6 2.3 2.7 6.9 3.3 5
2 0.4 0.7 0.6 0.7 0.9 0.9 1 1.7 0.9 0.9 1.4 0.4 1.4 1.4 1.9 1.4 2 1.6 1.7 1.6 2.9 1.7 2.3 2.1 1.9 2 3.7
3 1.1 0.4 1.3 0.7 0.3 0.7 1.1 0.7 1.3 0.6 0.9 1.6 1.3 1.1 2.1 1.4 0.7 1 1.3 0.6 2.7 1.7 0.7 1.4 0.9 0.9 3.1
4 1.3 0.6 1 0.9 0.7 0.7 1.1 0.6 1.1 0.9 1.1 1.1 1.4 1 1.1 1.3 1.7 1.4 2 1.3 1.1 1.4 1.7 0.9 1.9 1.4 3.9
5 0.7 0.3 0.7 1.1 0.4 1.3 1 0.7 1.4 1.3 0.9 1.1 1 0.4 1.9 1.4 0.6 1.3 2.1 0.8 2 1.6 0.1 2.3 1.9 1.1 2.3
6 0.9 0.6 1.4 1 0.6 0.9 1 0.9 1.6 1.3 0.7 1.1 1.4 0.7 2 1.4 0.7 1.4 2.1 1 3.4 2.6 1 2.4 1.1 1.1 4
7 1.7 0.9 1.6 1.3 0.7 0.7 2.4 0.9 1.3 1.6 0.9 1.3 2.4 1 1.4 1.3 0.7 1 1.9 1.1 2.1 2.4 0.9 2.4 2.6 1.6 1.7
8 0.6 0.1 1.7 0.6 0.3 1.3 0.4 0.9 1.6 0.9 0.3 1.6 1.1 0.6 2.9 0.7 0.4 1.7 1 0.9 2.4 1.6 1 3.1 1 1.9 4.4
9 0.9 0.9 1 0.7 0.7 1.6 0.7 0.9 1.7 1 0.9 2 0.7 1 2.4 0.7 1.1 1.7 1.2 1.7 2.3 1 1.4 2.9 1.6 1.6 3.2
$ 0.4 0.1 2.6 0.7 0.6 3.4 0.7 0.3 2.6 0.7 0.6 4 2.1 0.6 4.1 1 0.3 4.1 1.1 0.9 4.9 1.1 0.7 5.3 0.7 0.7 8
£ 0.4 0.9 2.6 0.9 1.1 3.4 0.6 1.6 2.6 0.7 0.9 3.6 0.9 1.1 3.7 1 1.6 3.7 0.6 1.7 4.9 1 1.4 5.4 1.6 2.3 6
; 0.3 1 2.1 2.3 1.1 4.3 2.4 1.1 3 2.9 1.6 4.4 2.4 1.6 5.4 1.7 1.3 5 2.3 2.9 5.6 3.7 2.3 6 3.9 3.4 7.6
. 2.1 1.3 7.3 3.1 2.7 6.3 2.1 2.6 6 3.1 3.9 10 2.6 5 9.9 5 4.6 12 3.4 4.3 12 6.6 4.6 11 7.9 11 24
, 2.9 1 6.9 1.3 1.1 6.1 1.7 1.1 7.7 2.1 1.7 8.3 3.1 0.9 7.9 2.6 1.9 9.9 4.6 2.7 11 4.1 2.1 12 6.7 1.9 11
+ 0.6 0.3 1.4 0.6 0.7 1.3 0.3 0.7 2.1 0.7 0.7 0.7 0.9 0.9 0.7 0.6 0.6 1.9 0.6 1 2.1 0.7 1.7 1.7 1 2 1.9
- 1 2.3 3 1.1 2.1 3.3 1 4 6.9 1.1 5 11 1.3 4 4.1 1.1 3.4 5.4 1.4 5.7 11 1.7 6.5 5.1 2.9 10 12
* 0.7 1.3 6.1 0.7 1.1 7.3 1.1 0.9 6.7 0.9 1 8.1 1.4 0.6 9.7 1.7 0.7 9.9 1.9 1.6 10 2 1.6 14 1.6 2 14
% 0.4 0.4 4.3 0.4 0.4 4.7 0.7 0.9 4 0.4 0.6 5.1 0.4 0.9 4.9 0.4 0.3 6.2 1.1 1 6.6 0.7 1 7.3 1 0.9 8.7
Avr. 1 0.7 2.8 1.1 0.8 2.9 1.1 0.9 3 1.1 1 3.5 1.3 1 3.7 1.4 1 4.2 1.8 1.3 4.6 1.9 1.4 5.1 2.2 1.8 6.1
Table 5.7 : The Influence of Print-Scan Operation on the Centroids and the Rates of Black Pixels of 71 Characters Written in 9 Different Font Sizes
Chapter 5: Experimental Results and Analysis
114
The average influence on the values R, X, and Y of the characters is shown in
Table (5.8) and Figure (5.4).
24 22 20 18 16 14 12 10 8
X 1 1.1 1.1 1.1 1.3 1.4 1.8 1.9 2.2
Y 0.7 0.8 0.9 1 1 1 1.3 1.4 1.8
R 2.8 2.9 3 3.5 3.7 4.3 4.6 5.1 6.1
Avr. 1.5 1.6 1.7 1.9 2 2.2 2.6 2.8 3.4
Table 5.8 : The Average Influence of Print-Scan Operation on the Centroids and Rates of Black Pixels of Characters Written in 9 Different Font Styles
Figure 5.4: The Average Influence of Print-Scan Operation on the Centroids and Rates of Black Pixels of Characters Written in 9 Different Font Styles
It is noticeable that characters with smaller font size are more affected by D⁄A
conversion than characters with larger font size. The difference rates are increasing
gradually when the font size is getting smaller. The print-scan conversion causes
additional noise around the inner and outer borders of characters in the scanned
image. The added or removed black pixels can significantly increase or decrease the
rate of black pixels and shift the centroid point (if the noise is not distributed evenly
0
1
2
3
4
5
6
7
24 22 20 18 16 14 12 10 8
X
Y
R
Avr.
Size of font
Influ
ence
rat
e
Chapter 5: Experimental Results and Analysis
115
to the character image) in smaller font characters more than other characters with
larger fonts.
Therefore, the amount of noise makes a considerable difference in small font-
size characters while it makes less when added to larger size character images. This
explains the reason why large bold fonts are more often used in real life applications
of document verification as shown in the sample RYANAIR boarding pass in
Appendix-D where the font of the protected information such as the flying date and
the passenger name in this e-ticket is written in font sizes 12 and 25 and Verdana
Bold font.
Due to this diversity in the centroid position and rates of black pixels of different
fonts, Method 1 has to be seen as content dependent. Therefore, different thresholds
must be pre-set for different document images depending on the font size and style
of the textual image. The determined thresholds have to be a part of the
preservative data attached to the document and used during verification. The
experimental results of the verification system of Method 1 are given in the next
section.
5.2.2 Results of the Verification System of Method 1
Characters in document images can become connected to each other after a
print-scan operation. Therefore, it is better to divide each line into a number of parts
as previously suggested in section (4.7) to avoid the high rate of false-positive errors
type caused by character connections with a resulting rejection of authentic
documents having connected characters. Two sample document images divided by
using Method 1 are shown in Appendix A.
Chapter 5: Experimental Results and Analysis
116
The 55 test documents have been printed and scanned without intentional
modifications. Also, each copy of the same 55 documents has been forged twice by
different people and scanned to the computer. The 55 unaltered and 110 forged
document images were then used to test the verification system of Method 1. Table
(5.9) show the results of the verification system of Method 1 on the above-
mentioned document images.
Image # Scanned Forged1 Forged2 Image # Scanned Forged1 Forged2 1 Rejected Rejected Rejected 29 Rejected Rejected Rejected 2 Verified Rejected Rejected 30 Verified Rejected Rejected 3 Verified Rejected Rejected 31 Rejected Rejected Rejected 4 Verified Rejected Rejected 32 Rejected Rejected Rejected 5 Rejected Rejected Rejected 33 Rejected Rejected Rejected 6 Verified Rejected Rejected 34 Rejected Rejected Rejected 7 Rejected Rejected Rejected 35 Rejected Rejected Rejected 8 Rejected Rejected Rejected 36 Rejected Rejected Rejected 9 Rejected Rejected Rejected 37 Rejected Rejected Rejected
10 Verified Rejected Rejected 38 Rejected Rejected Rejected 11 Rejected Rejected Rejected 39 Rejected Rejected Rejected 12 Verified Rejected Rejected 40 Rejected Rejected Rejected 13 Rejected Rejected Rejected 41 Rejected Rejected Rejected 14 Rejected Rejected Rejected 42 Verified Rejected Rejected 15 Rejected Rejected Rejected 43 Verified Rejected Rejected 16 Verified Rejected Rejected 44 Verified Rejected Rejected 17 Rejected Rejected Rejected 45 Verified Rejected Rejected 18 Verified Rejected Rejected 46 Verified Rejected Rejected 19 Verified Rejected Rejected 47 Verified Rejected Rejected 20 Rejected Rejected Rejected 48 Verified Rejected Rejected 21 Rejected Rejected Rejected 49 Verified Rejected Rejected 22 Rejected Rejected Rejected 50 Verified Rejected Rejected 23 Verified Rejected Rejected 51 Verified Rejected Rejected 24 Rejected Rejected Accepted 52 Verified Rejected Rejected 25 Rejected Rejected Rejected 53 Verified Rejected Rejected 26 Verified Rejected Rejected 54 Verified Rejected Rejected 27 Verified Rejected Rejected 55 Verified Rejected Rejected 28 Rejected Rejected Rejected
Table 5.9: The Experimental Results of Method 1 Verificatoin System on 55 Unaltered and 100 Forged Scanned Documents
The verification system frequently failed to verify genuine scanned documents
because it was sensitive to the print-scan operation. Twenty eight out of 55 genuine
scanned documents have been rejected by the system. The failure of the verifier is
due to the fact that scanned images are subject to rotation, additional noise, and
Chapter 5: Experimental Results and Analysis
117
skew. Those applied modifications can mislead the verifier in detecting lines in the
scanned images. The verifier can detect a wrong width size of each line in the
scanned image if there is any level of rotation.
Consequently, the lines will be divided in a different way from how they were
divided during the creation process. The number of generated sub-images from lines
will be not equal to the number stored in the attached preservative data. As a result,
the false positive error rate will be extremely high. In real life applications, innocent
clients can lose trust in any organization that uses a document verification system
which rejects their genuine documents.
On the other hand, the verifier can easily detect small a forgery in intentionally
altered documents even if the forged images are perfectly scanned without rotation
and noise by comparing the rates of black pixels and the centroid points of the
generated sub-images with those stored in the preservative data. Manipulating the
thresholds Tcentroid and Tblack can adjust the sensitivity of the verifier. The smaller
Tcentroid and Tblack, the more sensitive the verifier is. In the experiments, both
thresholds were set to 7. Forgery has been detected in 109 out of 110 altered
documents.
Only one document with a single forged character among 441 characters was
considered as a genuine copy during the verification process. The reason why the
verifier failed to detect forgery is that the number of added black pixels was
accidently equivalent to the removed black pixels which did not change the rate of
black pixels in the image. Also, the new distribution of black pixels in the forged
image did not make a significant change in the location of the centroid point.
Therefore, the modified image was accepted by the system.
Chapter 5: Experimental Results and Analysis
118
The exact locations of the altered parts were detected in Method 1 but there was
no possibility to recover the original contents. located in Table (5.10) shows the
statistical error of the verification system of Method 1.
Document Condition
Scanned (unaltered) Forged
Test
Results
Verified True Positive 49.1% False Negative 0.9 %
Rejected False Positive 50.9% True Negative 99.1%
Table 5.10: The Statistical Error of the Verification System of Method 1
5.3 The Experimental Results of Method 2 The same 55 scanned and 110 forged document images previously used in
Method 1 were also used to test Method 2. The whole contents of the document
images were selected as counterfeiters’ areas of interest to be protected. The text
images were recursively divided into 4 parts from the centroid point in Method 2
using the quad tree method. Two sample document images divided by Method 2 are
shown in Appendix B.
Two thresholds (Tmargin and TS) are required to stop the recursive division process
and another threshold (Tdiff) is needed to compare between the rate of black pixels of
the divided sub-images of the test scanned document and the original image as
explained in section (4.3.1) and (4.3.2). Different threshold values have been set for
document images depending on the contents of the document. Table (5.11) shows
the values used for each test image and the experimental results of the verification
system in Method 2 on the 55 scanned and 110 forged document images.
Chapter 5: Experimental Results and Analysis
119
Thresholds Matching Rates
Img# TS Tmargin Tdiff Scanned Forged 1 Forged 2 No. of Parts 1 20 20 2 100 98.4 96.9 64 2 25 20 3 100 40.9 27.3 22 3 20 20 1 100 17.2 94.8 58 4 20 20 1 100 96.6 89.7 58 5 20 20 6 100 76.1 98.5 67 6 25 20 0.5 100 78.6 71.4 28 7 20 20 0.5 100 45.7 65.2 46 8 20 20 2 100 94.2 67.3 52 9 20 20 3 100 25 42.2 64 10 20 20 3 100 92.2 21.9 64 11 30 20 0.5 100 68.8 93.8 16 12 25 20 1 100 96.8 74.2 31 13 20 20 2 100 60.9 95.3 64 14 20 20 2 100 91 88.1 67 15 30 20 1 100 87.5 0 16 16 30 20 0.5 100 68.8 0 16 17 20 20 0.5 100 70.3 62.5 64 18 20 20 1 100 89.1 90.6 64 19 30 20 2 100 93.8 6.25 16 20 20 20 2 100 95.3 81.3 64 21 20 20 1 100 93.8 95.3 64 22 20 20 1 100 98.4 90.6 64 23 20 20 0.5 100 53.1 67.2 64 24 20 20 1 100 95.3 95.3 64 25 20 20 1 100 89.1 82.8 64 26 20 20 1 100 90.6 93.8 64 27 30 20 0.5 81.25 37.5 68.8 16 28 20 20 1 100 59.4 96.9 64 29 20 20 0.5 100 84.2 63.2 19 30 20 20 2 100 94.7 100 19 31 30 20 2 100 37.5 75 16 32 30 20 1 100 75 93.8 16 33 10 20 3 100 97.2 99.6 250 34 10 20 4 100 94.5 96.9 256 35 20 20 4 100 79.7 56.3 64 36 20 20 1 100 87.5 81.3 64 37 20 20 6 100 96.9 98.4 64 38 20 20 3 100 79.7 98.4 64 39 20 20 5 100 92.2 79.7 64 40 20 20 4 100 96.9 48.4 64 41 20 20 1 100 93.8 90.6 64 42 10 20 3 100 98 98.4 250 43 20 20 1 100 92.2 98.4 64 44 20 20 1 100 70.3 92.2 64 45 20 20 2 100 69 91.4 58 46 20 20 0.5 100 65 55 40 47 20 20 0.5 100 72.5 70 40 48 30 20 0.5 100 37.5 31.3 16 49 20 20 1 100 88.2 94.1 34 50 20 20 2 100 57.7 63.5 52 51 20 20 0.5 100 35.9 85.9 64 52 20 20 0.5 100 64.5 83.9 31 53 20 20 0.5 100 50 53.1 64 54 20 20 0.5 100 53.8 32.7 52 55 25 20 2 100 75 100 52
Table 5.11 : The Experimental Results of the Verification System of Method 2 on Scanned and Forged Images
Chapter 5: Experimental Results and Analysis
120
The value of the Tmargin threshold is static in the experiments for all the test
document images but different values have been set for the other two thresholds TS
and Tdiff in order to minimize the error rate and obtain optimal results in verification.
The number of sub-images resulted from each document is shown in the right
side column of the table above. The table has drawn attention to the fact that the
smaller value of TS threshold leads to more sub-images derived from the document
image as in images 33, 34, and 42. On the other hand, the number of sub-images is
smaller when TS value is higher as in images 11, 15, 16, 19, 31, 32, and 48.
Among the 55 test scanned document images used in experiments, the verifier
succeeded in authenticating 98.2% of the documents after print-scan operation while
1.8% of the unaltered scanned documents were rejected by the verifier. The only
one unauthenticated out of 55 documents represents the false-positive (Type-α)
error rate in the system.
The verification system of Method 2 has successfully detected alterations in 108
out of 110 test forged images, the true-negative rate of counterfeit detection is
98.2%. On the other hand, the verifier failed to detect forgery in two forged text
images out of 110 images. This failure caused a 1.8% false-negative (i.e. Type-β)
error rate. The statistical error in the verification system of Method 2 is shown in
Table (5.12).
Document Condition
Scanned (unaltered) Forged
Test
Results
Verified True Positive 98.2% False Negative 1.8%
Rejected False Positive 1.8% True Negative 98.2%
Table 5.12: The Statistical Error of the Test Results of Method 2
Chapter 5: Experimental Results and Analysis
121
It is essential to mention that the two forged images that have been verified
and the unaltered image which has been rejected by the verifer have more white
areas than the other images which have been successfully authenticated or rejected.
The reason behind this failure is that some sub-images may have a small part of
a character located on the border while the rest of the sub-image is totally white. No
significant change occurs in the rate of black pixels of this sub-image when the
centroid in the parent image of this sub-image is shifted by a single bit. Also, any
added or removed black pixels to the text of this sub-image would not be able to
make a noticeble difference in the rate of the black pixels over the whole sub-image
which leads to a false-negative error. On this basis, it may be inferred that images
with extensive white parts can cause failure in the verification system. Therefore, it
is preferable to select areas of interest containing texts with few white areas.
5.4 The Experimental Results of the Extended Version of Method 2
The same test document images were also used to test the reliability and
efficiency of the modified version of Method 2 which recursively divides each image
into 2 parts either vertically or horizontally from its centroid point as discussed
earlier in section (4.4). An example of the division process of this method applied on
two sample images is shown in Appendix C. The experimental results of the
verification system of this method on the test documents are shown in Table (5.13).
Unlike in Method 2 where value of the value of Tmargin threshold was static, different
values for the three thresholds TS, Tmargin, and Tdiff were set to test each document
depending on the contents of the document.
Chapter 5: Experimental Results and Analysis
122
Thresholds Matching Rates
Img# TS Tmargin Tdiff Scanned Forged 1 Forged 2 No. of Parts 1 0.2 0.2 0.5 100 50 100 8 2 0.25 0.2 2 100 20 20 10 3 0.25 0.25 2 100 9.1 63.6 11 4 0.2 0.2 1 100 95.2 95.2 21 5 0.25 0.2 1 100 0 30 10 6 0.25 0.25 0.5 100 81.2 62.5 16 7 0.25 0.25 0.5 91. 7 75 75 12 8 0.2 0.2 1 100 Forged Forged 22 9 0.2 0.2 3 100 33.3 46.7 30 10 0.2 0.2 4 Rejected Forged Forged 28 11 0.2 0.2 5 89.3 64.3 78.6 28 12 0.2 0.2 1 100 64.3 89.3 28 13 0.2 0.2 1 100 75 100 28 14 0.2 0.2 1 100 81.25 62.5 16 15 0.25 0.2 1 100 66.7 0 9 16 0.2 0.2 1 100 100 0 28 17 0.2 0.2 1 100 63.3 86. 7 30 18 0.2 0.2 1 100 65 55 20 19 0.2 0.2 0.5 100 13.8 65.5 29 20 0.2 0.2 1 100 96.4 85.7 28 21 0.2 0.2 1 100 100 86. 7 30 22 0.2 0.2 1 100 96.4 92.9 28 23 0.2 0.2 1 100 67.9 100 28 24 0.2 0.2 0.5 96.9 62.5 46.9 32 25 0.2 0.2 1 100 96.2 30.8 26 26 0.2 0.2 1 100 92 68 25 27 0.2 0.2 1 100 100 88 25 28 0.2 0.2 1 100 90 73.3 30 29 0.2 0.2 1 100 81.8 72.7 11 30 0.2 0.2 1 100 Forged Forged 13 31 0.2 0.2 1 100 75 75 12 32 0.2 0.2 1 100 53.6 92.9 28 33 0.2 0.2 1 100 97.1 76.5 34 34 0.2 0.2 1 100 59.4 81.25 32 35 0.2 0.2 1 22.6 32.3 16.1 31 36 0.2 0.2 1 100 93.1 82.8 29 37 0.2 0.2 1 100 3.6 Forged 28 38 0.2 0.2 1 100 71.4 71.4 28 39 0.2 0.2 1 100 73.3 73.3 30 40 0.2 0.2 1 100 23.3 0 30 41 0.2 0.2 1 100 100 96.7 30 42 0.2 0.2 1 100 50 67.9 28 43 0.2 0.2 1 100 96.8 Forged 31 44 0.2 0.2 1 100 95.8 79.2 24 45 0.2 0.2 1 100 96.4 100 28 46 0.2 0.2 0.5 100 68.4 63.2 19 47 0.2 0.2 0.5 90 65 80 20 48 0.2 0.2 2 100 62.1 Forged 29 49 0.2 0.2 1 100 73.1 100 26 50 0.2 0.2 1 100 73.1 73.1 26 51 0.2 0.2 0.5 Rejected Forged Forged 18 52 0.2 0.2 1 100 53.8 92.3 13 53 0.2 0.2 1 100 57.1 71.4 28 54 0.2 0.2 1 100 65 55 20 55 0.2 0.2 1 100 12.5 12.5 8
Table 5.13: The experimental Results of the Verification System of the Extended Version of Method 2 on Scanned and Forged Images
Chapter 5: Experimental Results and Analysis
123
The verification system of this method shows higher rates of error than the
verification system of Method 2 because it failed to detect forgery in 9 out of 110
altered documents. Also, it rejected 7 out of 55 unaltered scanned documents. The
errors values during verification are written in bold font in Table (5.13). The words
Rejected and Forged in the table means that the number of the sub-images
generated during the verification process are not equal to the number of the sub-
images derived during the creation process. Therefore, the scanned image has been
considered as a forgery. The statistical error in the verification system of the
extended version of Method 2 is shown in Table (5.14).
Document Condition
Scanned (unaltered) Forged
Test
Results
Verified True Positive 87.3% False Negative 8.2%
Rejected False Positive 12.7% True Negative 91.8%
Table 5.14: The Statistical Error in the Verification System of the Extended Version of Method 2
5.5 A Comparison between the Results of the Proposed Methods
As the same test document images were used to test the verification system of
all the proposed methods in this thesis, it is essential to compare the results to
identify the most efficient and reliable algorithm in detecting significant changes and
verifying scanned document images. Table (5.15) shows the average rates of the
successfully authenticated rejected document images by the verification system of
the three proposed methods after the print-scan operation.
Chapter 5: Experimental Results and Analysis
124
Method 1 Method 2 The Extended version of Method 2
55 Unaltered scanned document images 49.1% 98.2% 87.3%
110 Forged scanned document images 99.1% 98.2% 91.8%
Average 74.1% 98.2% 89.55%
Table 5.15: The Average Rates of Successfully Verified and Rejected Documents by the Three Proposed Methods After Print-Scan Operation
The table above shows that Method 2 is more accurate and reliable for detecting
alterations in forged document as well as for verifying genuine scanned images as
this method has less sensitivity to the print scan operation than the other two
methods. Method 1 is the most sensitive and therefore it failed to verify more than
half of the unaltered scanned documents. On the other hand, it has successfully
rejected the vast majority of the forged document due its sensitivity. Overall, it is
better to rely on Method 2 as it shows the highest average rate of accuracy in
verification.
5.6 The Influence of Rotation on the Verification Process
A sample document has been rotated by 0.75 degrees clockwise and counter-
clockwise with 0.05 intervals and each rotated image has been compared with the
original document. The matching rates of the rotated and original images in the
verification system of Method 2 are given in Figure (5.5) and Table (5.16).
Chapter 5: Experimental Results and Analysis
125
Figure 5.5: The Rotation Influence on Document Verification
Clockwise (CW) Counter Clockwise (CCW)
Rotation Degree
Matching Rate
Rotation Degree
Matching Rate
0.05 100% -0.05 100%
0.1 100% -0.1 100%
0.15 99.09% -0.15 100%
0.2 99.09% -0.2 99.09%
0.25 99.09% -0.25 99.09%
0.3 97.24% -0.3 98.16%
0.35 98.16% -0.35 98.16%
0.4 95.41% -0.4 97.24%
0.45 96.33% -0.45 98.16%
0.5 96.33% -0.5 96.33%
0.55 95.45% -0.55 95.45%
0.6 95.45% -0.6 96.33%
0.65 96.33% -0.65 96.33%
0.7 93.39% -0.7 96.33%
0.75 93.39% -0.75 83.48%
Table 5.16: The Rotation Influence on Document Verification
5.7 Invisible Data Hiding The preservative data of a document can be invisibly embedded into the
document image itself by using any of the invisible data-hiding methods for binary
75.00%
80.00%
85.00%
90.00%
95.00%
100.00%
-0.7
5-0
.7-0
.65
-0.6
-0.5
5-0
.5-0
.45
-0.4
-0.3
5-0
.3-0
.25
-0.2
-0.1
5-0
.1-0
.05 0
0.05 0.1
0.15 0.2
0.25 0.3
0.35 0.4
0.45 0.5
0.55 0.6
0.65 0.7
0.75 Rotation Degree
Matching Rate
Chapter 5: Experimental Results and Analysis
126
images provided that they do not visually affect the binary images to the extent of
leading to different preservative data. The dots-shift data-hiding technique can be
implemented to embed a binary stream of data into a bi-level document image. The
main idea of the dots-shifting technique is to shift the dots of the letters (i & j) in the
document image very slightly (up, down, left, right, or diagonally) to encode 3 bits
as a maximum amount of data per dot as shown in Figure (5.6).
Figure 5.6 : An Example of Dot Shifting Technique to Embed 3 Bits of Data by Shifting the Dot in the Letter (i)
Theoretically, the proposed technique does not cause noticeable irregularities on
the host document image. To extract the hidden data, the locations of shifted dots
must be determined in order to detect the horizontal direction of shifting (left or
right). Also, the distance between the centroid of the dot and the character which
this dot belongs to must be calculated to find out the vertical shifting direction (up or
down).
The Data Hiding Capacity (DHC) depends on the quantity of the available dotted
letters (i & j) in documents written in English language. In a sample English text
book contains 2,188,153 words in 4466 pages (i.e. 489.98 words⁄page), the number
of the letters (i and j) has been computed. The percentages of frequencies of (i & j)
letters per 1000 words and per page have been calculated as shown in Table (5.17).
Chapter 5: Experimental Results and Analysis
127
No. of words & pages
No. of (i & j) Words = 2,188,153 Pages = 4,466
(i) = 569,016 260 (i ⁄ 1000 words) 127.4 (i ⁄ page)
(j) = 9572 4.3 (j ⁄ 1000 words) 2.143 (j ⁄ page)
Table 5.17 : The Percentages of Frequencies of (i & j) Letters per 1000 Words and Per Page in a Sample (2,188,153 words) Text Book
The data hiding capacity (bits ⁄page) for a text can be calculated by equation
(6.1) below:
Data Hiding Capacity (bits ⁄page) = 3 × [(No. of i) + (No. of j)] ... equation (6.1)
The average DHC (bits ⁄page) for that sample text book is:
DHC = 3 × [127.4 + 2.143]
= 388 bits (that is 48 ASCII chars)
The capacity differs from language to another because texts in some languages
have more dots than other languages, for instance there are more dotted letters in
the Arabic language than those in English. On the other hand, some languages like
Chinese or Hebrew do not have dotted letters which makes this technique not
applicable for those languages. Therefore, this data hiding techniques would be
useful only for some languages. More experiments are needed to prove this
hypothesis.
Chapter 6: Conclusions and Future Research
128
Chapter Six: Conclusions and Future Research
6.1 Recommendations to Improve the System
The methods proposed in this research can be developed by adding more
features or by replacing some algorithms with others. This chapter sheds light on
techniques that if they were applied, they could possibly enhance the performance,
reliability, and security level of the verification system. The main aim of making
these changes would be to make the verification system more reliable in detecting
forgery and in authenticating original copies. Also, the system could be improved by
employing an invisible data-hiding method to embed the preservative data. Various
suggestions and recommendations for developing the system are given in this
section. One or more of the techniques listed below could be added to the document
creation and validation algorithms in order to enhance the verification method.
6.1.1 Noise Removal
Noise can occur in documents when they are printed, sent, and rescanned into
the computer because of factors such as inconsistencies of printers, unclean flatbed
glass of scanners, humidity, or the exposure of hard copy to direct sunlight for long
time. Also, the careless handling of documents by their owners can be another cause
of noise in documents, for example if documents become slightly torn or bent. If the
amount of the noise is significant, the verifier would count this noise as additional
unwanted elements in the scanned document thereby leading to false detection of
tampering and the possiblity that the document will be considered as a forgery and
rejected by the system.
Chapter 6: Conclusions and Future Research
129
Therefore, it is essential to apply a noise removal method to binary images that
recognize the difference between the original text of the document and the added
noise followed by elimination of this undesirable noise (Agrawal & Doermann, 2009).
The challenge in the noise removal is that the real text of the imaged document
should not be damaged, no matter what the size of the text. In this research, Method
1 used a simple noise removal technique to eliminate isolated dots or lines from the
document image. However, a more accurate and efficient noise reduction technique
is recommended for use in the future.
6.1.2 De-skewing
Documents can be rotated clockwise or anticlockwise during printing and
scanning operations. In printers, if the paper tray inaccurately feeds the printer, the
text lines will be irregularly oblique on the printed paper. On the other hand, if the
document is not well placed in the flatbed glass of a scanner, the scanned text will be
slanted in the document image. This rotation in the printed/scanned document can
deceive the verifier as discussed earlier in section (5.6). As a result, the document
might be considered as a falsified paper when actually it is authentic. For that
reason, it is necessary to identify the orientation of the scanned image and to
compute its degree of rotation in order to deskew and straighten the text lines before
validating the document (Baird, 1995).
Indeed, unaltered documents are supposed to be verifiable despite the
unintentional rotation that can happen during scanning. However, the system often
fails to verify rotated documents with genuine contents because binary document
images are highly sensitive to any degree of rotation. The rotation causes shifting in
centroid positions. As a result, the image will be divided in different ways creating
dissimilar sub-images compared with the original. Consequently, the rotated image
will be considered as a false document which makes the verification system
Chapter 6: Conclusions and Future Research
130
vulnerable and causes a false-positive error. Also, the failure of Method 1 to verify
unaltered scanned images is because this method is very sensitive to rotation, as
discussed in Chapter Five.
Due to the substantial impact of rotation and skewing on the authentication
process, it has been vital to find the rotation degree in order to regulate the rotated
image and restore its original form before verifying it. Images with borders or with
plain vertical/horizontal lines can be adjusted simply by computing the rotational
degree. However, documents with no straight lines need other adjustment
techniques. To avoid the false-positive-error caused by rotation, one of the deskew
techniques proposed in the literature should be used to align scanned documents
prior to the verification process.
6.1.3 Optical Character Recognition (OCR)
If the actual formatted text of a document is used as preservative data and
embedded in the attached barcode, the OCR technique can be used to convert the
document image into formatted text (Wang & Bunke, 1997) and this text can be
compared with the preservative data previously embedded in the document itself.
The use of OCR could help to avoid the weaknesses of Method 1 in recognizing ASCII
characters as previously discussed in section 4.6.
However, OCR has limitations because it can be used only for text documents
that do not contain any images and graphs. Also, multilingual documents need an
OCR system which is able to recognize different languages. Therefore, the reliability
of the verification system will depend on the efficiency of the OCR. The failure of the
OCR to detect the text correctly could lead to false-positive errors. As a result, the
OCR software has to be carefully chosen and tested on the text used in typical
documents.
Chapter 6: Conclusions and Future Research
131
6.1.4 Data Compression
The size of the preservative data extracted from the original document can be
reduced using a data compression technique to embed the maximum amount of
information about the original document in the barcode in order to increase the
reliability of the verification system.
The range of black pixels rates needed to represent the binary sub-images can
be computed. Each value might be represented in fewer bits if the values were
reduced by scaling. For example, if the range is (R: min to max), it can be re-scaled
to (R‘:0..max-min) in order to diminish the size of the preservative data. Also, the
values of the calculated centroids can be ranged and encoded. Therefore, fewer bits
will be needed to represent the preservative data. Popular data compression
algorithms such as Run Length Encoding (RLE) can be used to reduce the size of the
data (Salomon, 2007).
6.1.5 Data Encryption
As discussed in Chapter 4, the preservative data are information about some
areas in the binary document image which may be subject to forgery. This
information could be the location of centroids, rate of black pixels, or a combination
of both. If counterfeiters know what the representation of the area of interest is,
they could re-generate the preservative data after altering the document image.
Then, this data can be stored in a barcode and attached to the document. The
verification system will authenticate this forged document because its details match
the data extracted from the barcode.
Therefore, it is recommended to encrypt the preservative data before saving that
data in the barcode. One of the most popular encryption algorithms such as DES
Chapter 6: Conclusions and Future Research
132
(Data Encryption Standard) or RSA which is proposed by Ron Rivest, Adi Shamir,
and Leonard Adleman, or any other secure algorithm can be used. A public/private
keys must be used during encryption where only the creator knows the encryption
secret key. On the other hand, the verification system has the decryption key which
is used to convert the ciphered data into plain preservative data (Mollin, 2007;
Konheim, 2007). If counterfeiters intend to produce a valid fake document, they
need to know not only the format of the preservative data but also the encryption
keys. Therefore, the use of a data encryption algorithm makes it harder for
counterfeiters to create modified documents which can be verified by the
authentication system.
6.1.6 Auto Selection of Data Embedding
The system could be modified to select a suitable data embedding technique
automatically if there were more than one technique available. This can be decided
depending on the size of the preservative data and the hiding capacity of the host in
each existing technique. Also, the selection priority could vary depending on the
system requirements. For example, if the 2D data matrix barcode and the dot
shifting discussed in section 6.2.5 were the two available embedding methods, the
system should be able to choose the one with higher data capacity to hold the
preservative data. The Dot-shifting method should be chosen only if there are
enough dots in the host document image to carry the preservative data. Otherwise,
the data would be encoded in a 2D barcode as shown in Figure (6.1).
Chapter 6: Conclusions and Future Research
133
Figure 6.1: A Flowchart for the Auto-Selection of Data Embedding Technique
6.2 Conclusions
The print-scan operation has a substantial impact on the contents of documents.
Given this fact, scanned images cannot be exact copies of their originals because of
the potential noise and rotations that can be introduced into the document during
print or scan operations. The proposed methods have been designed to verify
unaltered scanned document images as well as to detect the potential intentional
alternations in ordinary paper documents. The contribution of the proposed system is
the combination of using centroids, quadtree, and 2D Data Matrix barcodes that
generate a novel system to verify scanned documents efficiently. The unique use of
the above mentioned combination has created efficient methods to detect forgery
accurately in document images. The superior ability of the methods proposed in this
research to verify monochrome document images explains why those methods have
a higher level of accuracy in detection than other techniques previously proposed in
literature. The rates of black pixels and the centroid points were used as the
preservative data in the three methods. Those methods have shown different levels
Compute the number of dotted letters in
the host image The preservative data
Embed the data in a barcode
Use Dots-Shifting Data Hiding Technique
Is
DHC of the host enough to
hold the preservative
data?
Chapter 6: Conclusions and Future Research
134
of sensitivity to D/A conversion by low-priced scanners. Fifty five unaltered scanned
documents and 110 slightly altered documents were used in experiments to test the
verification systems of the proposed methods. The majority of the scanned images
have successfully been verified while the majority of the forged documents have
been rejected. Method 2 has a high level of reliability for distinguishing between
malicious and accidental modifications in the scanned documents. It is the least
sensitive method for artefacts introduced by the print-scan operation because it has
effectively managed to verify unaltered, but scanned, documents. Method 1 is
sensitive to D/A conversion but is also effective in detecting small intentional
alterations in forged documents. The locations of altered parts have also been
detected by the proposed methods.
Appendices
137
Appendix B
Appendix B: Two Sample Document Images Divided by Method 2
The depth levels of image division are shown in different colours. Level 1:Green, Level 2:Red, Level 3:Blue, and Level 4:Purple.
Appendices
138
Appendix C
Appendix C: Two Sample Document Images Divided by The Extended Version of Method 2
The depth levels of image division are shown in different colours. Level 1:Green, Level 2:Red, Level
3:Blue, Level 4:Purple, Level 5: Silver, Level 6:Yellow, Level 7:Black, and Level 8:Aqua.
References
140
List of References
Agrawal, M. & Doermann, D. (2009), "Clutter Noise Removal in Binary
Document Images", International Conference on Document Analysis
and Recognition (ICDAR '09).
Alattar A. M. and Alattar O M. (2004) “Watermarking electronic text
documents containing justified paragraphs and irregular line spacing”.
Security, Steganography, and Watermarking of Multimedia Contents
2004: pp. 685-695.
Arnold, M. (2003). Techniques and Applications of Digital Watermarking and
Content Protection. Norwood, MA: Artech House, Incorporated.
Baird, H.S. (1995), "The skew angle of printed documents". In Document
image analysis, IEEE Computer Society Press, Los Alamitos, CA, USA.
Berg, M. D. & Cheong, O. (2008), Computational Geometry: Algorithms and
Applications, 3rd ed, Springer-Verlag, Heidelberg.
Bourke, P. (1988), Calculating the Area and Centroid of A Polygon. Available
from: The University of Western Australia, Web site:
http://local.wasp.uwa.edu.au/~pbourke/geometry/polyarea/
[Accessed: March 11, 2009].
Brown, L. (1993), The New Shorter Oxford English Dictionary on Historical
Principles, 2nd ed, Oxford University Press Inc. New York, USA.
Chen, M., Memon, N. & Wong, E. K. (2004). Data Hiding in Document
Images. In Lu C. Multimedia Security: Steganography and Digital
Watermarking Techniques for Protection of Intellectual Property (pp.
References
141
231-247). Hershey, PA, USA: Ideal Group Publishing.
Chow, S. Serinken, N. & Seymour, S. (1993), Security Technology 1993,
Institute of Electrical and Electronics Engineers 1993 International
Carnahan Conference on Security Technology, Proceedings, vol. 6, pp.
11-14.
Chung, K.L. & Chen, P.C. (2005), "An efficient algorithm for computing
moments on a block representation of a greyscale image", Pattern
Recognition , vol. 38, no. 12, pp. 2578-2586.
Dai, M. Baylou, P. & Najim, M. (1992), "An efficient algorithm for
computation of shape moments from run-length codes or chain
codes", Pattern Recognition , vol. 25, no. 10, pp. 1119-1128.
Digimarc (2009), Digimarc (DMRC) Enriched Media, Enriching Lives.
Available from: https://www.digimarc.com [Accessed: February 12,
2009].
Easton, V.J. & McColl, J.H. (1997), Statistics Glossary v1.1. Available from:
University of Glasgow, Web site:
http://www.stats.gla.ac.uk/steps/glossary/index.html [Accessed:
September 19, 2009].
Finkel, R. A. and Bentley, J. L. (1974), "Quad trees a data structure for
retrieval on composite keys ", Acta Informatica, vol. 4, no. 1, pp. 1-9.
Flusser, J. (1998), "Fast calculation of geometric moments of binary
images", in: Proc. 22nd OAGM’98 Workshop Pattern Recognition
Medical Computer Vision, vol. no. pp. 265–274.
References
142
Fridrich J (1998b), "Methods for Detecting Changes in Digital Images", Proc.
of the 6th IEEE International Workshop on Intelligent Signal Processing
and Communication Systems, 5-6 Nov. 1998, Melbourne, Australia,
pp.173-177.
Fridrich J. (1998a), "Image Watermarking for Tamper Detection ", Proc.
IEEE International Conf. on Image Processing, ICIP-98, Vol.2, pp.404-
408.
Fridrich J. (1999a), "Methods for tamper detection in digital images," Proc.
ACM Workshop on Multimedia and Security, pp. 19--23, Orlando, FL,
October 30--31, 1999.
Fridrich J. (1999b), A Hybrid Watermark for Tamper Detection in Digital
Images, ISSPA'99 Conf., Brisbane, Australia, August 22-25, 1999.
Fridrich J., Goljan M., and Baldoza A. (2000), "New fragile authentication
watermark for images," in Proceedings of IEEE International
Conference on Image Processing, (Vancouver, Canada), September
10-13, 2000.
Fridrich, J. & Goljan, M. (1999). Protection of Digital Images Using Self
Embedding. Symposium on Content Security and Data Hiding in Digital
Media.
Fridrich, J. Goljan, M. & Lisonek, P. (2005), "Writing on Wet Paper", IEEE
Trans. on Signal Processing, vol. 53, no. 2, pp. 3923- 3935.
Furht, B., Muharemagic, E. & Socek, D. (2005). Multimedia Encryption and
Watermarking. New York: Springer.
Gou, H. & Wu, M. (2007), "Improving Embedding Payload in Binary Images
with "Super-Pixels"", Image Processing, 2007. ICIP 2007. IEEE
References
143
International Conference on, vol. 3, pp. 277- 280.
Han, S. Y., Jung E. H., and Cho S. Y. (2005). A Robust Digital Watermarking
Adopting 2D Barcode. Computer Analysis of Images and Patterns: 11th
International Conference, CAIP 2005, Versailles, France, September 5-
8, 2005. Proceedings, 3691 / 2005 p 717-723.
Har-Peled, S. (2008), Geometric Approximation Algorithms. Available from:
University of Illinois, Department of Computer Science Web site:
http://valis.cs.uiuc.edu/~sariel/teach/notes/aprx/ [Accessed: March
13, 2009].
Hsieh, M. & Tseng, D. (2004), "Perceptual Digital Watermarking for Image
Authentication in Electronic Commerce", Electronic Commerce
Research, vol. 4, no. 1-2, pp. 157-170.
Katzenbeisser, S. (2000), Information Hiding Techniques for Steganography
and Digital Watermarking, Artech House Books, Boston, MA.
Kim H. Y. and Queiroz R. L (2004a). ‘A public-key authentication
watermarking for binary images’. ICIP 2004: 3459-3462.
Kim, H. Y. and Queiroz, R. L. (2004b), “Alteration-Locating Authentication
Watermarking for Binary Images,” Int. Workshop on Digital
Watermarking 2004, (Seoul), Lecture Notes in Computer Science 3304,
pp. 125-136, 2004.
Kim, H.Y. & Afif, A. (2003), "Secure Authentication Watermarking for Binary
Images ", XVI Brazilian Symposium on Computer Graphics and Image
Processing (SIBGRAPI'03), pp. 199.
Kim, H.Y. & Afif, A. (2004), "A Secure Authentication Watermarking for
References
144
Halftone and Binary Images", International Journal of Imaging
Systems and Technology, vol. 14, no. 4, pp. 147 - 152.
Kim, Y. & Oh, I. (2004). Watermarking text Document images using edge
direction histogram. Pattern Recognition Letters, 25 (11), 1243-1251.
Konheim, A. G. (2007), Computer Security and Cryptography, Wiley & Sons,
Inc., Hoboken, New Jersey.
Koschan, A. & Abidi, M. (2008), Digital Color Image Processing, John Wiley &
Sons, Inc. Hoboken, New Jersey.
Kwok, S.H. Yang, C.C. & Tam, K.Y. (2000), "Watermark Design Pattern for
Intellectual property Protection in Electronic Commerce Applications",
Proceedings of the 33rd Hawaii International Conference on System
Sciences, vol. 6, pp. 6038.
Lan, T. & Tawfik, A. H. (1999). Fraud Detection and Self Embedding.
Proceedings of the seventh ACM international conference on
Multimedia (part 2), 33 – 36.
Li, B.C. (1993), "A New Computation of Geometric Moments", Pattern
Recognition , vol. 26, no. 1, pp. 109-113.
Li, C. Lou, D. & Liu, J. (2003), "Image Integrity and Authenticity Verification
Via Content-based Watermarks and a Public Key Cryptosystem",
Journal of Chinese Institute of Electrical Engineering, vol. 10, no. 1,
pp. 99-106.
Lin, C. & Chang, S. (2001). SARI: self-authentication-and-recovery image
watermarking system. International Multimedia Conference, volume 9,
References
145
628 – 629.
Liu, H. Rui, W. & Huang, J. (2007), "Binary Image Authentication using
Zernike Moments", IEEE International Conference on Image
Processing, ICIP 2007. , vol. 1, pp. 385 - 388.
Low, S. H., Maxemchuk, N. F. & Lapone, A. M. (1998). Document
Identification for Copyright Protection Using Centroid Detection. IEEE
Transactions on Communications, 46 (3), 372-383.
Lu H, Kot, A.C. and Cheng J (2003); ‘Secure data hiding in binary document
images for authentication’, Circuits and Systems, 2003. ISCAS '03.
Proceedings of the 2003 International Symposium on, Volume 3, 25-28
May 2003, Pages: III-806 - III-809.
Mei Q. G., Wong E. K. and Memon N. D. (2001), ‘Data hiding in binary text
documents,’ in Proc. of SPIE Security and Watermarking of Multimedia
Contents III, Aug. 2001, vol. 4314, pp.369–375.
Mohanty, S.P. Ramakrishnan, K.R. & Kankanhalli, M. (1999), "A dual
watermarking technique for images", Proceedings of the seventh ACM
international conference on Multimedia, vol. 2, pp. 49 - 51.
Mollin, R. A (2007), An Introduction to Cryptography, 2nd ed, Chapman &
Hall/CRC, New York.
Onwuegbuzie, A.J. & Daniel, L. G. (2003), "Typology of Analytical and
Interpretational Errors in Quantitative and Qualitative Educational
Research", Current Issues in Education, Vol.6, No.2.
Pratap, R. & Ruina, A. (2009), "Center of mass and gravity". In Introduction
to Statics and Dynamics, Oxford University Press, Oxford.
References
146
Pratt, W. K. (2007), Digital Image Processing, 4th ed, John Wiley & Sons,
Inc. Los Altos, California.
Puhan, N.B. & Ho, A.T. (2005), "Binary Document Image Watermarking for
Secure Authentication Using Perceptual Modeling", Signal Processing
and Information Technology, 2005. Proceedings of the Fifth IEEE
International Symposium, vol. no. pp. 393 - 398.
QuadTree (2009), QuadTree. Available from:, Wikipedia, the free
encyclopedia Web site: http://en.wikipedia.org/wiki/Quadtree
[Accessed: June 29, 2009].
Rogers, T. & Link, T. (1998), Amazing Applications of Probability and
Statistics. Available from: Intuitor.Com, Web site:
http://www.intuitor.com/statistics/ [Accessed: September 19, 2009].
Salomon, D. (2007), Data Compression: the Complete Reference, 4th ed,
Springer, London.
Schneider, M. (1996), "A robust content based digital signature for image
authentication", International Conference on Image Processing,
Proceedings. vol. 3, pp. 227 - 230.
Sharkas, M. Elshafie, D. & Hamdy, N. (2005), "A Dual Digital-Image
Watermarking Technique", PROCEEDINGS OF WORLD ACADEMY OF
SCIENCE, ENGINEERING AND TECHNOLOGY, vol. 5, pp. 136-139.
Signum Technologies (2009), Digital Watermarking from Signum
Technologies. Available from: http://www.signumtech.com [Accessed:
February 12, 2009].
References
147
Sun, Q.B. Feng, P.R. & Deng, R. (2001), "An optical watermarking solution
for authenticating printeddocuments", Information Technology: Coding
and Computing, 2001. Proceedings. International Conference on, pp.
65 - 70.
Sutcu, Y. Coskun, B. & Sencar, H.T. (2007), "Tamper Detection Based on
Regularity of Wavelet Transform Coefficients", IEEE International
Conference on Image Processing. ICIP 2007., vol. 1, pp. 397 – 400.
Tan, C.L., Huang, W. & Yu, Z. (2002), "Imaged Document Text Retrieval
Without OCR", IEEE Transaction on Pattern Analysis and Machine
Intelligence, vol. 24, no. 6, pp. 838-844.
The IET (2005) “The 2D Data Matrix Barcode”, Computing and Control
Engineering. December 2005, Volume 16, Issue 6, p. 39.
Tsai, J. Huang, W. & Chen, C. (2007), "A Feature-Based Digital Image
Watermarking for Copyright Protection and Content Authentication",
IEEE International Conference on Image Processing, ICIP 2007, vol. 5,
pp. 469 - 472.
Wang, P. S. & Bunke, H.(1997), Handbook on Optical Character Recognition
and Document Image Analysis, World Scientific Publishing Co Pte Ltd,
London.
Wong, P.W. (1998), "A public key watermark for image verification and
authentication", International Conference on Image Processing, 1998.
ICIP 98. Proceedings, vol. 1, pp. 455-459.
Xiaoyi, Y. Yunhong, W. & Tieniu, T. (2004), "Steganalysis of Data Hiding in
References
148
Binary Images", Proceedings of the Pattern Recognition, 17th
International Conference on (ICPR'04), vol. 4, pp. 877 - 880.
Yang H. and Kot A. C. (2004) ‘Data hiding for bi-level documents using
smoothing technique’. ISCAS (5) 2004: 692-695.
Yeung, M. M. and Mintzer, F. C. (1997) "Invisible watermarking for image
verification," in Proc. International Conference on Image Processing,
vol. 2, (Washington, DC), pp. 680--683, October 1997.
Yu, L. Niu, X. & Sun, S. (2005), "Print-and-Scan Model and the
Watermarking Countermeasure", Image and Vision Computing, vol. 23,
pp. 807-814.
Zhao, J. & Koch, E. (1995), "Embedding Robust Labels into Images for
Copyright Protection", Proceedings of the International Congress on
Intellectual Property Rights for Specialized Information, Knowledge and
New Technologies, vol. pp. 242-251.
Zhu, B., Wu, J. & Kankanhalli, M. S. (2003). Print signatures for document
authentication. Conference on Computer and Communications Security
(Proceedings of the 10th ACM conference on Computer and
communications security), 145 - 154.