+ All Categories
Home > Documents > An improved data mining technique for classification and detection of breast cancer from mammograms

An improved data mining technique for classification and detection of breast cancer from mammograms

Date post: 04-Dec-2016
Category:
Upload: saroj-kumar
View: 213 times
Download: 0 times
Share this document with a friend
8
ORIGINAL ARTICLE An improved data mining technique for classification and detection of breast cancer from mammograms Aswini Kumar Mohanty Manas Ranjan Senapati Saroj Kumar Lenka Received: 7 July 2011 / Accepted: 5 January 2012 Ó Springer-Verlag London Limited 2012 Abstract The high incidence of breast cancer in women has increased significantly in the recent years. Mammo- gram breast X-ray imaging is considered the most effec- tive, low-cost, and reliable method in early detection of breast cancer. Although general rules for the differentiation between benign and malignant breast lesion exist, only 15–30% of masses referred for surgical biopsy are actually malignant. Physician experience of detecting breast cancer can be assisted by using some computerized feature extraction and classification algorithms. Computer-aided classification system was used to help in diagnosing abnormalities faster than traditional screening program without the drawback attribute to human factors. In this work, an approach is proposed to develop a computer-aided classification system for cancer detection from digital mammograms. The proposed system consists of three major steps. The first step is region of interest (ROI) extraction of 256 9 256 pixels size. The second step is the feature extraction; we used a set of 26 features, and we found that these features are capable of differentiating between normal and cancerous breast tissues in order to minimize the classification error. The third step is the classification process; we used the technique of the asso- ciation rule mining to classify between normal and cancerous tissues. The proposed system was shown to have the large potential for cancer detection from digital mammograms. Keywords Data mining Computer-aided diagnosis Mammogram classification Statistical features Gray level co-occurrence matrix features Association rule mining Region of interest 1 Introduction Breast cancer is a leading cause of fatality among all cancers for women. However, the etiologies of breast cancer are unknown, and no single dominant cause has emerged. Still, there is no known way of preventing breast cancer, but early detection allows treatment before it is spread to other parts of the body. Currently, X-ray mam- mography is the single most effective, low-cost, and highly sensitive technique for detecting small lesions resulting in at least a 30% reduction in breast cancer deaths [1]. However, the sensitivity of mammography is highly chal- lenged by the presence of dense breast parenchyma, which deteriorates both detection and characterization tasks [2]. It may not be feasible to routinely perform a second reading by a radiologist due to financial, technical, and logistical restraints. Therefore, efforts were made to develop a computer-aided detection (CAD) system [3, 4]. CAD can be defined as a diagnosis made to improve radiologists’ performance by indicating the sites of potential abnor- malities, to reduce the number of missed lesions, and/or by providing quantitative analysis of specific regions in an A. K. Mohanty (&) SOA University, Bhubaneswar, Orissa, India e-mail: [email protected] M. R. Senapati Department of Computer Science, Krupajal Engineering College, Bhubaneswar 752002, Orissa, India e-mail: [email protected] S. K. Lenka Department of Computer Science, Modi Univesity, Lakshmangarh 332311, Rajasthan, India e-mail: [email protected] 123 Neural Comput & Applic DOI 10.1007/s00521-012-0834-4
Transcript

ORIGINAL ARTICLE

An improved data mining technique for classificationand detection of breast cancer from mammograms

Aswini Kumar Mohanty • Manas Ranjan Senapati •

Saroj Kumar Lenka

Received: 7 July 2011 / Accepted: 5 January 2012

� Springer-Verlag London Limited 2012

Abstract The high incidence of breast cancer in women

has increased significantly in the recent years. Mammo-

gram breast X-ray imaging is considered the most effec-

tive, low-cost, and reliable method in early detection of

breast cancer. Although general rules for the differentiation

between benign and malignant breast lesion exist, only

15–30% of masses referred for surgical biopsy are actually

malignant. Physician experience of detecting breast cancer

can be assisted by using some computerized feature

extraction and classification algorithms. Computer-aided

classification system was used to help in diagnosing

abnormalities faster than traditional screening program

without the drawback attribute to human factors. In this

work, an approach is proposed to develop a computer-aided

classification system for cancer detection from digital

mammograms. The proposed system consists of three

major steps. The first step is region of interest (ROI)

extraction of 256 9 256 pixels size. The second step is the

feature extraction; we used a set of 26 features, and we

found that these features are capable of differentiating

between normal and cancerous breast tissues in order to

minimize the classification error. The third step is the

classification process; we used the technique of the asso-

ciation rule mining to classify between normal and

cancerous tissues. The proposed system was shown to have

the large potential for cancer detection from digital

mammograms.

Keywords Data mining � Computer-aided diagnosis �Mammogram classification � Statistical features �Gray level co-occurrence matrix features �Association rule mining � Region of interest

1 Introduction

Breast cancer is a leading cause of fatality among all

cancers for women. However, the etiologies of breast

cancer are unknown, and no single dominant cause has

emerged. Still, there is no known way of preventing breast

cancer, but early detection allows treatment before it is

spread to other parts of the body. Currently, X-ray mam-

mography is the single most effective, low-cost, and highly

sensitive technique for detecting small lesions resulting in

at least a 30% reduction in breast cancer deaths [1].

However, the sensitivity of mammography is highly chal-

lenged by the presence of dense breast parenchyma, which

deteriorates both detection and characterization tasks [2]. It

may not be feasible to routinely perform a second reading

by a radiologist due to financial, technical, and logistical

restraints. Therefore, efforts were made to develop a

computer-aided detection (CAD) system [3, 4]. CAD can

be defined as a diagnosis made to improve radiologists’

performance by indicating the sites of potential abnor-

malities, to reduce the number of missed lesions, and/or by

providing quantitative analysis of specific regions in an

A. K. Mohanty (&)

SOA University, Bhubaneswar, Orissa, India

e-mail: [email protected]

M. R. Senapati

Department of Computer Science, Krupajal Engineering

College, Bhubaneswar 752002, Orissa, India

e-mail: [email protected]

S. K. Lenka

Department of Computer Science, Modi Univesity,

Lakshmangarh 332311, Rajasthan, India

e-mail: [email protected]

123

Neural Comput & Applic

DOI 10.1007/s00521-012-0834-4

image to improve diagnosis. CAD systems typically oper-

ate as automated ‘‘second opinion’’ or ‘‘double reading’’

systems [5].

Many techniques have been used to detect masses in the

mammograms. Youssry et al. [6] used a technique that

depends mainly on the difference between normal and

cancerous histograms and used four features for the clas-

sification process through a neural network classifier. The

four features are statistical ones, which are the mean and

the first three moments. Preprocessing techniques were

used such as histogram equalization and segmentation.

Kobatake et al. [7] presented a Computer-Aided Diagnostic

(CAD) system for the detection of malignant tumors on

digital mammograms and used 9 features to identify

malignant tumors. Yu et al. [8] presented a CAD system for

the automatic detection of clustered microcalcification

through two steps. The first one is to segment potential

microcalcification pixels by using wavelet and gray level

statistical features and to connect them into potential

individual microcalcification objects. The second step is to

check these potential objects by using 31 statistical fea-

tures. Neural network classifiers were used. Results are

satisfactory but not highly guaranteed because the learning

set was used in the testing set. Verma et al. [9] presented

technique that based on fuzzy-neural and feature extrac-

tion, and used 14 features for the proposed method. A

fuzzy technique in conjunction with three features was

used to detect a microcalcification pattern and a neural

network to classify it into benign or malignant. Senapati

et al. [10] have used neural network approach to classify

tumors into benign and malignant types. Hagihara et al.

[11] presented a CAD for breast cancers by improvement

of classifiers, used 5 features related to the concurrency

matrix, 3 features related to the density histogram, and one

feature related to the shape of the extracted region. Furuya

et al. [12] improvement of performance to discriminate

malignant tumors from normal tissue on mammograms by

feature selection and evaluation of features selection cri-

teria, and used four types features, first-order statistics

features, second-order statistic (co-occurrence) features,

density features, and shape features. Fogela et al. [13] used

the patient age as a feature besides radiographic features to

train artificial neural networks to detect breast cancer.

Brake et al. [14] studied the scale effect on the detection

process by using single scale and multi-scale detection

algorithms of masses in digital mammograms. The imple-

mentation of such techniques had to use different sets of

features which used to differentiate between normal and

cancer breast tissue from digital mammograms. In this

paper, the development of a computer-aided classification

system for cancer detection from digital mammograms is

presented. A new computer-aided classification system

used most effective sets of features that can differentiate

between normal and cancer breast tissue. The proposed

system consists of three major steps:

The first step is ROI extraction of 256 9 256 pixels size.

The second step is the feature extraction, where a set of 26

features are used and these features are capable of differ-

entiating between normal and cancerous breast tissues are

found. The third step is the classification process; the

technique of the association rule miner is used to classify

between normal and cancerous tissues.

2 Materials and methods

This study is done through two main phases: the learning

phase and the testing phase. Through the learning phase,

the system is learned by using normal and cancerous

images to differentiate between normal and cancerous

cases. In the testing phase, the performance of the system is

test by entering a test image to compute the correctness

degree of the system decision. The Fig. 1 shows a sche-

matic diagram for the proposed system.

2.1 Mammogram database

The mammogram images used in this paper are provided

by the University of South Florida, the digital database for

screening mammography (DDSM) [15]. The dataset con-

sists of digitized mammogram images, composed of both

oblique and cranio-caudal views. Each mammogram shows

one or more tumor mass marked by expert radiologists. The

position of individual masses is marked. The location of

the abnormalities in form of its boundary provided as chain

code where the first two values are the starting column and

row of the lesion boundary while other numbers correspond

to a specific direction on the X and Y coordinates. The

images are digitized from films using the Lumysis scanner

Fig. 1 Block diagram for the proposed system

Neural Comput & Applic

123

with 12 bits depth. Figure 2 shown below one spiculated

cancerous mass along with its ROI and detection.

2.2 Selection of ROI

Using the contour supplied by the DDSM for each mam-

mogram, the ROI of size 256 9 256 pixels is extracted

with mass centered in the window and is divided into two

sets: the learning set and the testing set. The learning set is

composed of 88 cancerous images and 88 normal images,

while the testing set contained 57 cancerous images and 32

normal images. The normal images are taken from the

same image that has cancerous regions. Different ROI and

its processing are shown below in Fig. 3.

2.3 Feature extraction

A typical mammogram contains a vast amount of hetero-

geneous information that depicts different tissues, vessels,

ducts, chest skin, breast edge, the film, and the X-ray

machine characteristics. In order to build a robust diag-

nostic system toward correctly classifying normal and

abnormal regions of mammograms, we have to present all

the available information that exists in mammograms to the

diagnostic system so that it can easily discriminate between

the normal and the abnormal tissue. However, the use of all

the heterogeneous information results to high-dimensioned

feature vectors that degrade the diagnostic accuracy of the

utilized systems significantly as well as increase their

Fig. 2 a The mammogram. b The specified ROI. c The detected spiculated with a spiculated cancerous mass

Fig. 3 a 600 9 600 pixels ROI containing a malignant MC cluster in

original mammogram (DDSM: B_3406_RIGHT_CC), b processed

ROI, c binarization on manually defined ROI, d binarized MC cluster

on 600 9 600 pixels ROI, e surrounding tissue ROI (ST-ROI),

f magnified 128 9 128 pixels sub region of ST-ROI

Neural Comput & Applic

123

computational complexity. Therefore, reliable feature vec-

tors should be considered that reduce the amount of irrel-

evant information, thus producing robust mammography

descriptors of compact size. In our approach, we examined

a set of 26 features were applied to the ROI using a window

of size 64 pixels with 64 pixels shift, that is, no overlap.

The features extracted in this study divided into two

categories: first-order statistics features and second-order

statistics (gray level co-occurrence matrix) features,

1. First-order statistics features: They provide different

statistical properties of the intensity histogram of an

image [16]. They depend only on individual pixel

values and not on the interaction or co-occurrence of

neighboring pixel values. In this study, first-order

textural features were calculated: mean value of gray

levels, standard deviation of gray levels, kurtosis,

skewness, variance and maximum of gray level, range

of gray level, entropy, second moment, and percentile.

2. Second-order statistics (gray level co-occurrence

matrix) features: The gray level co-occurrence matrix

(GLCM) is a well-established robust statistical tool for

extracting second-order texture information from

images [17, 18]. The GLCM characterizes the spatial

distribution of gray levels in the selected ROI sub

region. An element at location (i, j) of the GLCM

signifies the joint probability density of the occurrence

of gray levels i and j in a specified orientation h and

specified distance d from each other. Thus, for

different h and d values, different GLCMs are

generated. In this study, four GLCMs corresponding

to four different directions (h = 0�, 45�, 90� and 135�)

and one distance (d = 1 pixel) were computed for each

selected ROI sub region. Sixteen features were derived

from each GLCM. Specifically, the features studied are

energy, contrast, homogeneity, correlation, first-order

difference moment, entropy of co-occurrence, maxi-

mum of co-occurrence, shade, prominence, second-

order inverse difference moment, information correla-

tion2, sum of squares, sum average, sum entropy, and

difference entropy. Four values were obtained for each

feature corresponding to the four matrices.

2.4 Feature selection

The purpose of this step is to get the features that have the

ability of differentiation between normality and cancer to

be used in the classification process. In other words, the

discrimination power of the features is test. The input to

this test is two sets of values for each feature. One set

represents the normal case, and the other set represents the

cancerous case. We assume that each set follows a t dis-

tribution. The t-test checks the amount of overlapping

between the two distributions. If there is no overlapping,

then this feature has the ability of differentiation. But in

nature, it is not easy to find complete independent distri-

butions without overlapping. So, we determine a signifi-

cance level to consider the two sets come from two

different distributions. We chose this significance level to

be 5%. It means that the probability of incorrectly con-

sidering two independent distributions is 0.05 while the

truth is that the two sets come from the same distributions.

The test computes a value called the p value which is the

probability of observing one sample from the first set in the

second distribution. If the p value is less than the signifi-

cance level, then these two sets come from two different

distributions and this feature can differentiate [19].

To prepare the two sets of each feature, the feature

matrix resulted from the step of features extraction is used.

For each feature, we transfer the matrix of each image to a

vector. Thus, we have for each feature a number of vectors

equal to the numbers of the sample normal and cancerous

images. These vectors are concatenated under each other to

form the normal cluster and the cancerous cluster, and

these two sets are the input to the t = test step.

2.5 Classification

The classification process is consisting of feature dictio-

nary and association rule mining classifier where the details

follows divided into the learning phase and the testing

phase. In the learning phase, known data are given, and the

feature parameters are calculated by the processing which

precedes classification.

Separately, the data on a candidate region that has

already been decided as a tumor or as normal are given,

and the classifier is trained. We used the learning set for

this phase which consists of 88 cancerous ROI and 88

normal ROI. In the testing phase, unknown data are given

and the classification is performed using the classifier after

learning. Breast cancer image diagnosis assistance is the

task in the testing phase. We used a testing set for this

phase which consists of 57 cancerous ROI and 32 normal

ROI. Associative image classifier was used in our experi-

ment to find the sensitivity and specificity.

2.5.1 Creating the feature dictionary

The dictionary consists of the most typical statistical fea-

tures and texture features of the individual blocks of

mammograms in the training set. It is created by clustering

corresponding feature values into a chosen number of

groups. The clustering is performed using a k-means

algorithm with the histogram intersection measure for

comparing statistical features and texture features.

Centroids resulting from the clustering operation become

Neural Comput & Applic

123

the elements of the dictionary and are labeled with con-

secutive natural numbers. These identifiers are then used to

describe blocks of the images in the database. During the

classification phase, the previously created dictionary is

queried with statistical and texture feature values and

responds with the labels of the most similar entries.

2.5.2 Associative classifications

Associative classifiers are a recent, two-stage approach to

classification, in which a set of association rules between

the attribute values and category labels is first discovered,

and then a compact classifier is created by selecting the

most important rules for classification.

2.5.2.1 Association rules for classification Formally, an

association rule used for classification is an implication of

the form of X ? C, where item set X is a non-empty subset

of all possible items in the database, X ( I, I = {i1, i2,.

,

…, in}, and c is a class identifier, C

{c1, c2, …, cn}. Let

thus a rule item set be a pair\X, c[, containing the item set

X and a class label c. The rules are discovered in a training

set of transactions Dt. Each transaction is a triple of the

form \tid, Y, c[, containing a transaction identifier tid,

item set Y ( I, and a class label c.

In our approach, we discover the most interesting

association rules between the images in the training set,

described by dictionary entries, and their category labels.

This is a slight modification of the classic association rule

mining problem, as the implication consequent is always

limited to a class label. The aim of the mining is then to

discover the rules that are a subset of general association

rules and have the following form:

Rc : fstat feature1; . . .; stat featuren; texture1; . . .; texturemg) class label: ð1Þ

We adapt the existing methods of association rule mining

to create a classifier suitable for categorization of image

data. Direct application of any rule mining algorithm to a

transactional database containing images represented by

feature values in their particular locations would result in a

large number of irrelevant associations. Following this

observation, we consider only the existence, occurrence

count, and spatial proximity of features to create rules that are

sufficiently general to classify previously unseen examples.

The initial set of discovered rules is usually very large,

so it is necessary to limit the number of associations by

specifying the minimum support and confidence values and

employing various pruning techniques. We use the CBA

approach proposed in [20] to mine the rules along with

frequent item sets and then apply a pruning strategy to limit

their number.

2.5.2.2 Considering occurrence count Extending associ-

ation rules to include the information about item occur-

rence count in multimedia applications was first proposed

in [21]. We use this general idea to mine classification rules

with recurrent items and apply a selected number of such

associations to the problem of image classification. A slight

modification of calculating the support of such rules is

necessary, as a single transaction may increase the support

of an item set by more than one. The support of an item set

X may thus be calculated as [2]:

suppðXÞ ¼XDj j

k¼0

/ðX; tkÞDj j ; ð2Þ

where / is a function that returns the ratio by which a

transaction tk of a database D supports item set X and is

defined as:

/ðX; tÞ ¼ minaj

bj

!; j ¼ 1; . . .; n;

tk ¼ fa1i1; a2i2; . . .; aning;X ¼ fb1i1; b2i2; . . .; bning; a1 6¼ 0; b1 6¼ 0:

ð3Þ

The support of a rule with recurrent items is calculated

similarly as when considering simple association rules, by

counting the support of a set consisting of both the rule’s

antecedent and consequent.

The definition of confidence also remains unchanged

and may be calculated as supp (X [ Y)/supp(X). The defi-

nition of a frequent item set may be extended by including

an additional condition of maximum support supp(X) \ R,

apart from its minimum value supp(X) [ r, which helps to

minimize the number of uninteresting rules.

A modified version of the CBA algorithm, presented as

Algorithm 1, is used to mine all possible rules or only the

rules having a certain maximum number of items in the

antecedent.

In the lines 1–3, a first pass over the database is made to

find all sufficiently frequent item sets, which can be used to

build rules with a single value in the antecedent. The

maximum number of occurrences of every item in the

transactions of the database is also counted, similarly as in

the max-occur algorithm [21], to limit the number of item

recurrences while generating candidates. Line 6 generates

candidates using the Apriori method and includes another

occurrence of an existing item, whenever existing count is

below the maximum value. The count function returns the

item’s current number of occurrences in an item set, while

the � and � symbols denote item set merging and item

concatenation operations, respectively. Lines 7–12 are used

to independently calculate the support of each rule and the

support of its antecedent. These values are then used to

calculate confidence of the rules in line 14.

Neural Comput & Applic

123

2.5.2.3 Considering spatial proximity Apart from the

association rules between the number of particular features

present on the images and their category labels, we also

consider rules that include the information about spatial

proximity of the features. While mining for the association

rules, we check for spatial relationships between the

recurring features and only then include them multiple

times in the rules, when they form a single area on the

image. It is possible to mine such rules without any change

to the abovementioned algorithm by slightly changing the

representation of the images. For every element of the

dictionary, each transaction is scanned to find the largest

area covered by a single feature. The original number of

occurrences of every item is then reduced to that maximum

value before the association rule mining algorithm is

applied. Table 1 illustrates the difference between both

approaches to image representation.

2.5.2.4 Building the classifier Having found all the rules

with minimum and maximum support, as well as the

minimum confidence, we come to the problem of creating a

classifier, which will then be used to associate category

labels with previously unseen images. The final classifier

is created by first sorting the rules according to their

confidence and support in descending order and the number

of items in their antecedents in ascending order. Next, for

every rule in the sorted list, all elements of the training set

matching that rule are found and removed from further

processing. A rule is then added to the classifier if it

matches at least one element of the set. At each step of the

iteration, a default class of the classifier is selected that

minimizes the error of classification of the remaining data.

Lastly, when the rule or dataset is empty, the final classifier

is reduced to the first number of rules that decrease the

general error rate in classification.

2.5.2.5 Classification Classification is performed by

applying the first matching rule from the classifier to a

given image described by dictionary entries. A default class

label is then given to an image for which there is no

matching rule. An image matches a rule when it contains

each of the items of the rule’s antecedent with at least the

same occurrence count.

Table 2 shows an example representation of a few

photographs without considering spatial relationships

between the features, a possible classifier content, and

classification result. The classifier was created using the

CBA approach to limit the number of rules. Dictionary

Algorithm 1 CBA-RG with recurrent items

Table 1 An example of image representation when considering spatial proximity of features

Neural Comput & Applic

123

entries are identified by Bi (statistical) and Ti (texture)

labels. The dictionary size was 8 entries for color and 8 for

texture in this example. The first two images are matched

by the first and second rule of the classifier, respectively,

and are associated with C1 category label. The last two

images are classified using the default class value, as they

remain unmatched by any rule.

3 Results and discussions

3.1 Feature extraction and selection

The previously mentioned 26 features using a window size

of 64 pixels and a window shift of 64 pixels, that is, no

overlap is applied. Features are tested using a hypothesis

test to decide whether or not this feature can discriminate

between normal and abnormal tissues using a significance

level of 0.05. The hypothesis indicated that only 8 features

(kurtosis, spread, information correlation2 at angle 0�,

information correlation2 at angle 45�, information corre-

lation2 at angle 90�, and information correlation2 at angle

135�, difference entropy at angle 45�, and sum entropy at

angle 45�) cannot discriminate between the two clusters

because their p value is larger than the significance level

of 0.05.

3.2 Classification for specificity and sensitivity

We measured, quantitatively, the detection performance of

the classifiers by computing the sensitivity and specificity

on the data. Sensitivity is the conditional probability of

detecting cancer while there is really cancer in the image.

Specificity is the conditional probability of detecting nor-

mal breast while the true state of the breast is normal.

In the terms of the false-negative rate and the false-

positive rate:

Sensitivity ¼ 1� false - negative rate:

Specificity ¼ 1� false - positive rate:

False-negative rate: The probability that the classification

result indicates a normal breast while the true diagnosis is

indeed a breast disease (i.e. positive). This case should be

completely avoided since it represents a danger to the

patient.

False-positive rate: The probability that the classification

result indicates a breast disease while the true diagnosis is

indeed a normal breast (i.e. negative). This case can be

tolerated, but should be as infrequent as possible.

Table 3 shows the results of the association rule mining

classifier.

By testing the learning set, the system detected 88

images from 88 cancerous images and detected 88 images

from 88 normal images. This gives a sensitivity of 100%

and a specificity of 100%.

By testing the testing set, the system detected 55 images

from 57 cancerous images and detected 31 images from 32

normal images. This gives a sensitivity of 96.5% and a

specificity of 96.88%. Comparing the results obtained from

the association rule classifier in this study with the results

obtained from other study [22, 23], sensitivity and speci-

ficity results are much better than that study. These results

are not so bad in compare to other classification techniques,

but on the other hand it is also not too much satisfactory.

This returns to many reasons. The first reason comes from

the great variability in the database mammograms. The

cancer values and the normality values change extensively,

which leads to more overlapping between the normal

cluster space and the cancerous cluster space. The second

reason is the small number of used cases in learning the

system, which does not cover the entire space of each

cluster.

4 Conclusions

Automated breast cancer detection has been studied for

more than two decades Mammography is one of the best

methods in breast cancer detection, but in some cases

radiologists face difficulty in directing the tumors. We

have described a comprehensive of methods in a uniform

Table 2 An example of image representation, classifier content, and the classification results

Table 3 Results for association rule mining classifier

Parameter Association rule mining classifier

Learning set (%) Testing set (%)

Sensitivity 100 96.5

Specificity 100 96.88

Neural Comput & Applic

123

terminology to define general properties and requirements

of local techniques and to enable the readers to select the

efficient method that is optimal for the specific application

in detection of microcalcifications in mammogram images.

Although by now some progress has been achieved, there

are still remaining challenges and directions for future

research, such as, developing better preprocessing,

enhancement and segmentation techniques; designing

better feature extraction, selection and classification algo-

rithms; integration of classifiers to reduce both false posi-

tives and false negatives; employing high-resolution

mammograms and investigating 3D mammograms. The

CAD mammography systems for microcalcifications

detection have gone from crude tools in the research lab-

oratory to commercial systems. Mammogram image anal-

ysis society database is standard test set but defining

different standard test set (database) and better evaluation

criteria are still very important. With some rigorous eval-

uations, and objective and fair comparison could determine

the relative merit of competing algorithms and facilitate the

development of better and robust systems. The methods

like one presented in this paper could assist the medical

staff and improve the accuracy of detection. Our method

can reduce the computation cost of mammogram image

analysis and can be applied to other image analysis appli-

cations. In this method, a computer-aided classification

system for mass detection in the digitized mammograms of

the breast is presented. This system depends on selecting

some features and using them in the classification process.

The 8 features can not differentiate between normality and

cancer after testing their discrimination power is proved.

Association rule classifier gave the best results; a sensi-

tivity of 96.5% and a specificity of 96.88%.

References

1. Cheng H, Lui YM, Freimanis RI (1998) A novel approach to

microcalcification detection using fuzzy logic technique. IEEE

Trans Med Imaging 17(3):442–450

2. Sampat PM, Markey MK, Bovik AC (2005) Computer-aided

detection and diagnosis in mammography. In: Bovik AC (ed)

Handbook of image and video processing, 2nd edn. Academic

Press, Waltham, MA, pp 1195–1217

3. Winsberg F, Elkin M, Macy J, Bordaz V, Weymouth W (1967)

Detection of radiographic abnormalities in mammograms by

means of optical scanning and computer analysis. Radiology

89:211–215

4. Marx C, Malich A, Facius M, Grebenstein U, Sauner D,

Pfleiderer SOR, Kaiser WA (2004) Are unnecessary follow-up

procedures induced by computer-aided diagnosis (CAD) in

mammography? Comparison of Mammographic diagnosis with

and without use of CAD. Eur J Radiol 51:66–72

5. Sajda P, Spence C, Pearson J (2002) Learning Contextual Rela-

tionships in Mammograms Using a Hierarchical Pyramid Neural

Network. IEEE Trans Med Imaging 21(3):239–250

6. Youssry N, Abou-Chadi F, El-Sayad AM (2002) A neural net-

work approach for mass detection in digitized mammograms.

ACBME 109–116

7. Kobatake H, Murakami M, Takeo H, Nawano S (1999) Com-

puterized detection of malignant tumors on digital mammograms.

IEEE Trans Med Imaging 18(5):369–378

8. Yu S, Guan L (2000) A CAD system for the automatic detection

of clustered microcalcifications in digitized mammogram films.

IEEE Trans Med Imaging 19(2):115–126

9. Verma B, Zakos J (2001) A computer-aided diagnosis system for

digital mammograms based on fuzzy-neural and feature extrac-

tion techniques. IEEE Trans Inf Technol Biomed 5(1):46–54

10. Senapati MR, Mohanty AK, Dash S, Dash PK (2011) Local linear

wavelet neural network for breast cancer recognition. Neural

Comput Appl. doi:10.1007/s00521-011-0670-y

11. Hagihara Y, Hagihara Y, Wei J (2005) Enhancement of CAD

system for breast cancers by improvement of classifiers. Syst

Comput Jpn 36(9):65–76

12. Furuya S, Wei T, Hagihara Y, Shimizu A, Kabotake H, Nawano

S (2004) Improvement of performance to discriminate malignant

tumors from normal tissue on mammograms by feature selection

and evaluation of features selection criteria. Syst Comput Jpn

35(7):72–84

13. Fogela DB, Wasson EC, Boughton EM, Pm-to VW (1997) A step

toward computer-assisted mammography using evolutionary

programming and neural networks. Cancer Lett 119:93–97

14. Brake GM, Karssemeijer N (1999) Single and multiscale detec-

tion of masses in digital mammograms. IEEE Trans Med Imaging

18:628–639

15. http://marathon.csee.usf.edu/Mammography/Database.html

16. Gonzalez RC, Woods RE (2002) Digital Image processing.

Prentice-Hall Inc., New Jersey, pp 76–142

17. Haralick RM, Shanmugam K, Dinstein I (1973) Textural features

for image classification. IEEE Trans System Man Cybern SMC-

3:610–621

18. Walker RF, Jackway P, Longstaff ID (1995) Improving co-

occurrence matrix feature discrimination. In Proceedings of 3rd

Conference on Digital Image Computing: Techniques and

Applications (DICTA’95), pp. 643–648, Brisbane, Australia

19. Le CT (2003) Introductory biostatistics. Wiley, London

20. Liu B, Hsu W, Ma Y (1998) Integrating classification and asso-

ciation rule mining. In Proceedings of 4th International Confer-

ences on Knowledge Discovery and Data Mining, pp 80–86

21. Zaı̈ane OR, Han J, Zhu H (2000) Mining recurrent items in

multimedia with progressive resolution refinement. In Proceed-

ings of 16th International Conferences on Data Engineering,

pp 461–470

22. Thangavel K, Kaja Mohideen A (2009) Classification of micro-

calcifications using multi-dimensional genetic association rule

miner. Int J Recent Trends Eng 2(2):233–235

23. Ribeiro MX, Traina AJM, Traina C, Azevedo-Marques PM, Sao

Paulo Univ., Sao Carlos (2008) An association rule-based method

to support medical image diagnosis with efficiency. IEEE Trans

Multimed 10(2):277–285

Neural Comput & Applic

123


Recommended