+ All Categories
Home > Documents > LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION...

LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION...

Date post: 02-Apr-2019
Category:
Upload: ngodung
View: 223 times
Download: 0 times
Share this document with a friend
31
LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION NEURAL NETWORK WITH HARMONY SEARCH ALGORITHM MOHAMED HASSAN AHMED UNIVERSITI TEKNOLOGI MALAYSIA
Transcript
Page 1: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

i

LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION NEURAL

NETWORK WITH HARMONY SEARCH ALGORITHM

MOHAMED HASSAN AHMED

UNIVERSITI TEKNOLOGI MALAYSIA

Page 2: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION NEURAL

NETWORK WITH HARMONY SEARCH ALGORITHM

MOHAMED HASSAN AHMED

A dissertation submitted in partial fulfillment of

the requirements for the award of the degree of

Master of Science (Computer Science)

Faculty of Computing

Universiti Technologi Malaysia

JULY 2013

Page 3: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

iii

“To my beloved parents, wife and children, thank you for your endless support,

encouragement, patience, and understanding and to my lecturers and friends for their

respect and support. May ALLAH always revive our sweet memory”

Page 4: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

iv

ACKNOWLEDGMENT

All praises are due to ALLAH (SWT) whose favors to me is uncountable,

secondly I praise the prophet (PBUH) whose guidance safeguards me from the hell

fire. Next, I wish to extend my gratitude to all those who contributed this study either

directly or indirectly. To take the opportunity of the time, I have to thank my

supervisor, Prof. Dr. Hjh. Siti Mariyam Hj. Shamsuddin for her guidance, attention

and understanding throughout the period of this study. I also send my grateful

appreciation to my family for all their unforgettable patience support and

encouragement they provided me. I also send my gratitude to my examiners Dr. Nor

Azizah Ali and Dr. Haza Nuzly Abd Hamed for their helpful comments and

suggestions.

I am grateful to all my colleagues, friends, staff, and lecturers in Universiti

Teknologi Malaysia (UTM), especially those in the Faculty of Computing, and not

forgetting my colleagues, administration and stuff at SIMAD University for their

helpful support for the whole duration of this course of study.

Page 5: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

v

ABSTRACT

Training Radial Basis Function (RBF) neural network with Particle Swarm

Optimization (PSO) was considered as a major breakthrough, that overcome the

stuck to the local minimum of Back Propagation (BP) and time consuming and

computation expensive problems of Genetic Algorithm (GA). However, PSO proved

some problems to achieve the goal, i.e., it converged too fast so that it stuck to the

local optimum. Furthermore, particles may move to an invisible region. Therefore, to

realize the enhancement of the learning process of RBF and overcome these PSO

problems, Harmony Search Meta-Heuristic Algorithm (HSA) was employed to

optimize the RBF network and attain the desired objectives. The study conducted a

comparative experiments between the integrated HSA-RBF network and the PSO-

RBF network. The results proved that HSA increased the learning capability of RBF

neural network in terms of accuracy and correct classification percentage, error

convergence rate, and less time consumption with less mean squared error (MSE).

The new HSA-RBF model provided higher performance in most cases and promising

results with better classification proficiency compared with that of PSO-RBF

network.

Page 6: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

vi

ABSTRAK

Latihan Asas Fungsi Jejari jaringan saraf bersama serpihan kumpulan

optimum telah diambil kira sebagai penemuan yang besar yang dapat mengatasi

sekatan kepada minimum tempatan penyebaran belakang serta penggunaan masa dan

masalah pengiraan algoritma genetic yang mahal. Walau bagaimana pun serpihan

kumpulan optimum telah membuktikan sesetengah masalah untuk mencapai sasaran

contohnya penumpuannya terlalu laju supaya ia tersekat pada optimum tempatan.

Dalam pada itu serpihan mungkin beralih kepada bahagian tak dapat dilihat. Oleh itu,

untuk memahami penambahbaikan proses pembelajaran Latihan Asas Fungsi Jejari

dan mengatasi masalah serpihan kumpulan optimum, Pencarian Algoritma Harmoni

Meta-Heuristik dijalankan untuk optimumkan rangkaian Latihan Asas Fungsi Jejari

dan mencapai objektif yang disasarkan. Kajian ini di jalankan melalui perbandingan

eksperimen di antara gabungan jaringan Pencarian Algoritma Harmoni Meta-

Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan kumpulan

optimum dan Latihan Asas Fungsi Jejari. Keputusan menunjukkan Pencarian

Algoritma Harmoni Meta-Heuristik menambahkan keupayaan jaringan saraf Latihan

Asas Fungsi Jejari dari segi ketepatan dan peratusan klasifikasi, kesalahan kadar

penumpuan, dan pengurangan masa dengan pengurangan kesalahan purata persegi

(MSE). Penghasilan gabungan Pencarian Algoritma Harmoni Meta-Heuristik dan

Latihan Asas Fungsi Jejari ini menyediakan prestasi yang lebih tinggi di dalam

kebanyakan kes serta menjanjikan keputusan yang lebih baik kecekapan klasifikasi

berbanding dengan gabungan jaringan serpihan kumpulan optimum dan Latihan Asas

Fungsi Jejari.

Page 7: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

vii

TABLE OF CONTENTS

CHAPTER TITLE PAGE

DECLARATION ii

DEDICATION iii

ACKNOWLEDGMENT iv

ABSTRACT v

ABSTRAK vi

TABLE OF CONTENTS vii

LIST OF TABLES xi

LIST OF FIGURES xii

1 INTRODUCTION 1

1.1 Introduction 1

1.2 Problem Background 3

1.2.1 Meta-Heuristic Algorithms 4

1.2.2 Harmony Search Algorithm 6

1.3 Problem Statement 6

1.4 The Study Aim 7

1.5 Dissertation Objectives 8

1.6 Dissertation Scope 8

1.7 Dissertation Significance 9

1.8 The Organization of the Dissertation 9

2 LITERATURE REVIEW 10

2.1 Introduction 10

2.2 Artificial Intelligence 11

Page 8: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

viii

2.3 Artificial Neural Network (ANN) 11

2.3.1 The structure of Neural Network 12

2.3.1.1 Biological Neuron Structure 12

2.3.1.2 Artificial Neuron Structure 13

2.3.2 Neural Network Architecture 16

2.3.3 Artificial Neural Networks and BP algorithm 18

2.4 Radial Basis Function Neural Network 22

2.4.1 RBF Neural Network Architecture 23

2.4.2 RBF Activation Functions 24

2.4.3 Comparison of Traditional Neural Network and RBF 25

2.4.4 Training RBF Neural Networks 26

2.4.5 Construction of the Hidden Layer of RBF Networks 27

2.5 Heuristic Algorithm 31

2.5.1 Evolutionary Algorithms (EA) 32

2.5.1.1. Genetic Algorithm 34

2.5.2 Swarm Intelligence and Particle Swarm Optimization 35

2.5.2.1 PSO Learning Process 36

2.5.2.2 Evaluation of PSO equations 38

2.6 Harmony Search Algorithm (HSA) 40

2.6.1 Standard Harmony Search Algorithm 43

2.6.2 The Improved Harmony Search Algorithm 45

2.6.3 The Global-Best Harmony Search 47

2.6.4 Harmony Search Characteristics 48

2.7 Comparative Study of HSA with RBF Training 49

2.8 Summary 51

3 METHODOLOGY 52

3.1 The Framework of the Study 52

3.2 Phase I – Dataset Preparation 54

3.2.1 XOR Dataset 54

3.2.2 IRIS Dataset 55

3.2.3 Cancer Dataset 55

3.2.4 Heart Diseases Dataset. 56

3.2.5 Data Normalization 58

Page 9: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

ix

3.3 Phase II – Algorithm Development 58

3.3.1 HSA Algorithm 59

3.3.2 Particle Swarm Optimization (PSO) Algorithm 62

3.4.1 Training RBF Neural Networks 66

3.4.1.1 Training RBFN Using HS Algorithm. 68

3.4.1.2 PSO-Based RBFN Training Algorithm 70

3.4 Phase III – Performance Measurement 72

3.5 Summary 73

4 RESULTS AND DISCUSSION 74

4.1 Experimental Studies 74

4.2 Experimental Setup of Harmony Search Algorithm 75

4.3 Implementation of Radial Basis Function (RBF) Alone 79

4.3.1 Results of Iris Dataset with RBF Alone 79

4.3.2 Results of Iris Dataset with RBF Alone 80

4.3.3 Results of Cancer Dataset with RBF Alone 81

4.3.4 Results of Heart Disease Dataset with RBF Alone 82

4.4 Training RBF with Harmony Search Algorithm (HSA) 83

4.4.1 Results of XOR Dataset 83

4.4.2 Results of IRIS Dataset 84

4.4.3 Results of Cancer Dataset 86

4.4.4 Results of Heart Disease Dataset 88

4.5 Comparison of HSA-RBF and PSO-RBF 89

4.6 Validation Results 91

4.6.1 Validation Results of IRIS Dataset 92

4.6.2 Validation Results of Cancer Dataset 93

4.6.3 Validation Results of Heart Disease Dataset 94

4.7 Statistical Test of HSA-RBF Classifier 95

4.7.1 Statistical Report HSA-RBF with Iris Dataset 95

4.7.2 Statistical Report HSA-RBF with Cancer Dataset 96

4.7.3 Statistical Report HSA-RBF with Heart Dataset 97

4.7.4 Sensitivity and Specificity 97

4.8 Receiver Operating Characteristics (ROC) 98

4.9 Discussion 101

Page 10: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

x

4.10 Summary 101

5 CONCLUSION AND FUTURE WORK 102

5.1 Introduction 102

5.2 Discussion 102

5.3 Summary of the Work. 103

5.3.1 Achievements 103

5.3.2 Contribution of the Study 104

5.4 Suggestions for Future Work 104

5.5 Conclusion 105

REFERENCES 106

APPENDIX A 115

1 Optimizing Weights to Output Layer of IRIS Dataset 115

2 Optimizing Weights to Output Layer of Cancer Dataset 116

3 Optimizing Weights to Output Layer of Heart Dataset 120

APPENDIX B 122

APPENDIX C 128

APPENDIX D 148

Page 11: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

xi

LIST OF TABLES

TABLE NO. TITLE PAGE

‎2.1 The optimization terms in the musical context 41

3.1 XOR dataset. 54

3.2 Attribute information of iris dataset 55

3.3 Attribute information of cancer dataset 56

3.4 Attribute information of heart disease dataset 57

‎4.1 HS algorithm parameter settings In the testing process 77

4.2 Benchmark functions, their search range and global min. 78

‎4.3 Global Min, Max, Mean and (SD) of 100 runs of HSA. 78

4.4 Global Min, Max, Mean and (SD) of 5000 runs of HSA. 79

4.5 Result of RBF on XOR dataset 70

4.6 Result of HSA-RBF on XOR dataset 84

4.7 Result of HSA-RBF on Iris Dataset 85

4.8 Result of HSA-RBF on Cancer Dataset 86

4.9 Result of HSA-RBF on Heart Disease Dataset 88

‎4.10 Comparative summary of HSA-RBF and PSO-RBF 90

4.11 Iris 10 Fold Cross-Validation on HSA-RBF 92

4.12 Breast cancer 10 Fold Cross-Validation on HSA-RBF 93

4.13 Heart disease 10 Fold Cross-Validation on HSA-RBF 94

4.14 Iris statistical test report 96

4.15 Breast cancer statistical test report 96

4.16 Heart disease statistical test report 97

4.17 Sensitivity and Specificity 98

Page 12: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

xii

LIST OF FIGURES

FIGURE NO. TITLE PAGE

‎ 2.1 The structure of biological nerve cell 13

‎ 2.2 The structure of artificial neuron (Krogh, 2008) 14

‎ 2.3 Activation Functions: Step function, Sigmoid Function 15

‎ 2.4 Topology of simple Feed-forward Neural Network 16

‎ 2.5 Topology of simple Recurrent NN 17

‎ 2.6 A (k * m * n) feed-forward ANN 18

‎ 2.7 BPN training process 21

‎ 2.8 The structure of RBF Neural Network 23

‎ 2.9 Different classification mechanisms 26

‎ 2.10 RBF - with N inputs, L hidden units and M outputs 28

‎ 2.11 Flow chart of an Evolutionary Algorithm (EA) 33

‎ 2.12 The flow chart of basic genetic algorithm iteration. 35

‎ 2.13 PSO learning process 38

‎ 2.14 Analogy between improvisation and optimization, as in 41

‎ 2.15 The optimization process of improved HS algorithm. 46

‎ 3.1 The framework of the study 53

‎ 3.2 The Basic HSA Procedure 61

‎ 3.3 The Basic PSO Procedure 63

‎ 3.4 The network architecture of HSA-RBF network 69

‎ 3.5 The PSO-RBF network architecture 71

‎ 4.1 Output of XOR trained with RBF neural network alone 80

‎ 4.2 Convergence of IRIS trained with HSA-RBF 85

‎ 4.3 Convergence of cancer dataset trained with HSA-RBF 87

‎ 4.4 Convergence of heart dataset trained with HSA-RBF 89

Page 13: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

xiii

‎ 4.5 Comparison of RBF-Alone, HSA-RBF and PSO-RBF 90

‎ 4.6 ROC curve for HSA-RBF on the breast cancer dataset 99

‎ 4.7 ROC curve for HSA-RBF on the heart disease dataset 100

Page 14: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

Chapter 1

INTRODUCTION

1.1 Introduction

Artificial neural network’s mimicking ability to human talent and their

similarity to the structure of the neurons of the human brains attracted the eyes of

many researchers due to its unparalleled properties, such as adaptability, learning and

generalization capability (Kulluk et al., 2012). Basically the principles of the

Artificial Neural Networks (ANNs) were first formulated by McCulloch and Pitts in

1943(Graupe, 2007). According to Chan et al. (1995), Neural Network have not only

the competence to learn a complex nonlinear dataset from massive body of given

attributes, but can tolerate to fault and noisy condition in resemblance to human brain

as well.

One of the outstanding examples of neural networks is Radial Basis Function

(RBF). According to Gan et al. (2012), RBF neural Network which was originally

conceived by Broomhead and Lowe in 1988 has characterized with fast training

speed, strong learning capability and simple topological architecture. Idri et al.

(2010) and Gan et al. (2011) described that it consists of only three different layers

e.g. the input layer which accepts source dataset; the hidden layer that uses radial

basis function to compute its output, and the output layer which represents the result

Page 15: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

2

of the network. According to Kurban and Beşdok (2009) an Fernández-Navarro et al.

(2011), the activation function which is implemented is usually the Gaussian

function, although in some situations (e.g. time series) other functional forms

including thin-plate spline functions, multi-quadratic functions and sigmoidal

functions are applied.

All neural networks are classified into two main categories of training

algorithm, namely: supervised neural network and unsupervised neural network.

Bors (2001) proved that RBF usually subclasses under supervised category. The

supervised category works with sample of datasets labeled with the training dataset.

This dataset is presented to the inputs at the beginning of the learning process

to determine the correct outputs. As in Kattan et al. (2010) an output value that is

close to the desired output could be achieved through an iterative continuous process

and adjustment of the network weights. Although back-propagation algorithms

became one of the most popular methods used to train ANNs, however , it has two

drawbacks: firstly differentiable transfer function is required and secondly possibility

of trapping into the local minima is too high. Many Stochastic Global Optimization

(SGO) techniques such as evolutionary algorithms are adopted for the training of

ANNs in order to overcome the local minimum problems.

Harmony Search Algorithm (HSA), a powerful music-based meta-heuristic

SGO algorithm, not inspired by biological and physical processes is also adopted for

the training of ANNs (Kulluk et al., 2012) and (Kattan and Abdullah, 2011). HSA

performed better than the standard BP algorithm as reported in (Kattan et al., 2010).

The following section depicts the background of this study.

Page 16: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

3

1.2 Problem Background

The emergence of radial bases function as an alternative of ANN was first

perceived late 80’s, although their related counterpart – pattern recognition technique

– existed long ago (Bors, 2001). RBF Neural Network was originally perceived and

added to the ANN by Broomhead and Lowe (1988), who were inspired by the local

response observation in the biologic neurons. RBF Networks have been

implemented in a wide area of engineering and science fields, because of their

advantages over other well known networks such as: their optimized ability, simple

topological architecture, accuracy in dynamically nonlinear approximation and fast

and easy learning algorithms (Gan et al., 2012).

Bors (2001) mentioned in his paper that radial basis functions are entrenched

in two feed-forward neural network layers. In addition to this two visible layers, i.e.

the input layer and the output layer, there is a third hidden layer embedded in

between them for processing units called hidden units, in which RBF which is

generally a Gaussian function, is applied to each of them. Qasem and Shamsuddin

(2011) proved that the output layer of RBF has the characteristics of linear decision

boundary, where as the hidden units of this network are indeed a composition of non-

linear mapping.

According to Kurban et al. (2009), in various literatures, different algorithms

were proposed for training the RBF Network. It is necessary to find appropriate

training algorithms for the RBF Neural Network.

One of the most popular training algorithms in the domain of RBF Neural

Networks is the back-propagation technique, which is a gradient-descent method to

minimize the mean squared error between the desired outputs and the actual outputs

for the particular inputs to the networks. However, as in Kulluk et al. (2012), BP has

some drawbacks: the first is that it require a differentiable neuron transfer function

and the second is the high possibility to converging into local minima.

Page 17: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

4

To deal with this convergence problems, some researchers proposed two

derivative based algorithms for training RBF networks, such as the gradient descent

(GD) algorithm and Kalman filtering (KF). Kurban and Beşdok (2009) and Tuba et

al. (2009) proved that both algorithms need a prolonged time and have convergence

weaknesses to the local minima and procedure of discovering the optimal gradient. In

order to overcome this drawbacks, several global optimization methods could be

applied for training RBF networks in accordance with the various science and

engineering problems. Some of these algorithms are: genetic algorithms (GA), the

particle swarm optimization (PSO) algorithm, the artificial immune system (AIS)

algorithm and the differential evolution (DE) algorithm.

These meta-heuristic SGO techniques are inspired by biological processes

which has the characteristics of training algorithms that overcome the

aforementioned inefficiencies. Besides that, Harmony Search (HS) algorithms are

young meta-heuristic SGO methods which resemble the other SGO meta-heuristic

techniques except that they are inspired by music improvisation. Although HS have

been reported that they performed better than BP in adopting Feed Forward Neural

Networks (FFNN) as in (Kattan et al., 2010), they were not applied to RBFNNs so

far. The following section will illustrate more on meta-heuristic algorithm.

1.2.1 Meta-Heuristic Algorithms

In order to deal with the local minimum problem, many global optimization

techniques have been adopted for the training of RBF Neural Networks in this case.

Heuristic algorithms typically intend to find a good solution to an optimization

problem by ‘trial-and-error’ in a reasonable amount of computing time. Here

‘heuristic’ means to ‘find’ or ‘search’ by trials and errors. Generally, local search

methods are heuristic methods because their parameter search is focused on the local

variations, and the optimal or best solution can be well outside this local region.

However, a high-quality feasible solution in the local region of interest is usually

Page 18: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

5

accepted as a good solution in many optimization problems in practice if time is the

major constraint.

Meta-heuristic algorithms are higher-level heuristic algorithms. The word

‘meta-’ stands for ‘higher-level’ or ‘beyond’, so a literal meaning of meta-heuristic is

to find the solution through high level techniques, although certain trial-and-error

processes are still used. Broadly speaking, meta-heuristics are considered as higher-

level techniques or strategies which intend to combine lower-level techniques and

tactics for exploration and exploitation of the huge space for parameter search (Yang,

2009). Yang (2009) says that the word ‘meta-heuristic’ refers to modern high level

algorithms including Simulated Annealing (SA), Particle Swam Optimization (PSO),

Evolutionary Algorithms such as Genetic Algorithm (GA), and, certainly Harmony

Search Algorithm (HSA).

However, Ren et al. (2010) mentioned that, the Genetic Algorithm usually

spends a long time to find a solution. At the same time there may be premature and

slow convergence problems in GA. On the other hand, according to Dian et al.

(2011), PSO easily suffers from the partial optimism, which is related to the

regulation of its speed and direction. Moreover Grosan and Abraham (2011)

summarized the pitfalls of PSO in their book of “Intelligent Systems: A Modern

Approach”, as follows:

Particles tend to cluster, i.e., converge too fast and get stuck at local

optimum

Movement of particle carried it into infeasible region

Inappropriate mapping of particle space into solution space

These drawbacks motivated the proposal of HS which is a new meta-heuristic

algorithm to solve those aforesaid problems. A brief discussion about this algorithm

will come next.

Page 19: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

6

1.2.2 Harmony Search Algorithm

According to Kulluk et al. (2012) Harmony Search Algorithm (HSA) is a

meta-heuristic optimization algorithm motivated from the process of making music.

In HS algorithm each decision variable (musical instrument) generates a value (note)

in order to find the global optimum solution (best harmony). The method uses a

stochastic random search based on harmony memory consideration rate and pitch

adjustment rate instead of a gradient search. Nowadays HS algorithm has been

applied to many diverse optimization problems such as music composition, Sudoku

puzzle, timetabling, tour planning, logistics, web page clustering, text

summarization, Internet routing, robotics, power system design, structural design,

vehicle routing, heat exchanger design and so on.

As far to our knowledge, no study has been done for optimization of RBF

Neural Network with HSA. This attracted our attention in training RBF with HSA to

improve the performance of the network.

1.3 Problem Statement

According to kulluk et al. (2011); Kattan et al. (2010) and Hamed et al.

(2012) training RBF with BP faced some problems such as poor convergence and

trapping at the local minima. Genetic algorithm performed robust training without

suffering from local minimum problem. However its output production is time

consuming and computation expensive (Xie et al., 2011) and (Hamadneh et al.,

2012).

Particle swarm optimization attracted the attention of many researchers after

several experiments proved its better performance over GA. Although the

Page 20: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

7

experiments conducted by many researches showed a plausible achievement, Rini et

al. (2011) mentioned that PSO easily suffers from the partial optimism, which is

related to the regulation of its speed and direction. Moreover Grosan and Abraham

(2011) summarized the pitfalls of PSO in their book of “Intelligent Systems: A

Modern Approach”, with the following three problems: firstly, particles tend to

cluster, i.e., converge too fast and get stuck at local optimum. Secondly, movement

of particle carried it into infeasible region and finally, inappropriate mapping of

particle space into solution space.

Kulluk et al. (2011) proved that HS which is a new SGO meta-heuristic

algorithm is a good candidate and the most promising variant for training feed

forward type NNs. More over Soltani et al. (2011) confirmed that HS is not only

faster than PSO but has a significant convergence rate to reach the optimal solution.

So far no study related to the optimization of RBF Neural Network with HSA

has been done, therefore two questions that can be perceived from this research are

stated as below:

1. Since harmony search proved better performance compared to other

optimizing algorithms, can HS algorithm improve the learning capability of

RBF network?

2. How much significance can HSA provide in optimizing the RBF neural

network?

1.4 The Study Aim

In this study HSA will be employed to investigate the higher convergence

rate and the classification performance of RBF neural network’s learning capability

compared with PSO based RBF neural network

Page 21: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

8

1.5 Dissertation Objectives

The objectives of the study are as follows:

i. To identify existing literature of Radial Basis Function (RBF) and

Particle Swarm Optimization (PSO) algorithms and their interaction.

ii. To enhance the training process of RBF neural network by integrating

with the Harmony Search Algorithm (HSA).

iii. To compare the results between HSA-RBFN and PSO-RBFN in terms of

convergence rate and classification result.

1.6 Dissertation Scope

This study will be confined in the following points in order to fulfill the

aforementioned goals.

i. Four datasets which are XOR, Iris, Cancer and Heart disease

classification from UCI machine learning dataset have been used for

training and testing.

ii. The performance of HSA learning algorithm for RBF Network are

compared to PSO algorithm only.

iii. While enhancing RBF Network learning by integrating HSA error

function are minimization.

Page 22: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

9

1.7 Dissertation Significance

According to the process of work, there will analysis and continuous

experiments that will be targeted to the goals in order to improve the learning

capabilities of RBF and make it cost effective by integrating with it this superior

HSA which previous researches approved that it has great practical significance in

searching for optimal value of large-scale projects problems (Ren and Kezunovic,

2010)(Ren et al., 2010).

Performance metrics: In order to analyze and compare the training

capability of the HSA algorithm, four performance metrics will be taken into

consideration. These are: overall training time, sum of squared errors, training

accuracy and testing accuracy. Accuracy measures the ability of the classifier to

produce accurate results.

1.8 The Organization of the Dissertation

This dissertation consists of five chapters. Chapter 1 presents the introduction

of the dissertation, problems background, the problem statement, objectives, scope

and significance of the study. In Chapter 2, the literature reviews on ANN, BP, RBF,

PSO and HSA is discussed. Chapter 3 illustrates research methodology. Chapter 4

displays the experimental results, finally the conclusion and suggestions for future

work are explained in Chapter 5.

Page 23: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

106

REFERENCES

ABDULL HAMED, H. N., SHAMSUDDIN, S. M. & SALIM, N. 2012. Particle

Swarm Optimization For Neural Network Learning Enhancement. Jurnal

Teknologi, 49, 13–26.

ABDULLHAMED, H. N., SHAMSUDDIN, S. M. & SALIM, N. 2008. PARTICLE

SWARM OPTIMIZATION FOR NEURAL NETWORK LEARNING

ENHANCEMENT. Jurnal Teknologi, 49(D) Dis., p13–26.

ABRAHAM, A. 2005. Artificial neural networks. In: THORN, E. B. P. H. S. A. R.

(ed.) Handbook of Measuring System Design. John Wiley & Sons, Ltd.

ABRAHAM, A. 2005. Evolutionary Computation. In: THORN, P. H. S. A. R. (ed.)

Handbook of Measuring System Design. John Wiley & Sons, Ltd.

AL-BETAR, M. A. & KHADER, A. T. 2012. A harmony search algorithm for

university course timetabling. Springer Science+Business Media, LLC, 194,

p3-31.

AL-BETAR, M. A., KHADER, A. T. & ZAMAN, A. M. 2012 University Course

Timetabling Using a Hybrid Harmony Search Metaheuristic Algorithm. IEEE

TRANSACTIONS ON SYSTEMS, MAN, AND CYBERNETICS—PART

C: APPLICATIONS AND REVIEWS, VOL. 42, p664-681.

ALEJO, R., GARCÍA, V., SOTOCA, J., MOLLINEDA, R. & SÁNCHEZ, J. 2007.

Improving the performance of the RBF neural networks trained with

imbalanced samples. Computational and Ambient Intelligence, 162-169.

ALIA, O. M. & MANDAVA, R. 2011. The variants of the harmony search

algorithm: an overview. Artificial Intelligence Review, 36, 49-68.

ALIA, O. M. D. & MANDAVA, R. 2011. The variants of the harmony search

algorithm: an overview. Springer Science+Business Media B.V., DOI

10.1007/s10462-010-9201-y, 1-20.

Page 24: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

107

APHIRAK KHADWILARD, AND, P. L. & PONGCHAROEN, P. 2012. Full

factorial experimental design for parameters selection of Harmony Search

Algorithm. The Journal of Industrial Technology, Vol. 8, .

AYVAZ, M. T. 2007. Simultaneous determination of aquifer parameters and zone

structures with fuzzy c-means clustering and meta-heuristic harmony search

algorithm. Advances in Water Resources, 30, 2326-2338.

BAHOURA, M. & PARK, C. W. Year. FPGA-implementation of high-speed MLP

neural network. In: Electronics, Circuits and Systems (ICECS), 2011 18th

IEEE International Conference on, 2011. IEEE, 426-429.

BEG, A., CHANDANA PRASAD, P. & BEG, A. 2008. Applicability of feed-

forward and recurrent neural networks to Boolean function complexity

modeling. Expert Systems With Applications, 34, 2436-2443.

BERRY, E. A., DALBY, A. R. & YANG, Z. R. 2004. Reduced bio basis function

neural network for identification of protein phosphorylation sites: comparison

with pattern recognition algorithms. Computational biology and chemistry,

28, 75-85.

BODÉN, M. 2001. A guide to recurrent neural networks and backpropagation. The

DALLAS project. Report from the NUTEK-supported project AIS-8, SICS.

Holst: Application of data analysis with learning systems, 1-10.

BORS, A. G. Year. Introduction of the radial basis function (rbf) networks. In:

Online symposium for electronics engineers, 2001. 1-7.

BOUCKAERT, R. R. 2006. Efficient AUC learning curve calculation. AI 2006:

Advances in Artificial Intelligence. Springer.

BRADLEY, A. P. 1997. The use of the area under the ROC curve in the evaluation

of machine learning algorithms. Pattern recognition, 30, 1145-1159.

CEYLAN, H., HALDENBILEN, S. & BASKAN, O. 2008. Transport energy

modeling with meta-heuristic harmony search algorithm, an application to

Turkey. Energy Policy, 36, 2527-2535.

CHAKRABORTY, P., ROY, G. G., DAS, S., JAIN, D. & ABRAHAM, A. 2009. An

Improved Harmony Search Algorithm with Differential Mutation Operator.

Fundamenta Informaticae 95 p1-26.

CHAKRABORTY, R.C. (2010). Fundamental of Neural Network AI course lecture,

37-38

Page 25: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

108

CHAN, W., CHOW, Y. & LIU, L. 1995. Neural network: An alternative to pile

driving formulas. Computers and geotechnics, 17, 135-156.

CHEN, B., QIU, Y., FENG, Y., TAVNER, P. & SONG, W. Year. Wind turbine

SCADA alarm pattern recognition. In: Renewable Power Generation (RPG

2011), IET Conference on, 2011. IET, 1-6.

CHENG, Y., LI, L., LANSIVAARA, T., CHI, S. & SUN, Y. 2008. An improved

harmony search minimization algorithm using different slip surface

generation methods for slope stability analysis. Engineering Optimization, 40,

95-115.

DAS, S. R., PANIGRAHI, P. K., DAS, K. & MISHRA, D. 2012. Improving RBF

Kernel Function of Support Vector Machine using Particle Swarm

Optimization. International Journal, 2.

DHUBKARYA, D., NAGARIA, D. & KAPOOR, R. 2010. Implementation of a

Radial Basis Function Using VHDL. Global Journal of Computer Science

and Technology, 10.

DIAN, P. R., SITI, M. S. & SITI, S. Y. 2011. Particle Swarm Optimization:

Technique, System and Challenges. International Journal of Computer

Applications, 14, 19-27.

DIAO, R. & SHEN, Q. Year. Two new approaches to feature selection with harmony

search. In: Fuzzy Systems (FUZZ), 2010 IEEE International Conference on,

2010. IEEE, 1-7.

EBERHART, R. C. & SHI, Y. 2001. Particle Swarm Oprimization:Developments,

Applications and Resources. IEEE, p81-86.

EL CHOUBASSI, M., EL KHOURY, H., ALAGHA, C. E. J., SKAF, J. & AL-

ALAOUI, M. Year. Arabic speech recognition using recurrent neural

networks. In: Signal Processing and Information Technology, 2003. ISSPIT

2003. Proceedings of the 3rd IEEE International Symposium on, 2003. IEEE,

543-547.

ENGELBRECHT, A. P. 2007. Computational intelligence: an introduction,

Chichester,

FAWCETT, T. 2006. An introduction to ROC analysis. Pattern recognition letters,

27, 861-874.

FERNÁNDEZ-NAVARRO, F., HERVÁS-MARTÍNEZ, C., SANCHEZ-

MONEDERO, J. & GUTIÉRREZ, P. A. 2011. MELM-GRBF: a modified

Page 26: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

109

version of the extreme learning machine for generalized radial basis function

neural networks. Neurocomputing, 74, 2502-2510.

FESANGHARY, M., MAHDAVI, M., MINARY-JOLANDAN, M. & ALIZADEH,

Y. 2008. Hybridizing harmony search algorithm with sequential quadratic

programming for engineering optimization problems. Computer methods in

applied mechanics and engineering, 197, 3080-3091.

FORSATI, R., MAHDAVI, M., KANGAVARI, M. & SAFARKHANI, B. Year.

Web page clustering using harmony search optimization. In: Electrical and

Computer Engineering, 2008. CCECE 2008. Canadian Conference on, 2008.

IEEE, 001601-001604.

GALLEGOS-FUNES, F. J., GÓMEZ-MAYORGA, M. E., LOPEZ-BONILLA, J. L.

& CRUZ-SANTIAGO, R. 2009. Rank M-type Radial Basis function

(RMRBF) Neural network for Pap smear microscopic Image classification.

Apeiron, 16, 542.

GAN, M., PENG, H. & CHEN, L. 2012. A hybrid global–local optimization

approach to parameter estimation of RBF-type models. Information Sciences.

GAN, M., PENG, H. & DONG, X. 2011. A hybrid algorithm to optimize RBF

network architecture and parameters for nonlinear time series modeling.

Applied Mathematical Modelling.

GEEM, Z. & CHOI, J. Y. 2007. Music composition using harmony search algorithm.

Applications of Evolutionary Computing, 593-600.

GEEM, Z. W. & HWANGBO, H. Year. Application of harmony search to multi-

objective optimization for satellite heat pipe design. In: Proceedings of,

2006. Citeseer, 1-3.

GEEM, Z. W. & WILLIAMS, J. C. Year. Ecological optimization using harmony

search. In: Proceedings of the American Conference on Applied

Mathematics, 2008. World Scientific and Engineering Academy and Society

(WSEAS), 148-152.

GEEM, Z. W. 2008a. Harmony Search Applications in Industry. Springer-Verlag

Berlin Heidelberg 226, p117-134.

GEEM, Z. W. 2008b. Novel derivative of harmony search algorithm for discrete

design variables. Applied Mathematics and Computation, 199 p223-230.

GEEM, Z. W. 2009. Particle-swarm harmony search for water network design.

Engineering Optimization, Vol. 41, 297–311.

Page 27: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

110

GEEM, Z. W. 2010. State-of-the-Art in the Structure of Harmony Search Algorithm.

Springer-Verlag Berlin Heidelberg, SCI 270, 1-10.

GEEM, Z. W., KIM, J. H. & LOGANATHAN, G. V. 2001. A New Heuristic

Optimization Algorithm: Harmony Search. Simulation Councils Inc., 76, p60-

68.

GEEM, Z. Year. Harmony search algorithm for solving sudoku. In: Knowledge-

Based Intelligent Information and Engineering Systems, 2007. Springer, 371-

378.

GRAUPE, D. 2007. Principles of artificial neural networks, SINGAPORE., World

Scientific Publishing Company Incorporated.

GROSAN, C. & ABRAHAM, A. 2011. Intelligent Systems: A Modern Approach,

Springer.

HAMADNEH, N., SATHASIVAM, S., TILAHUN, S. L. & CHOON, O. H. 2012.

Learning Logic Programming in Radial Basis Function Network via Genetic

Algorithm. Journal of Applied Sciences, 12, 840-847.

HASAN, S. 2010. Enhanced self organizing map with particle swarm optimization

for classification problems. Universiti Teknologi Malaysia, Faculty of

Computer Science and Information System.

HASSAN, N. & AKAMATSU, N. 2004. A new approach for contrast enhancement

using sigmoid function. The International Arab Journal of Information

Technology, vol. 1, 221-226.

IDRI, A., ZAKRANI, A. & ZAHI, A. 2010. Design of radial basis function neural

networks for software effort estimation. IJCSI International Journal of

Computer Science Issues, 7.

KATTAN, A., ABDULLAH, R. & SALAM, R. A. Year. Harmony search based

supervised training of artificial neural networks. In: Intelligent Systems,

Modelling and Simulation (ISMS), 2010 International Conference on, 2010.

IEEE, 105-110.

KHADWILARD, A., LUANGPAIBOON, P. & PONGCHAROEN, P. 2012. Full

factorial experimental design for parameters selection of Harmony Search

Algorithm. The Journal of Industrial Technology, Vol. 8.

KONAKA, A., COITB, D. W. & SMITH, A. E. 2006. Multi-objective optimization

using genetic algorithms: A tutorial. Reliability Engineering and System

Safety, 91, p992-1007.

Page 28: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

111

KRIESEL, D. 2007. A brief introduction to neural networks. Retrieved August, 15,

2011.

KROGH, A. 2008. What are artificial neural networks? Nature biotechnology, 26,

195-197.

KULLUK, S., OZBAKIR, L. & BAYKASOGLU, A. 2012. Training neural networks

with harmony search algorithms for classification problems. Engineering

Applications of Artificial Intelligence, 25, 11-19.

KUMAR, M. & YADAV, N. 2011. Multilayer perceptrons and radial basis function

neural network methods for the solution of differential equations: A survey.

Computers & Mathematics with Applications.

KUOK, K. K., AND, S. H. & SHAMSUDDIN, S. M. 2010. Particle Swarm

Optimization Feed forward Neural Network for Hourly Rainfallrunoff

Modeling in Bedup Basin, Malaysia. International Journal of Civil &

Environmental Engineering IJCEE, Vol: 9 p20-39.

KURBAN, T. & BEŞDOK, E. 2009. A comparison of RBF neural network training

algorithms for inertial sensor based terrain classification. Sensors, 9, 6312-

6329.

LEANDRO DOS SANTOS COELHO & BERNERT, D. L. D. A. 2009. An

improved harmony search algorithm for synchronization of discrete-time

chaotic systems. Chaos, Solitons and Fractals 41, p2526–2532.

LI, L. J., HUANG, Z. B., LIU, F. & WU, Q. H. 2007. A heuristic particle swarm

optimizer for optimization of pin connected structures. Computers and

Structures, 85, p340-349.

LI, L., HUANG, Z., LIU, F. & WU, Q. 2007. A heuristic particle swarm optimizer

for optimization of pin connected structures. Computers & structures, 85,

340-349.

LUCKS, M. B. & OKI, N. Year. Radial basis function networks with quantized

parameters. In: Computational Intelligence for Measurement Systems and

Applications, 2008. CIMSA 2008. 2008 IEEE International Conference on,

2008. IEEE, 23-27.

MAHDAVI, M., FESANGHARY, M. & B, E. D. 2007. An improved harmony

search algorithm for solving optimization problems. Applied Mathematics

and Computation, 188 p1567-1579.

Page 29: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

112

MOHSEN, A., KHADER, A. & RAMACHANDRAM, D. 2010. An optimization

algorithm based on harmony search for rna secondary structure prediction.

Recent Advances In Harmony Search Algorithm, 163-174.

NAGESH KUMAR, D., SRINIVASA RAJU, K. & SATHISH, T. 2004. River flow

forecasting using recurrent neural networks. Water resources management,

18, 143-161.

NOMAN, S., SHAMSUDDIN, S. & HASSANIEN, A. 2009. Hybrid learning

enhancement of RBF network with particle swarm optimization. Foundations

of Computational, Intelligence Volume 1, 381-397.

OMRAN, M. G. H. & MAHDAVI, M. 2008. Global-best harmony search. Applied

Mathematics and Computation, 198 p643-656.

PAN, Q. K., SUGANTHAN, P., TASGETIREN, M. F. & LIANG, J. 2010. A self-

adaptive global best harmony search algorithm for continuous optimization

problems. Applied Mathematics and Computation, 216, 830-848.

PANCHAL, A. 2009. Harmony search in therapeutic medical physics. Music-

inspired harmony search algorithm, 189-203.

QASEM, S. N. & SHAMSUDDIN, S. M. 2011. Radial basis function network based

on time variant multi-objective particle swarm optimization for medical

diseases diagnosis. Applied Soft Computing, 11, 1427-1438.

QUAN-KE PAN, P.N. SUGANTHAN, M. FATIH TASGETIREN & LIANG, J. J.

2010. A self-adaptive global best harmony search algorithm for continuous

optimization problems. Applied Mathematics and Computation, 216, p830–

848.

REN, J. & KEZUNOVIC, M. Year. Use of recursive wavelet transform for

estimating power system frequency and phasors. In: Transmission and

Distribution Conference and Exposition, 2010 IEEE PES, 2010. IEEE, 1-6.

RINI, D. P., SHAMSUDDIN, S. M. & YUHANIZ, S. S. 2011. Particle Swarm

Optimization: Technique, System and Challenges. International Journal of

Computer Applications (0975 – 8887), Vol 14 p19-27.

SANTOS COELHO, L. & DE ANDRADE BERNERT, D. L. 2009. An improved

harmony search algorithm for synchronization of discrete-time chaotic

systems. Chaos, Solitons & Fractals, 41, 2526-2532.

SCHWENKER, F., KESTLER, H. A. & PALM, G. 2001. Three learning phases for

radial-basis-function networks. Neural networks, 14, 439-458.

Page 30: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

113

SHOPOVA, E. G. & VAKLIEVA-BANCHEVA, N. G. 2006. BASIC—A genetic

algorithm for engineering problems solution. Computers & chemical

engineering, 30, 1293-1309.

SHOPOVA, E. G. & VAKLIEVA-BANCHEVA, N. G. BASIC—A genetic

algorithm for engineering problems solution. Computers and Chemical

Engineering, Volume 30, p1-17.

SIDDIQI, A. A. 2012. Implications of using Artificial Intelligence Technology in

Modern Warfare. ICCIT.

SOLTANI, M. A. Z., HAGHIGHAT, A. T., RASHIDI, H. & CHEGINI, A. T. G.

2011. A Couple of Algorithms for K-Coverage Problem in Visual Sensor

Networks. International Conference on Communication Engineering and

Networks. Singapore: IACSIT Press.

SOLTANI, M. A. Z., HAGHIGHAT, A. T., RASHIDI, H. & CHEGINI, T. G. 2011.

A Couple of Algorithms for K-Coverage Problem in Visual Sensor Networks.

In: IPCSIT (ed.) International Conference on Communication Engineering

and Networks. Singapore: IACSIT Press.

Sultan Noman, Qasem Mohammed (2008) Learning enhancement of radial basis

function network with particle swarm optimization. Masters thesis, Universiti

Teknologi Malaysia, Faculty of Computer Science and Information System.

TALBI, E. G. 2009. METAHEURISTICS: FROM DESIGN TO IMPLEMENTATION,

Hoboken, New Jersey, Wiley Online Library.

TAN, K., YU, Q. & ANG, J. 2006. A coevolutionary algorithm for rules discovery in

data mining. International journal of systems science, 37, 835-864.

TSIMBOUKAKIS, N. & TAMBOURATZIS, G. Year. Neural Networks for Author

Attribution. In: Fuzzy Systems Conference, 2007. FUZZ-IEEE 2007. IEEE

International, 2007. IEEE, 1-6.

UMARANI, D. R. & SELVI, V. 2010. PARTICLE SWARM OPTIMIZATION

EVOLUTION, OVERVIEW AND APPLICATIONS. International Journal of

Engineering Science and Technology, Vol. 2 p2802-2806.

VALLE, Y. D., VENAYAGAMOORTHY, G. K., MOHAGHEGHI, S.,

HERNANDEZ, J.-C. & HARLEY, R. G. 2008. Particle Swarm Optimization:

Basic Concepts, Variants and Applications in Power Systems. IEEE

TRANSACTIONS ON EVOLUTIONARY COMPUTATION, VOL. 12,

p171-195.

Page 31: LEARNING ENHANCEMENT OF RADIAL BASIS FUNCTION …eprints.utm.my/id/eprint/36531/5/MohamedHassanAhmedMFSKSM2013.pdf · Heuristik dan Latihan Asas Fungsi Jejari serta gabungan serpihan

114

VASEBI, A., FESANGHARY, M. & BATHAEE, S. 2007. Combined heat and

power economic dispatch by harmony search algorithm. International

Journal of Electrical Power & Energy Systems, 29, 713-719.

WANG, J. Y. 2009. Data mining analysis(breast-cancer data). AI term project.

WANG, W. Year. Face Recognition Based On Radial Basis Function Neural

Networks. In: Future Information Technology and Management Engineering,

2008. FITME'08. International Seminar on, 2008. IEEE, 41-44. West Sussex,

John Wiley & Sons Ltd.

XIE, T., YU, H. & WILAMOWSKI, B. Year. Comparison between traditional neural

networks and radial basis function networks. In: Industrial Electronics

(ISIE), 2011 IEEE International Symposium on, 2011. IEEE, 1194-1199.

XU, S. & CHEN, L. Year. A novel approach for determining the optimal number of

hidden layer neurons for FNN’s and its application in data mining. In:

International Conference on Information Technology and Applications:

iCITA, 2008. 683-686.

YANG, X. S. 2009. Harmony search as a metaheuristic algorithm. Music-inspired

harmony search algorithm, 1-14.

YANG, X.-S. (2009). Harmony Search as a Metaheuristic Algorithm. In: (ED.), Z.

W. G. (ed.) Music-Inspired Harmony Search Algorithm. Springer-Verlag

Berlin Heidelberg.

YU, H., XIE, T., HAMILTON, M. & WILAMOWSKI, B. Year. Comparison of

different neural network architectures for digit image recognition. In: Human

System Interactions (HSI), 2011 4th International Conference on, 2011.

IEEE, 98-103.


Recommended