+ All Categories
Home > Documents > [American Institute of Aeronautics and Astronautics 13th AIAA/ISSMO Multidisciplinary Analysis...

[American Institute of Aeronautics and Astronautics 13th AIAA/ISSMO Multidisciplinary Analysis...

Date post: 14-Dec-2016
Category:
Upload: eliot
View: 212 times
Download: 0 times
Share this document with a friend
8
Visual Design Space Exploration using Contextual Self-Organizing Maps Brett Nekolny 1 , Trevor Richardson 2 , and Eliot Winer 3 Iowa State University, Ames, IA, 50011, USA Self-organizing maps (SOMs) and contextual maps are methods of visualizing high dimensional data in a low dimensional space. SOMs have previously been applied to visualize characteristics of optimization problems by generating maps of the component variables to compare interactions and relationships between design variables. In this paper, SOMs and contextual maps are explored as a visualization method to directly visualize the design space. Using the techniques described in the paper, high dimensional datasets are reduced to a 2D, human readable, visual map. Preliminary results show that the topology of three optimization functions using varying dimensionality can be clustered and visualized using contextual maps, and information can be gathered from these clusters including objective function values and variability amongst differing areas of the design space. This paper focuses on the use of contextual maps to extract valuable information such as modality and curvature to aid in future work such as selection of appropriate optimization algorithm and initial point for a solution run. I.Introduction isualization of complex design and optimization problems remains a difficult problem due to the inability to efficiently see data in greater than three dimensions. Accordingly, as optimization problems increase in dimensionality, the interplay and relationships between variables become harder to discern. In order to comprehend these problematic relationships, it is necessary to extract the variables and view them independently in one, two, or three dimensional (1D, 2D, 3D) spaces. Several research projects 1,2,3,4 examined various methods to visualize design spaces and other characteristics of optimization problems. For example, Eddy and Lewis created an environment to view the optimization process of multidimensional design, but require that the user either view the progress of the performance objectives, or view design variables in three or less dimensions. Two other alternatives to extracting variables to view them graphically are either statistical comparison or dimensionality reduction. Dimensionality reduction describes an attempt to display higher dimensional data by means of low dimensional output, which can be done by discarding variables and showing only those that describe the strongest correlations, or convert a high dimensional space to a low dimensional space. Kohonen’s Self-Organizing Map 5 (SOM) is a class of artificial neural networks with the ability to reduce dimensionality to a low dimensional space and furthermore visualize the resulting network. This type of neural network can provide the ability to train and then inspect a given dataset regardless of dimensionality. A. Self-Organizing Map The self-organizing map algorithm utilizes competitive learning strategies to train its (typically) two dimensional network of neurons. This two dimensional network provides connections between neurons (four immediate neighbors per neuron in the center) and forms a lattice structure that can be visualized in a perceivable 2D space as shown in Figure 1. Each neuron has an associated weight vector, as shown in Equation (1), with a dimensionality equivalent to the input space or data, as shown in Equation (2). w j = <w j1 , w j2 , ..., w jk > (1) x = <x 1 , x 2, ..., x k > (2) 1 Research Assistant, Department of Mechanical Engineering, Human Computer Interaction, Virtual Reality Applications Center, 1620 Howe Hall, Iowa State University, Ames, IA, 50011, USA, Student Member. 2 Research Assistant, Department of Mechanical Engineering, Virtual Reality Applications Center, 1620 Howe Hall, Iowa State University, Ames, IA, 50011, USA, Student Member. 3 Associate Professor, Department of Mechanical Engineering, Human Computer Interaction, Virtual Reality Applications Center, 1620 Howe Hall, Iowa State University, Ames, IA, 50011, USA, Senior Member. V 13th AIAA/ISSMO Multidisciplinary Analysis Optimization Conference 13 - 15 September 2010, Fort Worth, Texas AIAA 2010-9326 Copyright © 2010 by Eliot Winer. Published by the American Institute of Aeronautics and Astronautics, Inc., with permission.
Transcript

Visual Design Space Exploration using

Contextual Self-Organizing Maps

Brett Nekolny1, Trevor Richardson

2, and Eliot Winer

3

Iowa State University, Ames, IA, 50011, USA

Self-organizing maps (SOMs) and contextual maps are methods of visualizing high dimensional

data in a low dimensional space. SOMs have previously been applied to visualize characteristics of

optimization problems by generating maps of the component variables to compare interactions and

relationships between design variables. In this paper, SOMs and contextual maps are explored as a

visualization method to directly visualize the design space. Using the techniques described in the

paper, high dimensional datasets are reduced to a 2D, human readable, visual map. Preliminary

results show that the topology of three optimization functions using varying dimensionality can be

clustered and visualized using contextual maps, and information can be gathered from these

clusters including objective function values and variability amongst differing areas of the design

space. This paper focuses on the use of contextual maps to extract valuable information such as

modality and curvature to aid in future work such as selection of appropriate optimization

algorithm and initial point for a solution run.

I.Introduction

isualization of complex design and optimization problems remains a difficult problem due to the inability to

efficiently see data in greater than three dimensions. Accordingly, as optimization problems increase in

dimensionality, the interplay and relationships between variables become harder to discern. In order to comprehend these

problematic relationships, it is necessary to extract the variables and view them independently in one, two, or three

dimensional (1D, 2D, 3D) spaces. Several research projects1,2,3,4

examined various methods to visualize design spaces and

other characteristics of optimization problems. For example, Eddy and Lewis created an environment to view the

optimization process of multidimensional design, but require that the user either view the progress of the performance

objectives, or view design variables in three or less dimensions. Two other alternatives to extracting variables to view them

graphically are either statistical comparison or dimensionality reduction. Dimensionality reduction describes an attempt to

display higher dimensional data by means of low dimensional output, which can be done by discarding variables and

showing only those that describe the strongest correlations, or convert a high dimensional space to a low dimensional space.

Kohonen’s Self-Organizing Map5 (SOM) is a class of artificial neural networks with the ability to reduce dimensionality to a

low dimensional space and furthermore visualize the resulting network. This type of neural network can provide the ability

to train and then inspect a given dataset regardless of dimensionality.

A. Self-Organizing Map

The self-organizing map algorithm utilizes competitive learning strategies to train its (typically) two dimensional

network of neurons. This two dimensional network provides connections between neurons (four immediate neighbors per

neuron in the center) and forms a lattice structure that can be visualized in a perceivable 2D space as shown in Figure 1.

Each neuron has an associated weight vector, as shown in Equation (1), with a dimensionality equivalent to the input

space or data, as shown in Equation (2).

wj = <wj1, wj2, ..., wjk> (1)

x = <x1, x2, ..., xk> (2)

1 Research Assistant, Department of Mechanical Engineering, Human Computer Interaction, Virtual Reality Applications Center, 1620

Howe Hall, Iowa State University, Ames, IA, 50011, USA, Student Member. 2 Research Assistant, Department of Mechanical Engineering, Virtual Reality Applications Center, 1620 Howe Hall, Iowa State

University, Ames, IA, 50011, USA, Student Member. 3 Associate Professor, Department of Mechanical Engineering, Human Computer Interaction, Virtual Reality Applications Center, 1620

Howe Hall, Iowa State University, Ames, IA, 50011, USA, Senior Member.

V

13th AIAA/ISSMO Multidisciplinary Analysis Optimization Conference13 - 15 September 2010, Fort Worth, Texas

AIAA 2010-9326

Copyright © 2010 by Eliot Winer. Published by the American Institute of Aeronautics and Astronautics, Inc., with permission.

Figure 1: Diagram of a Self-Organizing Map

6

Equation (1) represents a sample weight vector for one neuron in the lattice, containing a variable number of weights where

k is equal to the dimensionality of the inputs. Equation (2) represents a sample input vector taken from the input space.

The training algorithm for SOMs is a winner-takes-all strategy coupled with a neighborhood influence surrounding the

winning neuron. The winning neuron is discovered by finding the minimum Euclidean distance between the input vector

and each neuron in the map. This method of calculating a winning neuron, and by that a neighborhood of influence, aids in

classifying each new input vector. After a winner is found, a time-varying neighborhood surrounding the winner is updated

using the node update shown in Equation (3).

wj(n+1) = wj(n) + η(n) · hj,i(x)(n) · ( x – wj(n) ) (3)

η is a time-varying learning rate, defining how much the value of each neuron will change throughout the training process,

and hj,i is the neighborhood influence calculated by distance from the winner and time. A fully trained SOM provides a

continuous input space, a spatially discrete output space of neurons, and a nonlinear model of the given dataset7.

B. Contextual Map

Further advances in self-organizing maps provided a modification called the contextual self-organizing map7. Contextual

maps expand upon the trained SOM by generating labels for each neuron in the lattice. The input vectors are fed back into

the map once it is fully trained so that each neuron can acquire a label to describe its weight vector. These labels not only

describe the content of each individual neuron but can also be used to understand the similarity that a neuron has to the

surrounding neurons. Figure 2 shows an example of a contextual generated from research lead by T. Kohonen, where the

letters on nodes are the contextual labels (some nodes have multiple labels).

Figure 2: World Poverty Contextual Map8

Figure 3: World Poverty Map8:

Colors Transferred from Contextual Map

Each label in this map corresponds to a specific country on the world, and the map is grouping these countries based

upon poverty level indicators. The colors displayed within this map were defined by the researchers to connect the nodes

with their countries’ locations on a global map. This way, once the SOM finishes, the resulting colors of each node on the

map can be placed on the world map so that the geographic relationships can be shown. The results of this technique are

shown in Figure 3. This is only one example usage of the contextual SOM. In the case of the research for this paper, a

coloring scheme can be employed to convey additional information about the underlying nature of the dataset. The

principles used to generate this contextual map will be used to gain an understanding and visualization of optimization

problems.

II.Background

The background of this paper describes the current state of research in the area of self-organizing maps applied to

optimization. The authors referenced below have proven that SOMs can be utilized to further understand a design space

visually, and their idea for objective function visualization can be adopted to display the results of this paper.

Su, Shao, and Lee introduced SOM-based optimization (SOMO)9 as a method for solving optimization problems using a

modified self-organizing map. The difference between SOMO and a traditional SOM is the modification made to the

winning node calculation. SOMO’s calculation of the winner uses the objective function, meaning that the 'distance' is the

resulting evaluation of the objective function using the node weight vectors as inputs. This method was also able to plot the

two dimensional lattice in 3D space using the objective function value for each node as the third dimension. The visual

output of the SOMO method proves a viable method for visualizing the design space. The drawbacks of this method are that

it is constrained to continuous optimization problems, and was untested at a dimensionality greater than 30.

Matthews10

used self-organizing maps and applied them to conceptual design of gas turbines and aircraft wings. In both

of these design scenarios the author generated test points from running physical simulations. Matthews was able to use the

component maps of the SOM, which are the extraction of each variable from the trained map, to analyze the design space

interaction. With the extraction of these variables into separate maps, it was possible to inspect the implicit relationships

between variables, which principle component analysis was not able to detect. The author combined the use of SOMs with

the Tanimoto metric10

to automatically extract these implicit relationships between the variables comprising the map. After

the Tanimoto metric had identified potential areas of interest, human intervention was required to verify these as areas of

interest, as the computer only identified potential areas of interest. Overall, the work by Matthews proved to advance the

role of self-organizing maps in design and optimization. This method of examining the component maps could fit in place to

be a next-step after an initial examination of a trained contextual map.

Bishop and Svensen developed the Generative Topographic Mapping (GTM)11

, a modification to the traditional self-

organizing map which maps higher dimensional data to a low dimensional space, similar to the SOM, using latent variables.

The GTM also optimizes its node weight values for the provided dataset, which eliminates the limitations of the SOM in

terms of training parameters, but adds to the computational expense. The GTM was later used by Holden and Keane12

in a

comparison of methods to visualize the aircraft design space, furthermore, this study compared both SOMs and GTMs at

completing the task of visualizing the design space. The SOMs benefited from the ability to analyze individual variables

independently, but the GTM provided an exact solution to the design problem. While the GTM gave a promising result in

solving an eight dimensional problem, the authors noticed a decline in ability to discern a promising solution when

increasing the space to 14 dimensions. It is also important to note that the GTM suffers from increased computation time,

and the inability to extract variables from the design space for evaluation.

Milano, Koumoutsakos, and Schmidhuber13

utilized the Kohonen SOM to reduce the dimensionality and visualize

optimization problems. They were able to train the Kohonen map on the input space of an optimization function following

the rules of the traditional SOM. After the map was trained, they calculated the objective function values associated with

each node's weights. With these objective function values and a trained and organized map the authors were able to draw

contour lines on the two dimensional lattice to show the structure of the input space. The authors also displayed these same

results in three dimensions using the objective function value as the third dimension. This idea follows the same lines as the

visualization completed in the SOMO algorithm in that both methods take the two dimensional lattice and plot it in 3D

space using the objective function value as the third dimension.

The research described above provides a basis for the work in this article, contextual self-organizing maps used to

construct an understanding of the design space, and provide a means to visualize the resulting map. The method described in

this paper will be trained without modifications to the SOM algorithm, and will expand on the ideas presented above. The

first requirement of SOMO and the Milano, et al. methods is the availability of an objective function. Without an objective

function, the SOMO map cannot train and Milano, et al. cannot display the results of their training. Additionally, Matthews

and Holden and Keane require the user to inspect the component graphs of the trained map (variables extracted to show

their pattern), which can be difficult given high dimensionality or very complex problems. The contextual SOM method

described in this paper does not require an objective function, but simply a set of data points. For simplicity in testing, these

points were generated from well-known problems.

This method will plot the resulting lattice map in a two dimensional space to show the topological pattern of the input

space after undergoing dimensionality reduction9,13

. Also, given a visual of the design space, using the idea of extracting the

component variables10,12

can enhance the understanding of the input space by further exploring the interplay between

variables.

This paper attempts to investigate the application of the contextual SOM with optimization functions by:

1. Visualizing the resulting SOM to show the structure of the design space.

2. Verifying the contextual maps display details of the design space by using functions that can be visualized in lower

dimensions.

3. Expanding these initial problems to new problems of higher dimensionality, and unknown design spaces.

Some of the potential results and applications of this work can be tied to the following ideas:

1. Gain information about the characteristics of an optimization problem such as modality (multi-modal), curvature,

linearity, and degree.

2. Provide means for selection of a solution algorithm: numerical, heuristic, etc.

3. Determine distinctiveness of design points and design space, meaning that an initial search area may be found using

contextual maps.

III.Methodology

The algorithm used in this paper follows combined methodologies described above for contextual self-organizing maps.

First, the empty map is generated and then initialized with random weights. Next, the map is provided an initial

neighborhood radius and learning rate, which are standard training parameters. In this case, the initial neighborhood is set to

the entire map, and the learning rate initialized at 0.1 (recommended value from Neural Network7 text) but then decays

exponentially. The map is trained using inputs chosen by randomly selecting points in the design space. The training portion

of this method utilizes the design variable values and the objective function values. An objective function is not necessary to

utilize this method, but was the method by which the data was generated. The training is based upon the winner-takes-all

method as described above where the Euclidean distance determines the winning node. Once the wining node is found, the

update, Equation (3) is used to influence the surrounding nodes in the lattice.

After training is complete, the contextual map phase is executed, which returns the input data to the map, calculates the

winning node for each input data vector, and assigns its associated objective function value to the contextual map label.

Given that the lattice or network contains many fewer nodes than input vectors, each node ends with a number of associated

contextual map labels. Each node’s final label is found by gathering statistical measures to evaluate the node's performance

with respect to the objective function. The first step is to calculate the mean, standard deviation, and minimum value.

Secondly, the data for each node is normalized for coloring purposes described below.

The HSV14

or 'hue saturation value' color scheme was used to display a combination of the mean, standard deviation and

minimum values for the labels of each node. Visually, the mean objective function value can be found by examining the

color of the node, where green is closest and red is farthest from optimal. In other words, red corresponds to a high mean

and green corresponds to a low mean. The standard deviation can be determined by looking at the vibrance of each color.

The more vibrant colors have lower standard deviations and the darker colors have higher standard deviations. The

minimum value is displayed using the whiteness of each node. In essence, the more pigment allowed in a color (further

from white), the lower the minimum value contained within the node. To re-connect with the idea of coloring the map, this

coloring scheme makes it possible to discern the mean, standard deviation, and minimum value of a given neuron and how

its properties relate to other regions of the design space. As an example, a vibrant red node will have a high mean, low

standard deviation, and low minimum value. Alternatively, a dark colored green node will have a low mean, high standard

deviation, and moderate minimum value. Lastly, a white colored node will have a high minimum value. These color

properties provide the ability to compare properties of various nodes within a given map. The understanding of the makeup

of each node allows an investigator to learn more about the design space and optimization problem.

The idea of using contextual self-organizing maps has previously not been applied to the area of optimization, but has

remained in the domain of database organization, and language processing. This paper utilizes contextual maps to explore

the design space in a multivariable visual manner without the necessity of comparing multiple single variable maps to

explain the trends within the design space. Additionally, the HSV color scheme is a beneficial addition to the contextual

SOM because it can display the characteristics of a given node without overwhelming the viewer.

IV.Results

The following presents the results of training the described contextual self-organizing maps on three different functions:

Rosenbrock’s Function15

, Griewank’s Function17

, and Ackley’s Function18

. Data points were generated randomly for each of

these functions at sample sizes (SS) of 1000 and 10000. Each problem was tested with two, five, and ten dimensional

datasets. Each data point generated was accompanied by an evaluation of the objective function for contextual purposes

described in Methodology.

Figure 4 displays the result of plotting the design space of

Rosenbrock’s Valley function. This function was generated and

evaluated as a 2D problem, and plotted in three dimensions with

the vertical axis being the objective function value. Rosenbrock’s

Function (banana function) is a unimodal function with a

parabolic shaped valley that is scalable to n dimensions.

Figure 5 displays contextual SOMs trained to datasets

produced using the Rosenbrock’s Function. Sample sizes (SS) of

1000 and 10000 at 2D, 5D, and 10D are shown. In each case, the

self-organization and contextual labeling reduce the

dimensionality of the visualization to a two dimensional, visible,

space. As described in Methodology, the nodes have set positions

within the lattice, but the containing weight vectors are modified

to best cover the input space during training. This method insures

that the resulting adjacent nodes contain data samples of similar

design variables.

As a proof of concept for the contextual SOM, the

preliminary test cases were 2D functions. This allowed

comparison with observed results of known

optimization problems. One example, the resulting 2D

contextual map in Figure 5A, shows distinctive

properties of Figure 4 (steep edges and low area in the

center). Due to the coloring scheme described in

Methodology, when Figure 4 has a large slope, the

corresponding node areas will contain high standard

deviation causing dark colored nodes. When it has low

mean and is relatively flat, it will result in bright green

nodes. In each 2D case, the map organizes itself in the

same layout as the XY plane in Figure 4 and uses the

described coloring scheme to represent the Z-axis.

This direct representative behavior of the graph can

only be expected for 2D problems because there is no

dimensionality reduction in producing the resulting

map.

Although dimensionality is reduced in both the 5

and 10 dimensional cases, information regarding the

characteristics of the design space can still be retained

from the resulting maps. In Figure 5D, the red

highlighted area is generally a bright green color. As

described in the methodology section, this represents

an area with both low mean and low standard deviation

amongst the data points contained in the nodes. The

edges, which are also in one large connected area,

appear dark brown. This color shows that the mean

value is higher (farther from green), and there is a high

standard deviation (darker color) of the data that each

of the nodes contain. This same trend is generally

observed in each of the remaining subfigures of Figure

5. In each case, this information may be useful to tell a

designer that there is a large area of the given design

space with similar low objective function values, and

there is also a large area ascending away from these low areas into higher values. This information could be enough for the

designer to gain a quick understanding of what the overall design space may look like.

The next function examined was the Griewank Function shown in Figure 6. This function, as in the case of

Rosenbrock’s Valley, is n dimensional. The Griewank function has an overall circular shape filtering to the center, however

has many peaks and valleys and is therefore considered multimodal.

Figure 7 shows that in the 2D training of the Griewank Function, the contextual SOM captures the overall look of the

Figure 4: Rosenbrock's Function16

Figure 5: Rosenbrock’s Function Results

function it has been trained to. It shows both the large

circular curvature toward the center and the sharper

sloping peaks in the corners. These features are

brought out by the large bright green area in the

center as it fades to the outside becoming darker and

getting further away from green. Although the figure

shows the general trends of the data, the 2D scenario

fails to show the multimodal behavior of the graph.

This could be caused by the normalization of the

standard deviation calculated within each node. If

multimodal behavior is captured, the resulting

contextual SOM should show spotted regions of dark

nodes meaning that even though the values are close

in the design space (Euclidean Distance), their

respective objective function values are far apart

causing the node to show color representing a high

standard deviation of the data within said node.

Examples of this observation can be found in

Figure 7(C-F). In the figures, dark nodes scattered

throughout suggest that the data is highly variable

within small regions of the design space. There is no

indication of any large flat or steep regions like the

results of Rosenbrock’s Valley function.

The final function tested was Ackley’s Function shown in Figure 8. This function is also n dimensional. Ackley’s Path

is multimodal as was observed in the case of Griewank’s Function.

Again, the 2D results of the contextual SOM shown in Figure 9 retain the overall shape of the 2D Ackley’s Path function

plotted in Figure 8. Much like the case of the Griewank Function, the 2D results do not capture the multimodal nature of

the function, possibly due to the same reason as before. Results show that there is a large flat area with very high minimum

values (colors close to white), and a small region where the standard deviation increases and the mean objective function

value decreases (colors getting darker and increasingly green).

In the 5D cases (Figure 9C & 9D), results indicate a large flat

plane with high minimum values surrounding a small region with

very high standard deviation sloping toward lower values (shown

by the colors approaching green) similar to the 2D results. Once

again, although the information may be helpful to determine overall

trends, this, as in the 2D test, does not capture the multimodal

behavior of the function.

The scattered regions of dark nodes in the 10D results (Figure

9E & 9F) suggest that the function is multimodal, however, neither

the 1000 or 10000 sample set capture a high plane and small region

to indicate a possible global minima similar to 2D and 5D. This

shows that using different dimensionality can pull out different

characteristics of a given design problem.

Figure 7: Griewank Function Results

Figure 8: Ackley Function18

Figure 6: Griewank Function17

V.Summary and Conclusions

The presented examples in this research are to show the effectiveness of using contextual identifiers in creating a self-

organizing map for use in optimization problems. The results first demonstrate that the contextual self-organizing map can

group similar data points within the design space. Once the data is grouped by the SOM, the map can be colored based on

the statistical properties of each resulting node. With the map colored according to mean, standard deviation, and minimum

value, investigators can gain quick insight into overall behaviors of a given design space.

Future work for this research will include testing larger contextual SOMs (more nodes to train the dataset to), larger

datasets (to train the maps from), and higher dimensionality datasets. Limiting the size of each SOM in the examples

demonstrated may cause a loss of useful information due to the greater reliance on the statistical properties within each

resulting node. Doing so may limit the contextual abilities of the map to represent the overall results. Also, training each

map from a larger dataset may create a more continuous design space resulting in a more understandable, two-dimensional

representation of the design space and its properties. Finally, Solutions to higher dimensional problems are becoming more

and more in demand as the complexity of products continue to increase. Allowing the SOM to organize and the contextual

SOM to visualize a high dimensional dataset could help industry save time and costs by allowing a greater understanding of

the design space in a two dimensional (human readable) format. In conclusion, the immediate goals of our future work are

to gain an understanding of the required map sizes and training parameters that best fit the data sets given certain size and

dimensionality.

Figure 9: Ackley Function Results

VI.References 1 Eddy, J., Lewis, K., “Multidimensional Design Visualization in Multiobjective Optimization”, 9th AIAA/ISSMO Symposium on

Multidisciplinary Analysis and Optimization, Atlanta, GA, AIAA, AIAA-2002-5621. 2 Winer, E.H., Bloebaum, C.L., “Visual Design Steering for Optimization Solution Improvement,” Structural and Multidisciplinary

Optimization,22(3), p. 219-229. 3 Chiu, P-W., Naim, A.M., Lewis, K.E., Bloebaum, C.L., “The hyper-radial visualization method for multi-attribute decision-making

under uncertainty”, International Journal of Product Development, Vol. 9, No. 1-3, 2009, p. 4-31. 4 Stump, G., Yukish, M., Simpson, T.W., “The Advanced Trade Space Visualizer: An Engineering Decision-Making Tool,” 10th

AIAA/ISSMO Multidisciplinary Analysis and Optimization Conference, Albany, NY, AIAA, AIAA-2004-4568. 5 Kohonen, T., “The self-organizing map”,Neurocomputing, Vol. 21, No. 1-3, November 6, 1998, p 1-6. 6 Koua, E.L., Kraak, M-J., “Geovisualization to support the exploration of large health and demographic survey data”, International

Journal of Health Geographics,Vol. 3, No. 21, June 4, 2004. 7 Haykin, S., “Neural Networks A Comprehensive Foundation”, Prentice Hall Publishing, Second Edition, 1999, ISBN: 978-

0132733502. 8 Kaski, S., “Example Application of the SOM: World Poverty Map”, Neural Networks Research Centre, September 1997.

[http://www.cis.hut.fi/research/som-research/worldmap.html. Accessed 2/12/2010.] 9 Su, M.C., Zhao, Y.X., Lee, J., “Som-based Optimization”, IEEE International Conference on Neural Networks, 2004, 781-786. 10 Matthews, P.C. “The Application of Self Organizing Maps in Conceptual Design”, PhD thesis, Cambridge University, Engineering

Department, 2001. 11 Bishop, C.M., Svensen, M., Williams, C.K.I., “GTM: The generative topographic mapping”, Neural Computation MIT Press, Vol.

10, No. 1, 1998, p 215. 12 Holden, C.M.E., Keane, A.J., “Visualization Methodologies in Aircraft Design”, Proceedings of the 10th AIAA/ISSMO

Multidisciplinary Analysis and Optimization Conference, Albany, NY, 2004, p 1-13. 13 Milano, M.. Koumoutsakos, P., Schmidhuber, J., “Self-organizing nets for optimization”, IEEE Transactions on Neural Networks,

Vol 15, No. 3, 2004, p. 756-758. 14 Clark, A.C, Wiebe, E.N., “Color Principles – Hue, Saturation, and Value”, NC State University, May 2000,

[http://www.ncsu.edu/scivis/lessons/colormodels/color_models2.html. Accessed 2/12/2010.] 15 Hedar, A., “Test Functions for Unconstrained Global Optimization – Rosenbrock Function”, Global Optimization Methods and

Codes, [http://www-optima.amp.i.kyoto-u.ac.jp/member/student/hedar/Hedar_files/TestGO_files/Page364.htm. Accessed 8/4/2010.] 16 “Rosenbrock’s function”, MedWiki, [http://medlibrary.org/medwiki/Rosenbrock's_function. Accessed 8/24/2010.]

17 Hedar, A., “Test Functions for Unconstrained Global Optimization – Griewank Function”, Global Optimization Methods and

Codes, [http://www-optima.amp.i.kyoto-u.ac.jp/member/student/hedar/Hedar_files/TestGO_files/Page364.htm. Accessed 8/4/2010.] 18 Hedar, A., “Test Functions for Unconstrained Global Optimization – Ackley Function”, Global Optimization Methods and Codes,

[http://www-optima.amp.i.kyoto-u.ac.jp/member/student/hedar/Hedar_files/TestGO_files/Page364.htm. Accessed 8/4/2010.]


Recommended