9
Hindawi Publishing Corporation International Journal o Computer Games echnology V olume , Art icle ID ,  pages http://dx.doi.org/.// Research Article Single- versus Multiobjective Optimization for Evolution of Neural Controllers in Ms. Pac-Man Tse Guan Tan, Jason Teo, and Kim On Chin Evolutionary Computing Laboratory, School of Engineering and Information echnology, Universiti Malaysia,  Jalan (UMS), Kota Kinabal u, Sabah, Malaysia Correspondence should be addressed to se Guan an; [email protected] Received September ; Revised December ; Accepted January Academic Editor: Abdennour El Rhalibi Copyright © se Guan an et al. Tis is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproductio n in any medium, provided the original work is properly cited. Te objective o this study is to ocus on the automatic generation o game articial intelligence (AI) controllers or Ms. Pac-Man agent by using articial neural network (ANN) and multiobjective articial evolution. Te Pareto Archived Evolution Strategy (P AES ) is use d to gen erate a Pa reto op timal set o ANN s tha t op timize the coni cti ng obj ective s o max imi zing Ms. Pa c-Man scores (screen-capture mode) and minimizing neural network complexity. Tis proposed algorithm is called Pareto Archived Evolution Strategy Neural Network or PAESNet. Tree dierent architectures o PAESNet were investigated, namely, PAESNet with xed number o hidden neurons (PAESNet F), PAESNet with varied number o hidden neurons (PAESNet V), and the PAESNet with multiobjective techniques (PAESNet M). A comparison between the single- versus multiobjective optimization is conducted in both training and testing processes. In general, thereore, it seems that PAESNet F yielded better results in training phase. But the PAESNet M successully reduces the runtime operation and complexity o ANN by minimizing the number o hidden neurons needed in hidden layer and also it provides better generalization capability or controlling the game agent in a nondeterministic and dynamic environment . 1. Introduction A number o optimization solution techniques have been intr oduce d or solving Mul ti-Ob jectiv e Prob lems (MOP s) []. An MOP has a set o conicting objective unctions subject to certain constraints which are to be minimized or maximized []. Among these techniques, Evolutionary Algorithms (EAs) are particularly suited or handling MOPs [,  ] because o its population approach that can help in nding a set o trade-osolutions in single simulation run, instead o having to perorm a series o separate runs such as in the case o traditional optimization techniques. Moreov er, EAs have been successully used in solving complex prob- lems such as discontinuities, multimodality, disjoint easible spaces, and noisy unction evaluations [ ]. A large range o pract ical app licat ions o Mul ti-Ob jectiv e Evo luti onary Algorithms (MOEAs) to real-lie problems across a host o dierent disciplines can be ound in the reerence texts by Deb [] and Coello et al. [ ]. Tere are several types o eective MOEAs such as Pareto Archived Evolution Strate gy (PAES) [], Strength Pareto Evolutionary Algorithm [ ], No ndo minat ed Sor ting Geneti c Alg ori thm II [ ], and Pa re to- rontier Dierential Evolution [ ]. Genera lly , MOEAs are able to solv e separ ate distinct  varied dimensional optimization problems. In other words, MOEAs outperormed single-objective EAs without com- binin g and resorting those mult iple problems into single weigh ted-sum objective. Such weighted-s um methods are disadvantageous in obtaining suitable mode or combining dierent objectives into a single-objective unction, which caus ed high-c ost eecti ve [ ]. Furth ermore, each evolu- tionary run generates single solution; the second solution will only be generated afer the weights are changed; these processes will be repeated or obtaining other solutions [ , ]. Another distinct advantage o MOEAs is its capability in generating a complete set o Pareto optimal solutions in a single run with provides users a cho ice o solutions or trade- obetween dierent objectives. On theother han d, the dis advantages o MOEAs[ , , ] are (i) as the number o objectives increases, the coverage

Single- Versus Multiobjective Optimization for Evolution of Neural Controllers in Ms. Pac-Man

Embed Size (px)

DESCRIPTION

Es un paper de Inteligencia artificial con caso de estudio en pacman

Citation preview

  • Hindawi Publishing CorporationInternational Journal of Computer Games TechnologyVolume 2013, Article ID 170914, 7 pageshttp://dx.doi.org/10.1155/2013/170914

    Research ArticleSingle- versus Multiobjective Optimization for Evolution ofNeural Controllers in Ms. Pac-Man

    Tse Guan Tan, Jason Teo, and Kim On Chin

    Evolutionary Computing Laboratory, School of Engineering and Information Technology, Universiti Malaysia,Jalan (UMS), 88400 Kota Kinabalu, Sabah, Malaysia

    Correspondence should be addressed to Tse Guan Tan; [email protected]

    Received 19 September 2012; Revised 14 December 2012; Accepted 4 January 2013

    Academic Editor: Abdennour El Rhalibi

    Copyright 2013 Tse Guan Tan et al. This is an open access article distributed under the Creative Commons Attribution License,which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

    The objective of this study is to focus on the automatic generation of game artificial intelligence (AI) controllers for Ms. Pac-Managent by using artificial neural network (ANN) and multiobjective artificial evolution. The Pareto Archived Evolution Strategy(PAES) is used to generate a Pareto optimal set of ANNs that optimize the conflicting objectives of maximizingMs. Pac-Man scores(screen-capture mode) and minimizing neural network complexity. This proposed algorithm is called Pareto Archived EvolutionStrategy Neural Network or PAESNet. Three different architectures of PAESNet were investigated, namely, PAESNet with fixednumber of hidden neurons (PAESNet F), PAESNet with varied number of hidden neurons (PAESNet V), and the PAESNet withmultiobjective techniques (PAESNet M). A comparison between the single- versus multiobjective optimization is conducted inboth training and testing processes. In general, therefore, it seems that PAESNet F yielded better results in training phase. But thePAESNet M successfully reduces the runtime operation and complexity of ANN by minimizing the number of hidden neuronsneeded in hidden layer and also it provides better generalization capability for controlling the game agent in a nondeterministicand dynamic environment.

    1. Introduction

    A number of optimization solution techniques have beenintroduced for solving Multi-Objective Problems (MOPs)[1]. An MOP has a set of conflicting objective functionssubject to certain constraints which are to be minimizedor maximized [2]. Among these techniques, EvolutionaryAlgorithms (EAs) are particularly suited for handling MOPs[3, 4] because of its population approach that can help infinding a set of trade-off solutions in single simulation run,instead of having to perform a series of separate runs such asin the case of traditional optimization techniques. Moreover,EAs have been successfully used in solving complex prob-lems such as discontinuities, multimodality, disjoint feasiblespaces, and noisy function evaluations [5]. A large rangeof practical applications of Multi-Objective EvolutionaryAlgorithms (MOEAs) to real-life problems across a host ofdifferent disciplines can be found in the reference texts byDeb [6] and Coello et al. [3]. There are several types ofeffective MOEAs such as Pareto Archived Evolution Strategy

    (PAES) [7], Strength Pareto Evolutionary Algorithm 2 [8],Nondominated SortingGenetic Algorithm II [9], and Pareto-frontier Differential Evolution [10].

    Generally, MOEAs are able to solve separate distinctvaried dimensional optimization problems. In other words,MOEAs outperformed single-objective EAs without com-bining and resorting those multiple problems into singleweighted-sum objective. Such weighted-sum methods aredisadvantageous in obtaining suitable mode for combiningdifferent objectives into a single-objective function, whichcaused high-cost effective [6]. Furthermore, each evolu-tionary run generates single solution; the second solutionwill only be generated after the weights are changed; theseprocesses will be repeated for obtaining other solutions [11,12]. Another distinct advantage of MOEAs is its capabilityin generating a complete set of Pareto optimal solutions in asingle run with provides users a choice of solutions for trade-off between different objectives.

    On the other hand, the disadvantages ofMOEAs [6, 11, 12]are (i) as the number of objectives increases, the coverage

  • 2 International Journal of Computer Games Technology

    of the Pareto front sparser and become unable to providea comprehensive set of solutions over multiple dimensions,(ii) problematic inmaintaining good spread diverse solutionsalong the Pareto front, and (iii) the difficulty of fitness sharingdecision making in MOEAs which utilize multiple popula-tions. This open research question provides the motivationfor the work in this paper. In other words, Is a single-objective optimization technique better than multiobjectiveoptimization in real-life problems?

    Games are one of the commonused platforms for answer-ing research question by allowing the testing and comparisonof new and experimental approaches on a challenging butwell-defined problem [1317]. In this research, Ms. Pac-man has been chosen as the test-bed due to its ease ofuse in comparing the performances of the single-objectiveoptimization and multiobjective optimization techniques.

    In this study, a feed-forward artificial neural network(FFNN) is used and later evolved with PAES, a well-knownand simple MOEA, for computer-based players to learn andoptimally play Ms. Pac-man game. There are two distinctobjectives to be optimized: (i) maximize the Ms. Pac-mansgame scores and (ii) minimize the number of hidden neuronsused in the FFNN architecture.

    A comparative empirical experiment will be conducted inorder to verify the performances for the methods used.

    (1) Single-objective optimization: the first experimentuses fixed number of hidden neurons in the FFNNand only maximizes Ms. Pac-man game scores,namely, PAESNet F.

    (2) Single-objective optimization: the second experimentis using variable number of hidden neurons in theFFNNand onlymaximizesMs. Pac-man game scores,namely, PAESNet V.

    (3) Multiobjective optimization: the third experimentmaximizes the game scores as well as minimizes thehidden neurons in the FFNN, namely, PAESNet M.

    The main contribution of these proposed algorithms isto create computer-based agent that not only is able to makeintelligent decisions like human players in the dynamic gameenvironments, but also is highly beneficial to the real-worldproblems with the successful application of these techniques,such as in the application of robotics and other complexsystems.

    2. Other Related Researches

    Basically, most studies in the game of Ms. Pac-man haveonly focussed on hand-coded rule-based (RB) approachesor other specific methods [1821]. Although these methodscan achieve quite high scores, they are associated with somelimitations. Firstly, game domains contain highly complexsolution spaces that require a large number of rules in orderto represent a set of all possible situations and correspondingactions in game environments. For instance, Szita and Lor-incz [21] list 42 rules of a very basic hand-coded RB agentused in Ms. Pac-man game from the lists of action modulesand observations to control the behaviour of agent. Secondly,

    the computation time required to exhaustively explore thesearch space is very expensive indeed if large sets of rulesare used by the search strategies. Thirdly, there is a lack ofgeneralization across different game domains or platformsbecause they would only apply in that particular game orgenre of game.

    The intention of this research is to create game controllerscapable of general intelligent action without requiring anydomain-dependent solution and also trying to be proficientin other games by just changing the input and output valuesof ANN. Thus, the experimental results will be compared toan appropriate reference system created by Lucas [22]. Lucasused generalmethods in designing the game controller whichevolves ANN by using evolutionary strategy to play Ms. Pac-man.The input of the network is a handcrafted feature vectorthat consists of the distance to each normal ghost, distanceto each edible ghost, location of current node, distance tonearest pill, the distance to nearest power pill, and distance tonearest junction, whereas the calculated output is a score forevery possible next location given the agents current location.ES is applied to evolve ANN connection weights. The bestevolved agent with (10 + 10)-ES had an average score of 4781over 100 runs of the nondeterministic game.

    3. Methods and Parameter Setting

    This investigation has two modes of operation: training andtesting as shown in Figure 1. In the trainingmode, the FFNNsare trained using evolution-based algorithm. The agents willlearn to playmany games in order to optimizeweights, biases,and number of hidden neurons in FFNN architecture, asan effective mode for training. After the training process,the neural network is tested for generalization using theoptimized networks.

    3.1. Pareto Archived Evolution Strategy. The (1 + 1)-PAES fora two-membered PAES has been applied for simultaneouslyoptimizing network parameters and architecture to solve sin-gle, and multiobjective optimization problems. The resultingalgorithm is referred to as the PAESNet. Figure 2 shows theflowchart of PAESNet and fitness evaluation process. Thestrengths of PAES are listed as follows:

    (1) simple structure;(2) easy to implement;(3) (1 + 1)-PAES and (1 + )-PAES are based on

    local search method with lower computational effortrequired compared to population-based MOEAs.

    (4) a small number of parameters are needed;(5) the simplest possible nontrivial algorithm capable of

    generating diverse solutions in the Pareto optimal set[23].

    3.2. Single Objective: PAESNet with Fixed Number of HiddenNeurons and PAESNet with Varied Number of Hidden Neu-rons. Two systems are discussed in this section, which arethe PAESNet with fixed number of hidden neuron (PAES-Net F) and PAESNet with varied number of hidden neurons

  • International Journal of Computer Games Technology 3

    :

    Proposed systems

    Two modes

    (1)Training (2) Testing

    -Pareto Archived Evolution Strategy Neural Network (PAESNet)(1) With a fixed number of hidden neurons (PAESNet F)(2) With a varied number of hidden neurons (PAESNet V)

    Single-objective optimization (1 and 2):(i) Maximize the game scores

    Biobjective optimization:(3) With multiobjective (PAESNet M)

    (i) Maximize the game scores(ii) Minimize the number of hidden neurons

    -To choose the best evolved network from the best run ofPAESNet F, PAESNet V, and PAESNet M, respectively.-10 tests for each system.

    -To train the PAESNet F, PAESNet V, and PAESNet Mwith 500 evaluations, respectively.-10 experimental runs for each system.

    Figure 1: The overview of the study.

    (PAESNet V). The default number of hidden neurons isset to 20. In the initialization phase, the ANN weightsand biases are encoded into a chromosome from uniformdistribution with range [1, 1] to act as parent and its fitnessis evaluated. Subsequently, polynomial mutation operator isused with distribution index = 20.0 to create an offspringfrom the parent and the fitness is evaluated. After that, thefitnesses of the offspring and parent are compared. If theoffspring performs better than the parent, then the parentis replaced by the offspring as a new parent for the nextevaluation. Otherwise the offspring is eliminated and a newmutated offspring is generated. If the parent and the offspringare incomparable, the offspring is compared with a set ofpreviously nondominated individuals in the archive. Belowis the description of the archiving process in PAESNet.

    There are three possible situations that can occur betweenthe comparison of the offspring and archive [7, 24, 25]. First,if the offspring is dominated by a member of the archive,then the offspring is discarded and a new mutated offspringis created from the parent. Second, if the offspring dominatessome members of the archive, then the set of dominatedmembers is removed from the archive.The offspringwill thenbe added to the archive and it also becomes the parent ofthe next generation. Third, if the offspring and the archivemembers do not dominate each other, then the archivewill be maintained depending on the archive size. If thearchive is not full, the offspring will be directly copied tothe archive. Otherwise, in the scenario that the archive isfull, a neighborhood density measure is used to ensure thata well-spread distribution is maintained in the archive. If theoffspring has succeeded to increase the archive diversity, itwill replace the archive member in the most crowded gridlocation in order tomaintain themaximumarchive size. Notethat in this third situation, the offspring and the parent are thenondominated members of the archive. The neighborhood

    density measure is also applied for parent selection of thenext generation from both of them. If the offspring residesin the less crowded area than the parent, then the offspring isselected.

    3.3. Multiobjective: PAESNet with Multiobjective (PAES-Net M). The structure of this proposed algorithm is similarto the algorithms in Section 3.2 except for the architectureof the ANN. In this proposed algorithms, two objectives areinvolved. The first objective is to maximize the game scoreswhile the second objective is to minimize the number ofneurons in the FFNN. The initial value of hidden neurons isset to 20.

    3.4. Feed-Forward ANN. The typical FFNN is composed ofthree layers: input, hidden, and output layers [26].The follow-ing is used to describe the feed-forward ANN architecture:

    (1) ,, and are the numbers of input neurons, hiddenneurons, and output neurons, respectively;

    (2) and

    are the weights connecting input unit ,

    = 1, . . . , , to hidden unit , = 1, . . . , , and fromhidden unit to output unit , = 1, . . . , ;

    (3) is the input signal.

    The net input of a neuron is calculated using the weightedsum of inputs from all neurons in the previous layer, asfollows:

    =0

    . (1)

    Log-sigmoid (logsig) is used as the activation function inthe hidden and output layers. Based on [27], logsig has beenidentified as suitable activation function in ANN for creatingneural-based Ms. Pac-man agent.

  • 4 International Journal of Computer Games Technology

    Mutation operation

    No

    Comparison between offspring and archive members

    Yes

    No

    Yes

    Archive update operation

    Select new parent solution from offspring and parent solutions

    Fitness evaluation

    End

    Save offsprings score

    save parents score

    Fitness evaluation process

    Save game score

    Yes

    No

    Win

    Die

    Alive

    Control object

    Feedforward neural network

    Pac-Man plays

    Random weights, biases, and activehidden neurons initialization

    (create parent solution)

    (new game start)

    Fitness evaluation(new game start)

    Fitness evaluation(new game start)

    The setup mode for PAESNet M

    Add to archive and

    Terminate?

    (create offspring solution)

    Is parent better than offspring?

    Game inputs, weights, and biases

    Game over?

    Figure 2: The flowchart of PAESNet optimizer.

    3.5. Experimental Setting. The FFNN architecture of thismodel has a 5-20-1 structure, which consists of 5 inputs and1 output together with one hidden layer of 20 neurons. Thisnumber of hidden neurons was suggested by Lucas [22]. TheEuclidean distance is applied to calculate the distance in themaze as the inputs of the network were obtained based on thefollowing information:

    (1) the closest distance from agent to a pill;

    (2) the closest distance from agent to a power pill;

    (3) the closest distance from agent to a ghost;

    (4) the closest distance from agent to an edible ghost;

    (5) the closest distance from agent to a fruit.

    4. Experimental Results and Discussions

    The results obtained from the analysis of training and testingperformances can be compared in the tables and figuresbelow.

    4.1. Training Results. Table 1 presents the training resultsof mean, standard deviation (SD), minimum (Min) andmaximum (Max) values obtained from the best game scores,and number of hidden neurons in each run. The statistics(mean of scores, mean of hidden neurons) for PAESNet F,PAESNet V, and PAESNet M were (7161, 20), (5935, 9.7),and (5734, 8), respectively. According to the mean values ofscores, the results showed that PAESNet F has the highestaverage score. However, the best scores are comparable acrossall three approaches (7430 in PAESNet F, 7190 in PAESNet V,and 7170 in PAESNet M).On the other hand, taking themean

  • International Journal of Computer Games Technology 5

    Table 1: The training results.

    Run PAESNet F PAESNet V PAESNet MScore Neurons Score Neurons Score Neurons

    1 6710 20 7190 13 7170 72 7050 20 5810 11 5690 103 7290 20 6260 11 6400 84 7190 20 5450 6 4940 55 7430 20 5190 8 4670 96 6930 20 5520 6 5990 67 7080 20 6650 11 5280 98 7280 20 5980 9 5260 99 7320 20 5500 12 6120 810 7330 20 5800 10 5820 9

    Min 6710 20 5190 6 4670 5Max 7430 20 7190 13 7170 10Mean 7161 20 5935 9.7 5734 8SD 219.47 0 612.78 2.41 739.58 1.56Score: best score; Neurons: number of hidden neurons.

    Table 2: Win rates for training results across all 10 runs.

    Proposed algorithm PAESNet F PAESNet V PAESNet MPAESNet F 90% 90%PAESNet V 10% 60%PAESNet M 10% 40% ( versus ) + (B versus A) = 100%.

    values of hidden neurons, we observed that the PAESNet Mreduces the number of hidden neurons from 20 to 8 which isaround 60% improvement. This emphasizes the advantagesof MOEA approach in terms of computational complexity inFFNN.

    Additionally, the average scores of all the proposedalgorithms, PAESNet F (7161), PAESNet V (5935), andPAES-Net M (5734), are relatively higher when compared to Lucas(4781) [22] for training to play theMs. Pac-man.Hence, this isfurther a proof that the proposed systems with PAES are ableto usefully and automatically generate Ms. Pac-Man agentsthat display some intelligent playing behavior.

    Table 2 lists the win rates (WR) for each comparison,which is the number of runs an artificial controller wins pertotal number of runs as shown in (2). Firstly, for PAESNet Fversus PAESNet V, WR = 90%, PAESNet F won 9 out of 10runs compared to PAESNet V except Run 1, and the resultis same for PAESNet F versus PAESNet M. Subsequently, forPAESNet V versus PAESNet M, WR = 60%, PAESNet Vwon 6 out of 10 runs compared to PAESNet M except Run 3,Run 6, Run 9 and Run 10.The results clearly show that PAES-Net F outperformed the other two competing approaches.This result may be explained by the fact that PAESNet F isconcerned with a single-objective of maximizing the gamescores, while that of the PAESNet M is to find the set oftrade-off solutions between the scores and number of hiddenneurons. The acceptance of trade-off solutions is due to con-vergence performance and diversity preservation in Pareto

    3610

    4180

    4940

    5990

    7170

    0

    1000

    2000

    3000

    4000

    5000

    6000

    7000

    8000

    1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20

    Gam

    e sc

    ores

    Number of hidden neurons

    Run 1Run 2Run 3Run 4Run 5Run 6

    Run 7Run 8Run 9Run 10Global Pareto

    Global Pareto-frontier solutions obtained

    Figure 3:The global Pareto-frontier solutions obtained across all 10runs using multiobjective optimization.

    optimal front. Due to these two criteria, the multiobjectiveoptimization is harder than single-objective optimization.Another possible explanation for this is that multiobjectiveoptimization is dealing with two search spaces, which aredecision variable space and objective space compared tosingle-objective optimization just involving one search space(decision variable space). This factor may influence theperformance of PAESNet M:

    WR = runs wontotal of runs played

    100%. (2)

    The global Pareto-frontier solutions obtained with thegoal of maximizing scores and minimizing hidden neuronsacross all 10 runs using multiobjective optimization areillustrated in Figure 3. The global Pareto solutions are shownby the dotted line. As can be seen from the figure, thePAESNet M reported significantly decreases the number ofhidden neurons needed from 20 to the range of 3 to 7 nodesin the hidden layer as the optimized networks and the gamescores achieved were 3610, 4180, 4940, 5990, and 7170.

    4.2. Testing Results. After the training phase, the best evolvednetworks were used to test the generalization ability of themodels in order to score as high as possible.The selected bestnumbers of neurons in the hidden layer are 20, 13, and 7 forPAESNet F, PAESNet V, and PAESNet M, respectively, as theoptimum networks, as shown in Table 3.

  • 6 International Journal of Computer Games Technology

    Table 3: The best evolved networks from proposed algorithms.

    Proposed algorithm Number of hidden neuronsPAESNet F 20PAESNet V 13PAESNet M 7

    Table 4: The testing results.

    Run PAESNet F PAESNet V PAESNet M1 3090 4160 42502 2720 2740 27003 2020 2590 28904 4170 2100 53605 2500 2280 25106 2580 1080 27607 2780 4020 28608 1990 4820 20809 2090 2290 313010 4360 4880 3950

    Min 1990 1080 2080Max 4360 4880 5360Mean 2830 3096 3249SD 836.62 1286.48 983.04

    Table 5: Win rates for testing results across all 10 runs.

    Proposed algorithm PAESNet F PAESNet V PAESNet MPAESNet F 30% 20%PAESNet V 70% 40%PAESNet M 80% 60% ( versus ) + ( versus ) = 100%.

    Table 4 presents the testing results of the three proposedalgorithms. We can observe that both of the max andmean scores in PAESNet M (5360, 3249) were higher thanPAESNet F (4360, 2830) and PAESNet V (4880, 3096). Basedon mean values, PAESNet M was shown to have betterperformance compared to PAESNet F and PAESNet V.

    Table 5 lists the win rates for each comparison. Firstly, forPAESNet F versus PAESNet V, WR = 70% and PAESNet Vwon 7 out of 10 runs compared to PAESNet F, except Run 4,Run 5, and Run 6. Next, for PAESNet F versus PAESNet M,WR = 80% PAESNet M won 8 out of 10 runs compared toPAESNet F, except Run 2 and Run 10. Lastly, for PAESNet Vversus PAESNet M, WR = 60% and PAESNet M won 6 outof 10 runs compared to PAESNet V, except Run 2, Run 7, Run8, and Run 10. From this data, we can see that PAESNet Mresulted in the highest value of win rate compared to the twoalgorithms.ThePAESNet M successfully found the appropri-ate network architecture and parameters by maximizing thegame scores andminimizing the hidden neurons.Overall, thetesting results have shown that FFNNs and PAES have strongpotential for controlling game agents in the game world.

    5. Conclusions

    In this study, the FFNN is evolved with the PAES MOEA forthe computer player to automatically learn and optimally playthe game of Ms. Pac-man which is called PAESNet. Threeforms of PAESNet, PAESNet F, PAESNet V, and PAESNet M,were introduced to solve single- andmultiobjective optimiza-tion problems and compared to each other in the trainingand testing processes.The Pareto optimal front resulted fromeach MOEA run provided a set of NNs which maximizedthe scores of Ms. Pac-man and at the same time minimizedthe size of the controller. In the training process, PAESNet Foutperformed PAESNet V and PAESNet M. However, in thetesting process, PAESNet M outperformed the other twoalgorithms. One of the most significant findings to emergefrom this study is that the generalization performance ofthe neural networks could improve significantly by evolvingthe architecture and connection weights (including biases)synchronously via aMOEA approach as opposed to fixing thenetwork architecture and optimizing the scoring componentonly using a single-objective optimization approach.

    Acknowledgment

    This research is funded under the Science Fund ProjectSCF52-ICT-3/2008 granted by the Ministry of Science, Tech-nology and Innovation, Malaysia.

    References

    [1] C. Zheng and P. Wang, Application of flow and transportoptimization codes to groundwater pump-and-treat systems:Umatilla Army Depot, Oregon, Tech. Rep., University ofAlabama, Tuscaloosa, Ala, USA, 2001.

    [2] C. A. C. Coello, Evolutionary multi-objective optimization:a critical review, in Evolutionary Optimization, R. Sarker,M. Mohammadian, and X. Yao, Eds., pp. 117146, KluwerAcademic, Boston, Mass, USA, 2002.

    [3] C. A. C. Coello, G. B. Lamont, and D. A. van Veldhuizen,Evolutionary Algorithms for Solving Multi-Objective Problems,Springer, New York, NY, USA, 2007.

    [4] W. Q. Ying, Y. X. Li, C. Y. S. Phillip, Y. Wu, and F. H. Yu,Geometric thermodynamicalselection for evolutionary multi-objective optimization, Chinese Journal of Computers, vol. 33,no. 4, pp. 755767, 2010.

    [5] C. M. Fonseca and P. J. Fleming, An overview of evolutionaryalgorithms in multiobjective optimization, Evolutionary Com-putation, vol. 3, pp. 116, 1995.

    [6] K. Deb, Multi-Objective Optimization Using Evolutionary Algo-rithms, Wiley, New York, NY, USA, 2001.

    [7] J. D. Knowles and D. W. Corne, The Pareto archived evolutionstrategy: a new baseline algorithm for Pareto multiobjectiveoptimisation, in Proceedings of the IEEE Congress on Evolution-ary Computation (CEC 99), pp. 98105,Washington, DC, USA,July, 1999.

    [8] E. Zitzler, M. Laumanns, and L. Thiele, SPEA2: improvingthe strength Pareto evolutionary algorithm, Tech. Rep. 103,Computer Engineering and Network Laboratory (TIK), SwissFederal Institute of Technology (ETH), Zurich, Switzerland,2001.

  • International Journal of Computer Games Technology 7

    [9] K. Deb, A. Pratap, S. Agarwal, and T. Meyarivan, A fastand elitist multiobjective genetic algorithm: NSGA-II, IEEETransactions on Evolutionary Computation, vol. 6, no. 2, pp. 182197, 2002.

    [10] H. A. Abbass, R. Sarker, and C. Newton, PDE: a Pareto-frontier differential evolution approach for multi-objectiveoptimization problems, in Proceedings of the IEEE Conferenceon Evolutionary Computation, pp. 971978, May 2001.

    [11] C. A. C. Coello, G. T. Pulido, and E. M. Montes, Current andfuture research trends in evolutionary multiobjective optimiza-tion, in Information Processing with Evolutionary Algorithms:From Industrial Applications to Academic Speculations, M.Grana, R. Duro, A. dAnjou, and P. P. Wang, Eds., pp. 213231,Springer, London, UK, 2005.

    [12] C. A. C. Coello, Recent trends in evolutionary multiobjec-tive optimization, in Evolutionary Multiobjective Optimization:Theoretical Advances and Applications, A. Abraham, L. Jain, andR. Goldberg, Eds., pp. 732, Springer, London, UK, 2005.

    [13] K. T. Chang, K.O. Chin, J. Teo, andA.M. Jilui-Kiring, Evolvingneural controllers using GA for warcraft 3-real time strategygame, in Proceedings of the IEEE 6th International ConferenceonBio-InspiredComputing:Theories andApplications, pp. 1520,September 2011.

    [14] K. T. Chang, J. H. Ong, J. Teo, and K. O. Chin, The evolutionof gamebots for 3D first person shooter (FPS), in Proceedings ofthe IEEE 6th International Conference on Bio-Inspired Comput-ing: Theories and Applications, pp. 2126, September 2011.

    [15] K. T. Chang, J. Teo, K. O. Chin, and B. L. Chua, Automatic gen-eration of real time strategy tournament units using differentialevolution, in Proceedings of the IEEE Conference on SustainableUtilization and Development in Engineering and Technology, pp.101106, October 2011.

    [16] C. H. Ng, S. H. Niew, K. O. Chin, and J. Teo, Infinitemario bross AI using genetic algorithm, in Proceedings of theIEEE Conference on Sustainable Utilization and Development inEngineering and Technology, pp. 8589, October 2011.

    [17] J. H. Ong, J. Teo, and K. O. Chin, Interactive evolutionary pro-gramming for mobile games rules generation, in Proceedings ofthe IEEEConference on Sustainable Utilization andDevelopmentin Engineering and Technology, pp. 95100, Semenyih, Malaysia,October 2011.

    [18] A. Fitzgerald and C. B. Congdon, RAMP: a rule-based agentfor Ms. Pac-man, in Proceedings of the IEEE Congress on Evo-lutionary Computation (CEC 09), pp. 26462653, Trondheim,Norway, May 2009.

    [19] E. Galvan-Lopez, J. M. Swafford, M. ONeill, and A. Brabazon,Evolving a Ms. Pacman controller using grammatical evolu-tion, in Proceedings of the Applications of Evolutionary Com-putation, EvoApplicatons: EvoCOMPLEX, EvoGAMES, EvoIASP,EvoINTELLIGENCE, EvoNUM, and EvoSTOC, pp. 161170,Istanbul, Turkey, April 2010.

    [20] D. Robles and S. M. Lucas, A simple tree search method forplayingMs. Pac-man, in Proceedings of the IEEE Symposium onComputational Intelligence and Games (CIG 09), pp. 249255,September 2009.

    [21] I. Szita and A. Lorincz, Learning to play using low-complexityrule-based policies: illustrations throughMs. Pac-man, Journalof Artificial Intelligence Research, vol. 30, pp. 659684, 2007.

    [22] S. M. Lucas, Evolving a neural network location evaluator toplay Ms. Pac-man, in Proceedings of the IEEE Symposium onComputational Intelligence and Games, pp. 203210, Essex, UK,April 2005.

    [23] J. D. Knowles and D. W. Corne, Approximating the nondom-inated front using the Pareto Archived Evolution Strategy,Evolutionary Computation, vol. 8, no. 2, pp. 149172, 2000.

    [24] L. T. Bui and S. Alam, An introduction to multi-objective opti-mization, in Multi-Objective Optimization in ComputationalIntelligence:Theory and Practice, L. T. Bui and S. Alam, Eds., pp.119, IGI Global, Hershey, Pa, USA, 2008.

    [25] T.Wong, P. Bigras, and K. Khayati, Causality assignment usingmulti-objective evolutionary algorithms, in Proceedings of theIEEE International Conference on Systems,Man andCybernetics,pp. 3641, October 2002.

    [26] D. Svozil, V. Kvasnicka, and J. Pospchal, Tutorial: introductionto multi-layer feed-forward neural networks, Chemometricsand Intelligent Laboratory Systems, vol. 39, no. 1, pp. 4362, 1997.

    [27] T. G. Tan, J. Teo, and P. Anthony, Comparative investigation ofnon-linear activation functions in neural controllers for search-based game AI engineering, Artificial Intelligence Review, 2011.

  • Submit your manuscripts athttp://www.hindawi.com

    VLSI Design

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    International Journal of

    RotatingMachinery

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Hindawi Publishing Corporation http://www.hindawi.com

    Journal of

    EngineeringVolume 2014

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Shock and Vibration

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Mechanical Engineering

    Advances in

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Civil EngineeringAdvances in

    Acoustics and VibrationAdvances in

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Electrical and Computer Engineering

    Journal of

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Distributed Sensor Networks

    International Journal of

    The Scientific World JournalHindawi Publishing Corporation http://www.hindawi.com Volume 2014

    SensorsJournal of

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Modelling & Simulation in EngineeringHindawi Publishing Corporation http://www.hindawi.com Volume 2014

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Active and Passive Electronic Components

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Chemical EngineeringInternational Journal of

    Control Scienceand Engineering

    Journal of

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Antennas andPropagation

    International Journal of

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014

    Navigation and Observation

    International Journal of

    Advances inOptoElectronics

    Hindawi Publishing Corporation http://www.hindawi.com

    Volume 2014

    RoboticsJournal of

    Hindawi Publishing Corporationhttp://www.hindawi.com Volume 2014