APPLICATION OF FUZZY BEHAVIOR COORDINATION AND Q LEARNING IN ROBOT NAVIGATION

Similar documents
APPLICATION OF FUZZY BEHAVIOR COORDINATION AND Q LEARNING IN ROBOT NAVIGATION

COMPACT FUZZY Q LEARNING FOR AUTONOMOUS MOBILE ROBOT NAVIGATION

Q Learning Behavior on Autonomous Navigation of Physical Robot

AUTONOMOUS FIVE LEGS RESCUE ROBOT NAVIGATION IN CLUTTERED ENVIRONMENT

Behaviour-Based Control. IAR Lecture 5 Barbara Webb

Robot Architectures. Prof. Yanco , Fall 2011

Robot Architectures. Prof. Holly Yanco Spring 2014

Key-Words: - Fuzzy Behaviour Controls, Multiple Target Tracking, Obstacle Avoidance, Ultrasonic Range Finders

Fuzzy Logic Based Robot Navigation In Uncertain Environments By Multisensor Integration

Subsumption Architecture in Swarm Robotics. Cuong Nguyen Viet 16/11/2015

Sonar Behavior-Based Fuzzy Control for a Mobile Robot

The Necessity of Average Rewards in Cooperative Multirobot Learning

Dipartimento di Elettronica Informazione e Bioingegneria Robotics

Control Arbitration. Oct 12, 2005 RSS II Una-May O Reilly

A Mobile Robot Behavior Based Navigation Architecture using a Linear Graph of Passages as Landmarks for Path Definition

Hybrid Neuro-Fuzzy System for Mobile Robot Reactive Navigation

Reinforcement Learning for CPS Safety Engineering. Sam Green, Çetin Kaya Koç, Jieliang Luo University of California, Santa Barbara

Learning Reactive Neurocontrollers using Simulated Annealing for Mobile Robots

Mobile Robot Navigation Contest for Undergraduate Design and K-12 Outreach

Evolving High-Dimensional, Adaptive Camera-Based Speed Sensors

NAVIGATION OF MOBILE ROBOT USING THE PSO PARTICLE SWARM OPTIMIZATION

Fuzzy Logic for Behaviour Co-ordination and Multi-Agent Formation in RoboCup

Control of motion stability of the line tracer robot using fuzzy logic and kalman filter

A Novel Hybrid Fuzzy A* Robot Navigation System for Target Pursuit and Obstacle Avoidance

Simple Target Seek Based on Behavior

Review of Soft Computing Techniques used in Robotics Application

Adaptive Action Selection without Explicit Communication for Multi-robot Box-pushing

Path Following and Obstacle Avoidance Fuzzy Controller for Mobile Indoor Robots

CS594, Section 30682:

LOCAL OPERATOR INTERFACE. target alert teleop commands detection function sensor displays hardware configuration SEARCH. Search Controller MANUAL

Strategy for Collaboration in Robot Soccer

GA-based Learning in Behaviour Based Robotics

Sensor Data Fusion Using Kalman Filter

The Architecture of the Neural System for Control of a Mobile Robot

Biologically Inspired Embodied Evolution of Survival

Online Evolution for Cooperative Behavior in Group Robot Systems

Robotic Systems ECE 401RB Fall 2007

MULTI-LAYERED HYBRID ARCHITECTURE TO SOLVE COMPLEX TASKS OF AN AUTONOMOUS MOBILE ROBOT

Evolved Neurodynamics for Robot Control

Keywords Multi-Agent, Distributed, Cooperation, Fuzzy, Multi-Robot, Communication Protocol. Fig. 1. Architecture of the Robots.

Unit 1: Introduction to Autonomous Robotics

Localisation et navigation de robots

Obstacle avoidance based on fuzzy logic method for mobile robots in Cluttered Environment

Reactive Planning with Evolutionary Computation

A Reactive Robot Architecture with Planning on Demand

Mechatronics 19 (2009) Contents lists available at ScienceDirect. Mechatronics. journal homepage:

A Lego-Based Soccer-Playing Robot Competition For Teaching Design

An Experimental Comparison of Path Planning Techniques for Teams of Mobile Robots

COS Lecture 1 Autonomous Robot Navigation

Traffic Control for a Swarm of Robots: Avoiding Group Conflicts

Crucial Factors Affecting Cooperative Multirobot Learning

Transactions on Information and Communications Technologies vol 6, 1994 WIT Press, ISSN

Learning to Avoid Objects and Dock with a Mobile Robot

Overview Agents, environments, typical components

Fuzzy-Heuristic Robot Navigation in a Simulated Environment

Generating Personality Character in a Face Robot through Interaction with Human

CYCLIC GENETIC ALGORITHMS FOR EVOLVING MULTI-LOOP CONTROL PROGRAMS

A simple embedded stereoscopic vision system for an autonomous rover

Reinforcement Learning Simulations and Robotics

Using Policy Gradient Reinforcement Learning on Autonomous Robot Controllers

Traffic Control for a Swarm of Robots: Avoiding Target Congestion

ROBOCODE PROJECT AIBOT - MARKOV MODEL DRIVEN AIMING COMBINED WITH Q LEARNING FOR MOVEMENT

Adaptive Neuro-Fuzzy Controler With Genetic Training For Mobile Robot Control

Moving Obstacle Avoidance for Mobile Robot Moving on Designated Path

Cooperative Behavior Acquisition in A Multiple Mobile Robot Environment by Co-evolution

A Probabilistic Method for Planning Collision-free Trajectories of Multiple Mobile Robots

Mobile Robot embedded Architecture Based on CAN

Adaptive Humanoid Robot Arm Motion Generation by Evolved Neural Controllers

Fuzzy Logic Based Path Tracking Controller for Wheeled Mobile Robots

Gilbert Peterson and Diane J. Cook University of Texas at Arlington Box 19015, Arlington, TX

Autonomous Stair Climbing Algorithm for a Small Four-Tracked Robot

Learning Attentive-Depth Switching while Interacting with an Agent

MOBILE ROBOT WALL-FOLLOWING CONTROL USING A BEHAVIOR-BASED FUZZY CONTROLLER IN UNKNOWN ENVIRONMENTS

A Robust Neural Robot Navigation Using a Combination of Deliberative and Reactive Control Architectures

Safe and Efficient Autonomous Navigation in the Presence of Humans at Control Level

A Comparative Study on different AI Techniques towards Performance Evaluation in RRM(Radar Resource Management)

Swarm Intelligence W7: Application of Machine- Learning Techniques to Automatic Control Design and Optimization

Dr. Wenjie Dong. The University of Texas Rio Grande Valley Department of Electrical Engineering (956)

A Divide-and-Conquer Approach to Evolvable Hardware

Unit 1: Introduction to Autonomous Robotics

An Improved Path Planning Method Based on Artificial Potential Field for a Mobile Robot

Creating a 3D environment map from 2D camera images in robotics

INFORMATION AND COMMUNICATION TECHNOLOGIES IMPROVING EFFICIENCIES WAYFINDING SWARM CREATURES EXPLORING THE 3D DYNAMIC VIRTUAL WORLDS

Real-Time Bilateral Control for an Internet-Based Telerobotic System

Implicit Fitness Functions for Evolving a Drawing Robot

Vision System for a Robot Guide System

Verified Mobile Code Repository Simulator for the Intelligent Space *

Hybrid architectures. IAR Lecture 6 Barbara Webb

1, 2, 3,

RescueRobot: Simulating Complex Robots Behaviors in Emergency Situations

Introduction to Robotics

Simulation of Synchronous Machine in Stability Study for Power System: Garri Station as a Case Study

CSCI 445 Laurent Itti. Group Robotics. Introduction to Robotics L. Itti & M. J. Mataric 1

COOPERATIVE STRATEGY BASED ON ADAPTIVE Q- LEARNING FOR ROBOT SOCCER SYSTEMS

Behavior Emergence in Autonomous Robot Control by Means of Feedforward and Recurrent Neural Networks

Associated Emotion and its Expression in an Entertainment Robot QRIO

Decentralised Cooperative Control of a Team of Homogeneous Robots for Payload Transportation

Randomized Motion Planning for Groups of Nonholonomic Robots

Motion Control of a Three Active Wheeled Mobile Robot and Collision-Free Human Following Navigation in Outdoor Environment

Artificial Neural Network based Mobile Robot Navigation

Transcription:

APPLICATION OF FUZZY BEHAVIOR COORDINATION AND Q LEARNING IN ROBOT NAVIGATION Handy Wicaksono 1,2, Prihastono 1,3, Khairul Anam 4, Rusdhianto Effendi 2, Indra Adji Sulistijono 5, Son Kuswadi 5, Achmad Jazidie 2, Mitsuji Sampei 6 1 Department of Electrical Engineering, Sepuluh Nopember Institute of Surabaya, Indonesia 2 Department of Electrical Engineering, Petra Christian University, Surabaya, Indonesia (Tel : +62-31-2983115; E-mail : handy@petra.ac.id) 3 Department of Electrical Engineering, University of Bhayangkara, Surabaya, Indonesia 4 Department of Electrical Engineering, University of Jember, Jember, Indonesia 5 Department of Mechatronics, Electronics Engineering Polytechnic Institute of Surabaya, Surabaya, Indonesia 6 Department of Mechanical and Environmental Informatics, Tokyo Institute of Technology, Tokyo, Japan ABSTRACT Behavior based architecture is widely used in mobile robot because it makes the robot response faster. If robot only works to achieve simple task, it can use some primitive behaviors. But when the task is getting more complex, the behavior coordination is needed. In order to construct this coordinator, fuzzy logic can be applied as Fuzzy Behavior Coordinator (FBC). By using FBC, it can be seen from simulation that robot has smoother movement and shorter time to reach target in its navigation. When the robot is in new and uncertain field, it needs to learn. Q learning can be used to give intelligence behavior to robot. Robot can construct its own behavior by learning from its environment. By applying Q learning, the shortest path to reach target will be obtained after some episodes of robot training. Keywords: behavior-based robotics, fuzzy behavior coordinator, Q learning 1 INTRODUCTION Robot architecture that had been developed in early days is deliberative architecture. The robot senses its environment, thinks by creating a world model, and after that robot will act. This approach has these disadvantages : planning process (world modelling) takes a long time, no direct relationship between sensor and actuator, and a very sophisticated controller is needed [4]. To overcome them, Brooks [7] suggest reactive architecture (sometimes called behavior based architecture). In this approach, we do not need a world model mathematically. The real environment is the only model which needed by robot. Another advantage is all behaviors run in parallel, simultaneous, and asynchronous way. So the sophisticated controller isn t need anymore, and it can be replaced with a small-low cost microcontroller [4]. In behavior based architecture, the problem will arise when there are many behaviors in a robot. So it has to decide which behavior will be executed at one time. Because of that, robot must have behavior coordinator (sometimes called arbiter). First approach that is also suggested by Brooks [7] is Subsumption Architecture that can be classified as competitive method. In this method, at one time, there is only one behavior that applied in robot. It is very simple method and it gives the fast performance result, but it has disavantage of nonsmooth response and inaccuracy in robot movement. To overcome competitive method weakness, Arkin [2, 3] suggest Potential Fields Method that can be classified as cooperative method. In this method, at one time, there can be more than one behavior that applied in robot, so every behavior has contribution in robot s action. It results in a smoother response and more accurate robot movement. In the other hand, this method is more complex than competitive method. The complete list of behavior coordination method can be found in [12]. Another cooperative method is by using Fuzzy Logic as tool for coordinate the behaviors. There are many approaches to apply Fuzzy Behavior Based Control. Abreu [1] uses fuzzy behavior arbitration that combines some behaviors in its defuzzyfication process to control autonomous vehicle. Thongchai [15] use fuzzy to process each behavior, but use priority based arbitration to combine its behaviors. Saffiotti [13] suggest context dependent blending to improve ordinary <Page number for first page>

<even page number > The 5 th International Conference on Information & Communication Technology and Systems Fuzzy Logic weakness because of behaviors conflict. Beside the correct architecture and behavior coordination method, in order to overcome the unplanned thing in uncertain world, the robot will need a right learning mechanism. In supervised learning mechanism, teacher is needed, while in unsupervised learning the agent has to learn by itself. So the latter is more suitable for robotic application. Reinforcement learning is an unsupervised learning method which can learn from reward that given by environment directly (online) [8]. There are various methods to solve reinforcement learning problem. Q learning algorithm is the famous one. The advantage of Q learning is its off policy characteristic, simple algortihm, and it is convergent to optimal policy [16]. 2 MODEL, ANALYSIS, DESIGN, AND IMPLEMENTATION Figure 2. Subsumption architecture example for navigation The method above is very simple and fast, but it still has disavantage of non-smooth response and inaccuracy in robot movement. The cooperative method overcomes these weakness. In this method, at one time, there can be more than one behavior that applied in robot, so every behavior have contribution in robot s action. Arkin [2] suggest the potential fields method, which every object will be described as vector that have magnitude and direction. The result behavior is mixing between each behavior. The motor scheme for this method appear in the figure below. 2.1 Behavior Coordination Methods In behavior based robotics approach, methods pf behavior coordination are significant. The designer needs to know how robot coordinate its behaviors and take the action in the real world. There are two approaches : competitive and cooperative. In competitive method, at one time, there is only one behavior that applied in robot. The first suggestion in this class is Subsumption Architecture that suggested by Brooks [7]. This method divides behaviors to many levels, where the higher level behavior have higher priorities too. So it can subsume the lower level ones. The layered control system figure is given below. Figure 1. Layered control system [7] Figure 3. Motor schema for potential field method [3] This method results in a smoother response and more accurate robot movement. In the other hand, this method is more complex. It also can result in slow movement, it needs long time to reach the target and it may fail to avoid trap. Another cooperative method is by using Fuzzy Logic as tool to combine the behaviors. After input from sensor is fuzzified, then it will be processed in rule base with other fuzzified input. Rule base represent the behavior. Defuzzification process will combine the behaviors [1]. The scheme of fuzzy behavior coordination is shown below. The example of behavior coordination that using Subsumption This architecture can be seen below.

<document ID-Title-first author s name> < odd page number> Figure 4. Scheme of fuzzy behavior coordination 2.2 Robot Learning 2.2.1 Reinforcement Learning Mataric [10] says that an excellent behavior selection is important in behavior based control, because it decides which behavior that will be applied in certain time. This problem can be formulated in Masalah ini reinforcement learning (RL) frame work by deciding which policy that will map the state into the behavior which will maximize the received reward in agent s liftime. RL is classified as unsupervised learning, so its agent doesn t need any supervising to accomplish its task. Glorennec [8] also states that in supervised learning, which is mentioned too as learning from teacher, the learning system know the system error every time for each related input and output vector. This error can be used to modify the learning parameters. While in RL, also can be mentioned as learning from critic, the accepted signal is reward (positive, negative, or neutral) from behavior. This signal indicates what the robot has to do without any information about how to do it. So it can be said that RL is online and it can be appled directly to the robot. The general process scheme can be shown below. In many learning applications, the goal of RL is constructing a control policy which wiil map the discrete input area to discrete output area so the maximal cumulative reward can be achieved. The most simple and wellknown learning algorithm in robotic application is Q learning [16]. Q learning has some advantages. It is off policy (robot can follow any policy to produce optimal solution), it has simple algorithm, and it will give convergent result in an optimal policy. Here is the simplified algorithm of Q learning [14]. Initialize Q(s,a) arbitralily Repeat (for each episode) : Initialize s Repeat (for each step of episode): Choose a from s using policy derived from Q (e.g., - greedy) Take action a, observe r, s Apply [ r + γ max Q( s', a') Q( s, )] Q( s, a) Q( s, a) + α a' a where : action) s s ; until s is terminal Q(s,a) : component of Q table (state, s : state s : next state a : action a : next action r : reward α : learning rate γ : discount factor 2.3. Behaviors Description 2.3.1 Behaviors in Subsumption Architecture Figure 5. General process scheme reinforcement learning [11] Implementation of RL in robot application is started by Obelix, robot that push the box to the certain location [9]. RL also applied to another kid of robot, such as an underwater robot [11], a group of mobile robots which are used to do foraging [10]. 2.2.2 Q Learning There are some behaviors that robot should have in order to succeed in autonomous navigation. They are : Obstacle avoidance Search the target Wandering The scheme of these architectures can be seen on Figure 2. Wandering behavior is the lowest level behavior (in Subsumption Architecure) that will move the robot all around the arena in order to find the target. If the robot sees the obstacles, then it will avoid them because of presence of obstacle

<even page number > The 5 th International Conference on Information & Communication Technology and Systems avoidance behavior. And then when the robot find the target (a light source), it will move closer to the robot. After the distance is close enough, then the robot will stop. 2.3.2 Fuzzy Behaviors Behaviors that use in Fuzzy Behavior (FB) dan Modified Fuzzy Behavior (MFB) are almost same. Here are the descriptions of them. 2.3.2.1. Obstacle Avoidance Behavior Robot uses two distance sensors (which is put in left and right). Here are the triangular membership function (MF) for the sensor (their configuration are same). They can be divided as : small, medium, and big. Figure 6. Membership function of distance sensors MF of two sensors will be combined by a rule base of left & right motor. The rule bases are not the same, but one is mirror of the other. Here are the rule bases for them. Table 1. Rule base 1 for left motor Right D.S S M B Left D. S S P P N M Z Z N B P P N Table 2. Rule base 2 for right motor Right D.S. S M B Left D.S. S P Z P M P P P B N N N 2.3.2.2. Search Target Behavior Figure 7. Membership function of light sensors The MF of two sensors will be combined by a rule base of left & right motor. The rule bases are not the same, but one is mirror of the other. Here are the rule bases for them. Table 3. Rule base 3 for left motor Right L.S. S M B Left L.S. S Z Z P M Z Z P B N Z P Table 4. Rule base 4 for right motor Right L.S. S M B Left L.S. S Z Z N M Z Z Z B P P P 2.3.3 Q Learning Behavior In this paper, Q learning will be applied only in Search Target behavior. In the beginning the robot will explore its environment and learn form it. In this exploration phase, the robot will move randomly and (in the same time) update its Q table. After some times, the value of Q table s component will be optimal and steady. Now it s time to switch to exploitation phase, when the robot only use the optimal value of Q table to navigate to the target. 3. RESULT Robot used here has two distance sensors and two light sensors. It only uses two motors. The complete parts of robot can be shown below. Robot uses two light sensors (which is put in left and right). Here are the triangular membership function (MF) for the sensor (their configuration are same). They can be divided as : small, medium, and big.

<document ID-Title-first author s name> < odd page number> Right Distance Sensor Right Light Sensor Right Motor Body of robot Figure 8. Robot used in Simulation Left Distance Sensor Left Light Sensor Left Motor Webbots 5.5.2 software from Cyberbotics has been used to simulate and test the performance of robot. 3.1 Behavior Coordination Simulation Simulations that has been done will observe the performance of these behavior coordination methods : 1. Subsumption Architecture (SA) 2. Fuzzy Behavior Coordinator (FBC) The simulation will measure how many time that robot used to reach the target, the performance of robot movement that can be seen by the robot s trajectory, and how well robot in avoid obstacle. 3.1.1 Target Reaching Simulation This Simulation will be done to measure how many times that needed by robot to travel from the start position until the target is found. There are three start positions in the arena, that can be seen in figure below. 0.43, 2.03) 1:09 m 0:24 m 2 (0.42, 0, 0.41, Reach in Reach in 0.93) 1:05 m 0:13 m 3 (-0.41, 0,- 0.42, 4.13) Reach in 1:45 m Reach in 1:21 m From the table we can see that SA needs most time (1:09 m, 1:05 m, 1:45 m) to reach target from each position. It is happened because SA is a competitive coordinator type, so there is a chance to ignore the search target behavior when other behavior is active. The rough response from the robot can increase the time to reach target. FBC gives better result because evary behavior gives contribution in robot s decision making. But it tends to give slow result, because every behavior will processed in fuzzy inference engine. MFBC come with a fastest result, because the search target behavior is not a fuzzy behavior, and it considers as more important behavior than other. 3.1.2 Movement Performance Simulation In this simulation, robot s trajectory will be observed in order to understand robot s movement. Here are the simulation figure of robot with SA. Figure 9. Arena for the simulations Each simulation has been done 5 times, and the best result is shown in table below. Table 5. Target reaching test result Pos Start position of Robot Time to reach target (minutes) (x,y,z,teta) SA FBC 1 (-0.41, 0, Reach in Reach in Figure 10. Movement performance simulation result for robot uses SA From the figure we can see that SA gives a sharp turn of robot. Look at the circle with the dash-line. This sharp turn happened because of the immediate change in robot s behavior. It has to switch from one behavior to another directly. So it make the robot s movement are little bit rough, although it gives faster movement than another approach. While the simulation figure of robot with FBC can be seen below.

<even page number > The 5 th International Conference on Information & Communication Technology and Systems turn right. After that, the robot will apply five actions : forward, turn somewhat left, turn left, turn somewhat right, and turn right. 3.2.1 Simulation with three actions The robot will start in coordinat (x, y, z, teta) : (0.4, 0, 0, 1.8). Here is the simulation result in exploration stage. Figure 11. Movement performance simulation result for robot uses FBC From the figure above we can see that FBC gives smoother response, slighter turn. Once again, it is marked by the circle with the dash-line. This is happened because fuzzy process that mix all robot s behavior. So there is no immediate change between behaviors. In overall, robot movement is very smooth but it give the slowest motion. 3.1.3 Trap Simulation The last simulation that will be held here is called trap simulation. The robot will be put into the trap to see how good their performance to escape from the trap. Here is the result figure. Figure 13. Exploration Simulation Result 1 After 5 times simulation, the Q table components values are already optimal. Here is the final Q table. Q = [ -1.7-1.7-1.7 0 0 0 3.33 3.33 3.33] After that robot enter the exploitation stage. Here is the simulation result Figure. 12 Trap performance simulation result From the figure we can see that SA give better performance in escaping from the trap. SA robot succeed to escape, but FBC robot are failed during the simulation. Both of them are trapped. 3.2 Q learning behavior simulation In this simulation, Q learning will be applied just in Search Target behavior. In the beginning, the robot only uses three actions : forward, turn left and Figure 14. Exploitation Simulation Result 1 From the figure above it can be seen that after some training form environment, the robot can achieve its target accurately in short time. 3.2.1 Simulation with five actions

<document ID-Title-first author s name> < odd page number> The robot will start in coordinat (x, y, z, teta) : (0.4, 0, 0, 1.8). Here is the simulation result in exploration stage. smoother movement and shorter time to reach the target than the Subsumption Architecture. But the FBC still has weaknesses in avoiding the trap. It also can be seen from the simulation result that by adding learning capability (using Q learning) in behavior that using five action, the robot can achieve the target accurately in short time, with less time in exploration. 5. ACKNOWLEDGEMENT Figure 15. Exploration Simulation Result 2 After 3 times simulation, the Q table components values are already optimal. Here is the final Q table. Q = [ -1.7-1.7-1.7-1.7-1.7 0 0 0 0 0 3.33 3.33 3.33 3.33 3.33] After that robot enter the exploitation stage. Here is the simulation result. This work is being supported by Japan International Cooperation Agency (JICA) through Technical Cooperation Project for Research and Education Development on Information and Communication Technology in Sepuluh Nopember Institute of Technology (PREDICT - ITS). REFERENCES [1] Abreu A. (1999) Fuzzy Behaviors and Behavior Arbitration in Autonomous Vehicles. Proceedings of the Portuguese Meeting in Artificial Intelligence EPIA99, volume 1695 of LNC. [2] Arkin R.C. (1987) Motor Schema Based Navigation for a Mobile Robot : an Approach to Programming by Behavior. IEEE Int. Conf. on Robotics and Automation, pp. 264-271. [3] Arkin R.C. (1998) Behavior-Based Robotics,. England : Bradford Books. [4] Asadpour M., and Siegwart R. (2004) Compact Q-Learning for Micro-robots with Processing Constraints. Journal of Robotics and Autonomous Systems, Vol. 48, No. 1, pp. 49-61. Figure 16. Exploitation Simulation Result 2 From this simulation it can be seen that the robot achieve its optimal value in less time simulation (three compare to five), and it is fast to reach the target accurately. 4. CONCLUSION AND DISCUSSION This paper described the application of fuzzy logic and Q learning in mobile robot navigation system. From the simulation result, it can be seen that Fuzzy Behavior Coordinator gives the [6] Bekey G. A. (2005) Autonomous Robot : From Biological Inspiration to Implementation and Control.Massachusetts : MIT Press. [7] Brooks R. (1986) A Robust Layered Control System For a Mobile Robot. IEEE Journal of Robotics and Automation, vol. 2, no. 1, pp. 14 23. [8] Glorennec, P. Y. (2000) Reinforcement Learning : An Overview. Proceedings of European Symposium on Intelligent Techniques. Aachen, Germany. [9]Mahadevan, S. dan Connell, J. (1991) Automatic Programming of Behavior Based using Reinforcement Learning. Proceeding of the Eighth

<even page number > The 5 th International Conference on Information & Communication Technology and Systems International Workshop on Machine Learning, ppl. 328-332. [10] Mataric, M.J. (1996) Reinforcement Learning in Multirobot Domain. Autonomous Robots, vol. 4, no. 1, hal. 73-83. [11] Perez M.C. (2003) A Proposal of Behavior Based Control Architecture with Reinforcement Learning for an Autonomous Underwater Robot, Tesis Ph.D., University of Girona, Girona. [12] Pirjanian P. (1999) Behavior coordination mechanisms : State-of-the-art, Technical Report IRIS-99-375, Univ. Southern California. [13] Saffiotti A. (1997) Fuzzy Logic in Autonomous Robotics : Behavior Coordination. Proceedings of the 6 th IEEE Int. Conf. on Fuzzy Systems. pp. 573 478, Barcelona, Spain. [14] Sutton, R.S., dan Barto,A.G. (1998), Reinforcement Learning, an introduction. Massachusets : MIT Press. [15] Tongchai S, Suksakulchai S., Wilkes D.M., and Sarkar N.(2000) Sonar Behavior-Based Fuzzy Control for a Mobile Robot. Proc. of IEEE International Conference on Systems, Man and Cybernetics. Nashville, Tennesse. [16] Watkins, C., Dayan, P.(1992) Q- learning,thechnical Note. Machine Learning, Vol 8, pp.279-292.