Online Evolution for Cooperative Behavior in Group Robot Systems

  • Lee, Dong-Wook (Division for Applied Robot Technology, Korea Institute of Industrial Technology) ;
  • Seo, Sang-Wook (School of Electrical and Electronics Engineering, Chung-Ang University) ;
  • Sim, Kwee-Bo (School of Electrical and Electronics Engineering, Chung-Ang University)
  • Published : 2008.04.30

Abstract

In distributed mobile robot systems, autonomous robots accomplish complicated tasks through intelligent cooperation with each other. This paper presents behavior learning and online distributed evolution for cooperative behavior of a group of autonomous robots. Learning and evolution capabilities are essential for a group of autonomous robots to adapt to unstructured environments. Behavior learning finds an optimal state-action mapping of a robot for a given operating condition. In behavior learning, a Q-learning algorithm is modified to handle delayed rewards in the distributed robot systems. A group of robots implements cooperative behaviors through communication with other robots. Individual robots improve the state-action mapping through online evolution with the crossover operator based on the Q-values and their update frequencies. A cooperative material search problem demonstrated the effectiveness of the proposed behavior learning and online distributed evolution method for implementing cooperative behavior of a group of autonomous mobile robots.

Keywords

References

  1. L. E. Parker, "ALLIANCE: An architecture for fault-tolerant multirobot cooperation," IEEE Trans. on Robotics and Automation, vol. 14, no. 2, pp. 220-240, April 1998 https://doi.org/10.1109/70.681242
  2. P. J. 't Hoen, K. Tuyls, L. Panait, S. Luke, and J. A. La Poutre, "An overview of cooperative and competitive multiagent learning," Learning and Adaption in Multi-Agent System, LNAI 3898, pp. 1-46, 2006 https://doi.org/10.1007/11691839_1
  3. H. Asama, "Perspective of distributed autonomous robotic systems," Distributed Autonomous Robotic Systems 5, H. Asama, T. Arai, T. Fukuda, T. Hasegawa (Eds.), Springer, pp. 3-4, 2002
  4. T. Arai, E. Pagello, and L. E. Parker, "Advances in multirobot systems," IEEE Trans. on Robotics and Automation, vol. 18, no. 5, pp. 655-661, October 2002 https://doi.org/10.1109/TRA.2002.806024
  5. R. S. Sutton and A. G. Barto, Reinforcement Learning: An Introduction, MIT Press, Cambridge, MA, 1998
  6. J. S. R. Jang, C. T. Sun, and E. Mizutani, Neuro-Fuzzy and Soft Computing, Prentice Hall, 1997
  7. C. J. C. H. Watkins and P. Dayan, "Q-learning," Machine Learning, vol. 8, pp. 279-292, 1992
  8. L. P. Kaelbling, "On reinforcement learning for robotics," Proc. of Int. Conf. on Intelligent Robot Systems, pp. 1319-1320, 1996
  9. S. O. Kimbrough and M. Lu, "Simple reinforcement learning agents: Pareto beats Nash in an algorithmic game theory study," Information Systems and E-Business Management, vol. 3, no. 2, pp. 1-19, March 2005 https://doi.org/10.1007/s10257-003-0024-0
  10. L. E. Parker, C. Touzet, and D. Jung, "Learning and adaptation in multi-robot teams," Proc. of Eighteenth Symposium on Energy Engineering Sciences, pp. 177-185, 2000
  11. M. Nakamura, N. Yamashiro, and Y. Gong, "Iterative parallel and distributed genetic algorithms with biased initial population," Proc. of Congress on Evolutionary Computation, vol. 2, pp. 2296-2301, 2004
  12. A. L. Jaimes and C. A. Coello, "MRMOGA: A new parallel multi-objective evolutionary algorithm based on the use of multiple resolutions," Concurrency and Computation: Practice and Experience, vol. 19, no. 4, pp. 397-441, March 2007 https://doi.org/10.1002/cpe.1107
  13. T. Fukuda and T. Ueyama, Cellular Robotics and Micro Robotic System, World Scientific, 1994