Abstract
A good deal of progress has been made in the past few years in the design and implementation of control programs for autonomous agents. A natural extension of this work is to consider solving dificult tasks with teams of cooperating agents. Our interest in this area is motivated in part by our involvement in a Navy-sponsored micro air vehicle (MAV) project in which the goal is to solve dificult surveillance tasks using a large team of small inexpensive autonomous air vehicles rather than a few expensive piloted vehicles. Our approach to developing control programs for these MAVs is to use evolutionary computation techniques to evolve behavioral rule sets. In this paper we describe our architecture for achieving this, and we present some of our initial results.
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
R.C. Arkin and T. Balch. Cooperative multiagent robotic systems. In David Kortenkamp, R.P. Bonasso, and R. Murphy, editors, Artificial Intelligence and Mobil Robots, Cambridge, MA, 1998. MIT/AAAI Press.
T. Balch. Learning roles: Behavioral diversity in robot teams. In Collected Papers from the 1997 AAAI Workshop on Multiagent Learning, pages 7–12, Cambridge, MA, July 1997. AAAI Press.
J. Grefenstette. Learning rules from simulation models. In Proceedings of the 1989 International Association of Knowledge Engineers Conference, pages 117–122, Washington, DC, 1989. IAKE.
J. Holland. Escaping brittleness: The possibilities of general-purpose learning algorithms applied to parallel rule-based systems. In J. Carbonell R. Michalski and T. Mitchell, editors, Machine Learning: An Artificial Intelligence Approach, Los Altos, 1996. Morgan Kaufman.
K. A. De Jong and W. M. Spears. Learning concept classification rules using genetic algorithms. In IJCAI 91, Proceedings of the 12th International Conference on Artificial Intelligence, pages 651–656, Sydney, Australia, 1991. Morgan Kaufmann Publishers, Inc.
C. Kube and H. Zhang. Collective robotics: From social insects to robots. Adaptive Behavior, 2(2):189–219, 1993.
M. Mataric. Learning to behave socially. In D. Cliff, P. Husbands, J.A. Meyers, and S. Wilson, editors, From Animals to Animats 3 (Third International Conference on Simulation of Adaptive Behavior), pages 453–462. MIT Press, 1994.
A. Schultz and J. Grefenstette. Using a genetic algorithm to learn behaviors for autonomous vehicles. In Proceedings of the American Institute of Aeronautics and Astronautics Guidance, Navigation and Control Conference, pages 739–749, Hilton Head, SC, 1992. AIAA.
S. Smith. Flexible learning of problem solving heuristics through adaptive search. In William Kaufman, editor, Proceeding of the Eighth International Joint Confe-rence on Artificial Intelligence, pages 422–425, Karlsruche, Germany, 1983.
T. Soule and J.A. Foster. Effects of code growth and parsimony pressure on popu-lations in genetic programming. Evolutionary Computation, 6(4):293–309, 1999.
A. Wu, A.C. Schultz, and A. Agah. Evolving control for distributed micro air ve-hicles. In IEEE Computational Intelligence in Robotics and Automation Engineers Conference, 1999.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2000 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Bassett, J.K., De Jong, K.A. (2000). Evolving Behaviors for Cooperating Agents. In: RaÅ›, Z.W., Ohsuga, S. (eds) Foundations of Intelligent Systems. ISMIS 2000. Lecture Notes in Computer Science(), vol 1932. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-39963-1_17
Download citation
DOI: https://doi.org/10.1007/3-540-39963-1_17
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-41094-2
Online ISBN: 978-3-540-39963-6
eBook Packages: Computer ScienceComputer Science (R0)