Academia.eduAcademia.edu

Outline

Model for Team Formation for Reformation in Multiagent Systems

2002

Abstract

Team formation, i.e., allocating agents to roles within a team or subteams of a team, and the reorganization of a team upon team member failure or arrival of new tasks are critical aspects of teamwork. We focus in particular on a technique called "Team Formation for Reformation", i.e., teams formed with lookahead to minimize costs of reformation. Despite significant progress, research in multiagent team formation and reorganization has failed to provide a rigorous analysis of the computational complexities of the approaches proposed or their degree of optimality. This shortcoming has hindered quantitative comparisons of approaches or their complexity-optimality tradeoffs, e.g., is the team reorganization approach in practical teamwork models such as STEAM optimal in most cases or only as an exception? To alleviate these difficulties, this paper presents R-COM-MTDP, a formal model based on decentralized communicating POMDPs, where agents explicitly take on and change roles to (re)form teams. R-COM-MTDP significantly extends an earlier COM-MTDP model, by analyzing how agents' roles, local states and reward decompositions gradually reduce the complexity of its policy generation from NEXP-complete to PSPACE-complete to P-complete. We also encode key role reorganization approaches (e.g., STEAM) as R-COM-MTDP policies, and compare them with a locally optimal policy derivable in R-COM-MTDP, thus, theoretically and empirically illustrating the complexityoptimality tradeoffs.

References (11)

  1. Bernstein, D. S.; Zilberstein, S.; and Immerman, N. 2000. The complexity of decentralized control of MDPs. In UAI.
  2. Boutilier, C. 1996. Planning, learning & coordination in multiagent decision processes. In TARK.
  3. Grosz, B., and Kraus, S. 1996. Collaborative plans for complex group action. Artificial Intelligence 86(2):269- 357. Kitano, H.; Tadokoro, S.; and Noda, I. 1999. Robocup- rescue: Search and rescue for large scale disasters as a domain for multiagent research. In IEEE Conference SMC.
  4. Marsella, S.; Tambe, M.; and Adibi, J. 2001. Experiences acquired in the design of robocup teams: A comparison of two fielded teams. JAAMAS 4:115-129.
  5. Modi, P. J.; Jung, H.; Tambe, M.; Shen, W.-M.; and Kulkarni, S. 2001. A dynamic distributed constraint sat- isfaction approach to resource allocation. In Constraints Proc (CP).
  6. Peshkin, L.; Meuleau, N.; Kim, K.-E.; and Kaelbling, L. 2000. Learning to cooperate via policy search. In UAI.
  7. Pynadath, D., and Tambe, M. 2002. Multiagent teamwork: Analyzing the optimality complexity of key theories and models. In AAMAS.
  8. Shehory, O., and Kraus, S. 1998. Methods for task allo- cation via agent coalition formation. Artificial Intelligence 101(1-2):165-200.
  9. Tambe, M. 1997. Towards flexible teamwork. JAIR 7:83- 124. Tidhar, G.; Rao, A.; and Sonenberg, E. 1996. Guided team selection. In ICMAS.
  10. Xuan, P.; Lesser, V.; and Zilberstein, S. 2001. Communi- cation decisions in multiagent cooperation. In Agents.
  11. Yoshikawa, T. 1978. Decomposition of dynamic team deci- sion problems. Proceedings of the IEEE AC-23(4):627-632.