Academia.eduAcademia.edu

Cooperative Multiagent Systems

description31 papers
group35 followers
lightbulbAbout this topic
Cooperative Multiagent Systems (CMAS) refer to a field of study focused on the design, analysis, and implementation of systems composed of multiple autonomous agents that work collaboratively to achieve common goals, optimize performance, and solve complex problems through communication, coordination, and negotiation among agents.
lightbulbAbout this topic
Cooperative Multiagent Systems (CMAS) refer to a field of study focused on the design, analysis, and implementation of systems composed of multiple autonomous agents that work collaboratively to achieve common goals, optimize performance, and solve complex problems through communication, coordination, and negotiation among agents.
In this thesis, we study the problem of the optimal decentralized control of a partially observed Markov process over a finite horizon. The mathematical model corresponding to the problem is a decentralized POMDP (DEC-POMDP). Many... more
In cooperative multi-agent sequential decision making under uncertainty, agents must coordinate to find an optimal joint policy that maximises joint value. Typical algorithms exploit additive structure in the value function, but in the... more
We consider multiagent systems situated in unpredictable environments. Agents viewed as abductive logic programs with abducibles being literals the agent could sense or receive from other agents, must cooperate to provide answers to users... more
The growing use of autonomous tractor fleets with detachable implements presents complex logistical challenges in agriculture. Current systems often rely on simple heuristics and avoid implement swapping, limiting efficiency. A central... more
Bayesian games can be used to model single-shot decision problems in which agents only possess incomplete information about other agents, and hence are important for multiagent coordination under uncertainty. Moreover they can be used to... more
This paper formulates the optimal decentralized control problem for a class of mathematical models in which the system to be controlled is characterized by a finite-state discrete-time Markov process. The states of this internal process... more
We address a long-standing open problem of reinforcement learning in decentralized partially observable Markov decision processes. Previous attempts focussed on different forms of generalized policy iteration, which at best led to local... more
This paper formulates the optimal decentralized control problem for a class of mathematical models in which the system to be controlled is characterized by a finite-state discrete-time Markov process. The states of this internal process... more
Bayesian games can be used to model single-shot decision problems in which agents only possess incomplete informa- tion about other agents, and hence are important for mul- tiagent coordination under uncertainty. Moreover they can be used... more
HAL is a multi-disciplinary open access archive for the deposit and dissemination of scientific research documents, whether they are published or not. The documents may come from teaching and research institutions in France or abroad, or... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
Recent years have seen significant advances in techniques for optimally solving multiagent problems represented as decentralized partially observable Markov decision processes (Dec-POMDPs). A new method achieves scalability gains by... more
We address a long-standing open problem of reinforcement learning in decentralized partially observable Markov decision processes. Previous attempts focussed on different forms of generalized policy iteration, which at best led to local... more
To compare the ability of agents to learn in open worlds, we need a framework with clear definitions of open world environments and how they can vary. This paper provides such a framework, proposing clear scientific definitions for open... more
This work examines the mean-square error performance of diffusion stochastic algorithms under a generalized coordinate-descent scheme. In this setting, the adaptation step by each agent is limited to a random subset of the coordinates of... more
Characterization of the uncertainty in robotic manipulators is the focus of this paper. Based on the random matrix theory (RMT), we propose uncertainty characterization schemes in which the uncertainty is modeled at the macro (system)... more
Loosely interconnected cooperative systems such as cable robots are particularly susceptible to uncertainty. Such uncertainty is exacerbated by addition of the base mobility to realize reconfigurability within the system. However, it also... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
We formulate an approach to multiagent metareasoning that uses organizational design to focus each agent's reasoning on the aspects of its local problem that let it make the most worthwhile contributions to joint behavior. By... more
Many challenges in today's society can be tackled by distributed open systems. This is particularly true for domains that are commonly perceived under the umbrella of smart cities, such as intelligent transportation, smart energy grids,... more
We develop a Multi-Agent Reinforcement Learning (MARL) method to learn scalable control policies for target tracking. Our method can handle an arbitrary number of pursuers and targets; we show results for tasks consisting up to 1000... more
Recent years have seen significant advances in techniques for optimally solving multiagent problems represented as decentralized partially observable Markov decision processes (Dec-POMDPs). A new method achieves scalability gains by... more
Resolving multiagent team decision problems, where agents share a common goal, is challenging since the number of states and joint actions is exponential with the number of agents. Even if the resolution of such problems is theoretically... more
An information agent is viewed as a deductive database consisting of 3 parts: • an observation database containing the facts the agent has observed or sensed from its surrounding environment. • an input database containing the information... more
We consider multiagent systems situated in unpredictable environments. Agents viewed as abductive logic programs with abducibles being literals the agent could sense or receive from other agents, must cooperate to provide answers to users... more
An information agent is viewed as a deductive database consisting of 3 parts: • an observation database containing the facts the agent has observed or sensed from its surrounding environment. • an input database containing the information... more
We consider a class of sequential decision-making problems under uncertainty that can encompass various types of supervised learning concepts. These problems have a completely observed state process and a partially observed modulation... more
We consider multiagent systems situated in unpredictable environments. Agents viewed as abductive logic programs with abducibles being literals the agent could sense or receive from other agents, must cooperate to provide answers to users... more
An information agent is viewed as a deductive database consisting of 3 parts: • an observation database containing the facts the agent has observed or sensed from its surrounding environment. • an input database containing the information... more
We consider multiagent systems situated in unpredictable environments. Agents viewed as abductive logic programs with abducibles being literals the agent could sense or receive from other agents, must cooperate to provide answers to users... more
Recent years have seen significant advances in techniques for optimally solving multiagent problems represented as decentralized partially observable Markov decision processes (Dec-POMDPs). A new method achieves scalability gains by... more
Bayesian games can be used to model single-shot decision problems in which agents only possess incomplete informa- tion about other agents, and hence are important for mul- tiagent coordination under uncertainty. Moreover they can be used... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
This paper introduces a probabilistic algorithm for multi-robot decision-making under uncertainty, which can be posed as a Decentralized Partially Observable Markov Decision Process (Dec-POMDP). Dec-POMDPs are inherently synchronous... more
Planning under uncertainty for multiagent systems can be formalized as a decentralized partially observable Markov decision process. We advance the state of the art for optimal solution of this model, building on the Multiagent A*... more
A recent insight in the field of decentralized partially observable Markov decision processes (Dec-POMDPs) is that it is possible to convert a Dec-POMDP to a non-observable MDP, which is a special case of POMDP. This technical report... more
Yi Xiong Department of Systems Engineering and Engineering Management, The Chinese University of Hong Kong, Hong Kong, China, yxiong@se.cuhk.edu.hk Ningyuan Chen Rotman School of Management, University of Toronto, Toronto, Canada,... more
This article presents the state-of-the-art in optimal solution methods for decentralized partially observable Markov decision processes (Dec-POMDPs), which are general models for collaborative multiagent planning under uncertainty.... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
Planning under uncertainty for multiagent systems can be formalized as a decentralized partially observable Markov decision process. We advance the state of the art for optimal solution of this model, building on the Multiagent A*... more
There has been substantial progress on algorithms for single-agent sequential decision making using partially observable Markov decision processes (POMDPs). A number of efficient algorithms for solving POMDPs share two desirable... more
The main objective of emergency medical assistance (EMA) services is to attend patients with sudden diseases at any possible location within an area of influence. This usually consists in providing "in situ" assistance and, if necessary,... more
Decentralized POMDPs provide a rigorous framework for multi-agent decision-theoretic planning. However, their high complexity has limited scalability. In this work, we present a promising new class of algorithms based on probabilistic... more
This article applies a performance metric to the multi-robot patrolling task to more efficiently distribute patrol areas among robot team members. The multi-robot patrolling task employs multiple robots to perform frequent visits to known... more
Abstract. Stroke is the third highest cause of mortality and the first cause of disabled people in western countries. A significant number of the people who survive live with serious physical and psychological disabil-ities and require... more
This work examines the mean-square error performance of diffusion stochastic algorithms under a generalized coordinate-descent scheme. In this setting, the adaptation step by each agent is limited to a random subset of the coordinates of... more
In this paper we consider the problem of controlling multiple robots manipulating and transporting an object in three dimensions via cables. We develop robot configurations that ensure static equilibrium of the object at a desired pose... more
In this paper we consider the problem of controlling multiple robots manipulating and transporting an object in three dimensions via cables. We develop robot configurations that ensure static equilibrium of the object at a desired pose... more
Download research papers for free!