Jump to content

Decentralized partially observable Markov decision process

From Wikipedia, the free encyclopedia
This is an old revision of this page, as edited by Betelgeuse99 (talk | contribs) at 00:19, 3 November 2017 (Page started). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.
(diff) ← Previous revision | Latest revision (diff) | Newer revision → (diff)

[[Category:Userspace drafts created via the Article Wizard from {{Subst:CURRENTMONTHNAME}} {{Subst:CURRENTYEAR}}]]

{{Subst:Nul|<==do not change this line, it will set the date automatically}}

The decentralized partially observable Markov decision process (Dec-POMDP) [1][2] is a model for coordination and decision-making among multiple agents. It is a probabilistic model that can consider uncertainty in outcomes, sensors and communication (i.e., costly, delayed, noisy or nonexistent communication). it is a generalization of a Markov decision process (MDP) and a partially observable Markov decision process (POMDP) to consider multiple decentralized agents.

References

  1. ^ Bernstein, Daniel S.; Givan, Robert; Immerman, Neil; Zilberstein, Shlomo (November 2002). "The Complexity of Decentralized Control of Markov Decision Processes". Math. Oper. Res. 27 (4): 819–840. doi:10.1287/moor.27.4.819.297. ISSN 0364-765X.
  2. ^ Oliehoek, Frans A.; Amato, Christopher. A Concise Introduction to Decentralized POMDPs | SpringerLink. doi:10.1007/978-3-319-28929-8.