Decentralized partially observable Markov decision process
![]() |
This is not a Wikipedia article: It is an individual user's work-in-progress page, and may be incomplete and/or unreliable. For guidance on developing this draft, see Wikipedia:So you made a userspace draft. This draft was last edited 0 seconds ago (purge). Finished? Save your work by pressing the "Save page" button below, and a button will appear here allowing you to submit your draft for review. |
[[Category:Userspace drafts created via the Article Wizard from {{Subst:CURRENTMONTHNAME}} {{Subst:CURRENTYEAR}}]]
{{Subst:Nul|<==do not change this line, it will set the date automatically}}
The decentralized partially observable Markov decision process (Dec-POMDP) [1][2] is a model for coordination and decision-making among multiple agents. It is a probabilistic model that can consider uncertainty in outcomes, sensors and communication (i.e., costly, delayed, noisy or nonexistent communication). it is a generalization of a Markov decision process (MDP) and a partially observable Markov decision process (POMDP) to consider multiple decentralized agents.
References
- ^ Bernstein, Daniel S.; Givan, Robert; Immerman, Neil; Zilberstein, Shlomo (November 2002). "The Complexity of Decentralized Control of Markov Decision Processes". Math. Oper. Res. 27 (4): 819–840. doi:10.1287/moor.27.4.819.297. ISSN 0364-765X.
- ^ Oliehoek, Frans A.; Amato, Christopher. A Concise Introduction to Decentralized POMDPs | SpringerLink. doi:10.1007/978-3-319-28929-8.