Abstract

In this paper, we describe the partially observable Markov decision process (pomdp) approach to finding optimal or near-optimal control strategies for partially observable stochastic environments, given a complete model of the environment. The pomdp approach was originally developed in the operations research community and provides a formal basis for planning problems that have been of interest to the AI community. We found the existing algorithms for computing optimal control strategies to be highly computationally inefficient and have developed a new algorithm that is empirically more efficient. We sketch this algorithm and present preliminary results on several small problems that illustrate important properties of the pomdp approach. Introduction Agents that act in real environments, whether physical or virtual, rarely have complete information about the state of the environment in which they are working. It is necessary for them to choose their actions in partial ignorance and o...

Keywords

Partially observable Markov decision processObservableSketchComputer scienceMarkov decision processMathematical optimizationOptimal controlMarkov processProcess (computing)Control (management)Markov chainArtificial intelligenceAlgorithmMarkov modelMachine learningMathematics

Affiliated Institutions

Related Publications

Digital Natives, Digital Immigrants Part 1

The main purpose of this project is to develop mathematical competences through interactive educational environments of multi-device access in the ninth grade students of the In...

2001 On the Horizon The International Jour... 10294 citations

Publication Info

Year
1994
Type
article
Pages
1023-1028
Citations
561
Access
Closed

External Links

Citation Metrics

561
OpenAlex

Cite This

Anthony R. Cassandra, Leslie Pack Kaelbling, Michael L. Littman (1994). Acting Optimally in Partially Observable Stochastic Domains. , 1023-1028.