Learn R Programming

⚠️There's a newer version (1.2.4) of this package.Take me there.

R package pomdp - Infrastructure for Partially Observable Markov Decision Processes (POMDP)

Introduction

A partially observable Markov decision process (POMDP) models an agent decision process where the agent cannot directly observe the environment’s state, but has to rely on observations. The goal is to find an optimal policy to guide the agent’s actions.

The pomdp package provides the infrastructure to define and analyze the solutions of optimal control problems formulated as Partially Observable Markov Decision Processes (POMDP). The package uses the solvers from pomdp-solve (Cassandra, 2015) available in the companion R package pomdpSolve to solve POMDPs using a variety of exact and approximate algorithms.

The package provides fast functions (using C++, sparse matrix representation, and parallelization with foreach) to perform experiments (sample from the belief space, simulate trajectories, belief update, calculate the regret of a policy). The package also interfaces to the following algorithms:

  • Exact value iteration
    • Enumeration algorithm (Sondik 1971; Monahan 1982).
    • Two pass algorithm (Sondik 1971).
    • Witness algorithm (Littman, Cassandra, and Kaelbling 1995).
    • Incremental pruning algorithm (Zhang and Liu 1996; Cassandra, Littman, and Zhang 1997).
  • Approximate value iteration
    • Finite grid algorithm (Cassandra 2015), a variation of point-based value iteration to solve larger POMDPs (PBVI; see (Pineau, Gordon, and Thrun 2003) without dynamic belief set expansion.
    • SARSOP (Kurniawati, Hsu, and Lee 2008), point-based algorithm that approximates optimally reachable belief spaces for infinite-horizon problems (via package sarsop).

If you are new to POMDPs then start with the POMDP Tutorial.

To cite package ‘pomdp’ in publications use:

Hahsler M (2023). pomdp: Infrastructure for Partially Observable Markov Decision Processes (POMDP). R package version 1.1.2, https://CRAN.R-project.org/package=pomdp.

@Manual{,
  title = {pomdp: Infrastructure for Partially Observable Markov Decision
Processes (POMDP)},
  author = {Michael Hahsler},
  year = {2023},
  note = {R package version 1.1.2},
  url = {https://CRAN.R-project.org/package=pomdp},
}

Installation

Stable CRAN version: Install from within R with

install.packages("pomdp")

Current development version: Install from r-universe.

install.packages("pomdp",
    repos = c("https://mhahsler.r-universe.dev". "https://cloud.r-project.org/"))

Usage

Solving the simple infinite-horizon Tiger problem.

library("pomdp")
data("Tiger")
Tiger
## POMDP, list - Tiger Problem
##   Discount factor: 0.75
##   Horizon: Inf epochs
##   Size: 2 states / 3 actions / 2 obs.
##   Normalized: FALSE
## 
##   Solved: FALSE
## 
##   List components: 'name', 'discount', 'horizon', 'states', 'actions',
##     'observations', 'transition_prob', 'observation_prob', 'reward',
##     'start', 'terminal_values'
sol <- solve_POMDP(model = Tiger)
sol
## POMDP, list - Tiger Problem
##   Discount factor: 0.75
##   Horizon: Inf epochs
##   Size: 2 states / 3 actions / 2 obs.
##   Normalized: FALSE
## 
##   Solved:
##     Method: grid
##     Solution converged: TRUE
##     # of alpha vectors: 5
##     Total expected reward: 1.933439
## 
##   List components: 'name', 'discount', 'horizon', 'states', 'actions',
##     'observations', 'transition_prob', 'observation_prob', 'reward',
##     'start', 'solution'

Display the value function.

plot_value_function(sol, ylim = c(0, 20))

Display the policy graph.

plot_policy_graph(sol)

Acknowledgments

Development of this package was supported in part by National Institute of Standards and Technology (NIST) under grant number 60NANB17D180.

References

Cassandra, Anthony R. 2015. “The POMDP Page.” https://www.pomdp.org.

Cassandra, Anthony R., Michael L. Littman, and Nevin Lianwen Zhang. 1997. “Incremental Pruning: A Simple, Fast, Exact Method for Partially Observable Markov Decision Processes.” In UAI’97: Proceedings of the Thirteenth Conference on Uncertainty in Artificial Intelligence, 54--61.

Kurniawati, Hanna, David Hsu, and Wee Sun Lee. 2008. “SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces.” In In Proc. Robotics: Science and Systems.

Littman, Michael L., Anthony R. Cassandra, and Leslie Pack Kaelbling. 1995. “Learning Policies for Partially Observable Environments: Scaling Up.” In Proceedings of the Twelfth International Conference on International Conference on Machine Learning, 362–70. ICML’95. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.

Monahan, G. E. 1982. “A Survey of Partially Observable Markov Decision Processes: Theory, Models, and Algorithms.” Management Science 28 (1): 1–16.

Pineau, Joelle, Geoff Gordon, and Sebastian Thrun. 2003. “Point-Based Value Iteration: An Anytime Algorithm for POMDPs.” In Proceedings of the 18th International Joint Conference on Artificial Intelligence, 1025–30. IJCAI’03. San Francisco, CA, USA: Morgan Kaufmann Publishers Inc.

Sondik, E. J. 1971. “The Optimal Control of Partially Observable Markov Decision Processes.” PhD thesis, Stanford, California.

Zhang, Nevin L., and Wenju Liu. 1996. “Planning in Stochastic Domains: Problem Characteristics and Approximation.” HKUST-CS96-31. Hong Kong University.

Copy Link

Version

Install

install.packages('pomdp')

Monthly Downloads

756

Version

1.1.3

License

GPL (>= 3)

Issues

Pull Requests

Stars

Forks

Maintainer

Michael Hahsler

Last Published

December 21st, 2023

Functions in pomdp (1.1.3)

reward

Calculate the Reward for a POMDP Solution
projection

Defining a Belief Space Projection
sample_belief_space

Sample from the Belief Space
regret

Calculate the Regret of a Policy
round_stochastic

Round a stochastic vector or a row-stochastic matrix
policy

Extract the Policy from a POMDP/MDP
pomdp-package

pomdp: Infrastructure for Partially Observable Markov Decision Processes (POMDP)
policy_graph

POMDP Policy Graphs
simulate_MDP

Simulate Trajectories in a MDP
transition_graph

Transition Graph
update_belief

Belief Update
solve_MDP

Solve an MDP Problem
plot_policy_graph

POMDP Plot Policy Graphs
solve_SARSOP

Solve a POMDP Problem using SARSOP
simulate_POMDP

Simulate Trajectories in a POMDP
solve_POMDP

Solve a POMDP Problem using pomdp-solver
write_POMDP

Read and write a POMDP Model to a File in POMDP Format
value_function

Value Function
add_policy

Add a Policy to a POMDP Problem Description
MDP

Define an MDP Problem
Tiger

Tiger Problem POMDP Specification
POMDP

Define a POMDP Problem
colors

Default Colors for Visualization in Package pomdp
optimal_action

Optimal action for a belief
POMDP_accessors

Access to Parts of the POMDP Description
estimate_belief_for_nodes

Estimate the Belief for Policy Graph Nodes
plot_belief_space

Plot a 2D or 3D Projection of the Belief Space
Maze

Steward Russell's 4x3 Maze MDP