MDP | Define an MDP Problem |
observation_matrix | Extract the Transition, Observation or Reward Matrices from a POMDP |
optimal_action | Optimal action for a belief |
O_ | Define a POMDP Problem |
plot | Visualize a POMDP Policy Graph |
plot_belief_space | Plot a 2D or 3D Projection of the Belief Space |
plot_policy_graph | Visualize a POMDP Policy Graph |
plot_value_function | Plot the Value Function of a POMDP Solution |
policy | Extract the Policy from a POMDP/MDP |
policy_graph | Extract the Policy Graph (as an igraph Object) |
POMDP | Define a POMDP Problem |
read_POMDP | Read and write a POMDP Model to a File in POMDP Format |
reward | Calculate the Reward for a POMDP Solution |
reward_matrix | Extract the Transition, Observation or Reward Matrices from a POMDP |
round_stochastic | Round a stochastic vector or a row-stochastic matrix |
R_ | Define a POMDP Problem |
sample_belief_space | Sample from the Belief Space |
simulate_POMDP | # Simulate belief points |
solve_POMDP | Solve a POMDP Problem |
solve_POMDP_parameter | Solve a POMDP Problem |
Three_doors | Tiger Problem POMDP Specification |
Tiger | Tiger Problem POMDP Specification |
transition_matrix | Extract the Transition, Observation or Reward Matrices from a POMDP |
T_ | Define a POMDP Problem |
update_belief | Belief Update |
write_POMDP | Read and write a POMDP Model to a File in POMDP Format |