#partially_observable_markov_decision_process
Partially observable Markov decision process
Generalization of a Markov decision process
A partially observable Markov decision process (POMDP) is a generalization of a Markov decision process (MDP). A POMDP models an agent decision process in which it is assumed that the system dynamics are determined by an MDP, but the agent cannot directly observe the underlying state. Instead, it must maintain a sensor model and the underlying MDP. Unlike the policy function in MDP which maps the underlying states to the actions, POMDP's policy is a mapping from the history of observations to the actions.
Tue 23rd
Provided by Wikipedia
This keyword could refer to multiple things. Here are some suggestions:
0 searches
This keyword has never been searched before
This keyword has never been searched for with any other keyword.