|
mlpack
3.4.2
|
Public Types | |
| using | ActionType = typename PolicyType::ActionType |
| Convenient typedef for action. More... | |
Public Member Functions | |
| AggregatedPolicy (std::vector< PolicyType > policies, const arma::colvec &distribution) | |
| void | Anneal () |
| Exploration probability will anneal at each step. More... | |
| ActionType | Sample (const arma::colvec &actionValue, bool deterministic=false) |
| Sample an action based on given action values. More... | |
| PolicyType | The type of the child policy. |
Definition at line 27 of file aggregated_policy.hpp.
| using ActionType = typename PolicyType::ActionType |
Convenient typedef for action.
Definition at line 31 of file aggregated_policy.hpp.
|
inline |
| policies | Child policies. |
| distribution | Probability distribution for each child policy. User should make sure its size is same as the number of policies and the sum of its element is equal to 1. |
Definition at line 39 of file aggregated_policy.hpp.
|
inline |
Exploration probability will anneal at each step.
Definition at line 63 of file aggregated_policy.hpp.
|
inline |
Sample an action based on given action values.
| actionValue | Values for each action. |
| deterministic | Always select the action greedily. |
Definition at line 52 of file aggregated_policy.hpp.
References DiscreteDistribution::Random().
1.8.5