rlHybridStochasticActorPolicy
Policy object to generate hybrid stochastic actions for custom training loops and application deployment
Since R2024b
Description
This object implements a hybrid stochastic policy, which returns stochastic hybrid
actions given an input observation, according to two (one discrete and one continuous)
probability distributions. You can create an rlStochasticActorPolicy
object
from an rlHybridStochasticActor
, or extract it from an rlSACAgent
with a
hybrid action space. You can then train the policy object using a custom training loop or
deploy it for your application using generatePolicyBlock
or generatePolicyFunction
. If UseMaxLikelihoodAction
is set to
1
, the policy is deterministic and therefore it does not explore. For
more information on policies and value functions, see Create Policies and Value Functions.
Creation
Description
creates the hybrid stochastic policy object policy
= rlHybridStochasticActorPolicy(actor
)policy
from the rlHybridStochasticActor
actor actor
. It also sets the
Actor
property of policy
to the input
argument actor
.
Properties
Object Functions
generatePolicyBlock | Generate Simulink block that evaluates policy of an agent or policy object |
generatePolicyFunction | Generate MATLAB function that evaluates policy of an agent or policy object |
getAction | Obtain action from agent, actor, or policy object given environment observations |
getLearnableParameters | Obtain learnable parameter values from agent, function approximator, or policy object |
reset | Reset environment, agent, experience buffer, or policy object |
setLearnableParameters | Set learnable parameter values of agent, function approximator, or policy object |
Examples
Version History
Introduced in R2024b
See Also
Functions
getGreedyPolicy
|getExplorationPolicy
|generatePolicyBlock
|generatePolicyFunction
|getAction
|getLearnableParameters
|setLearnableParameters
Objects
rlMaxQPolicy
|rlEpsilonGreedyPolicy
|rlDeterministicActorPolicy
|rlAdditiveNoisePolicy
|rlStochasticActorPolicy
|rlHybridStochasticActor
|rlSACAgent