OrnsteinUhlenbeckProcessWrapper¶
- class torchrl.modules.OrnsteinUhlenbeckProcessWrapper(*args, **kwargs)[source]¶
Ornstein-Uhlenbeck exploration policy wrapper.
Presented in “CONTINUOUS CONTROL WITH DEEP REINFORCEMENT LEARNING”, https://arxiv.org/pdf/1509.02971.pdf.
The OU exploration is to be used with continuous control policies and introduces a auto-correlated exploration noise. This enables a sort of ‘structured’ exploration.
Noise equation:
\[noise_t = noise_{t-1} + \theta * (mu - noise_{t-1}) * dt + \sigma_t * \sqrt{dt} * W\]Sigma equation:
\[\sigma_t = max(\sigma^{min, (-(\sigma_{t-1} - \sigma^{min}) / (n^{\text{steps annealing}}) * n^{\text{steps}} + \sigma))\]To keep track of the steps and noise from sample to sample, an
"ou_prev_noise{id}"
and"ou_steps{id}"
keys will be written in the input/output tensordict. It is expected that the tensordict will be zeroed at reset, indicating that a new trajectory is being collected. If not, and is the same tensordict is used for consecutive trajectories, the step count will keep on increasing across rollouts. Note that the collector classes take care of zeroing the tensordict at reset time.Note
Once an environment has been wrapped in
OrnsteinUhlenbeckProcessWrapper
, it is crucial to incorporate a call tostep()
in the training loop to update the exploration factor. Since it is not easy to capture this omission no warning or exception will be raised if this is ommitted!- Parameters:
policy (TensorDictModule) – a policy
- Keyword Arguments:
eps_init (scalar) – initial epsilon value, determining the amount of noise to be added. default: 1.0
eps_end (scalar) – final epsilon value, determining the amount of noise to be added. default: 0.1
annealing_num_steps (int) – number of steps it will take for epsilon to reach the eps_end value. default: 1000
theta (scalar) – theta factor in the noise equation default: 0.15
mu (scalar) – OU average (mu in the noise equation). default: 0.0
sigma (scalar) – sigma value in the sigma equation. default: 0.2
dt (scalar) – dt in the noise equation. default: 0.01
x0 (Tensor, ndarray, optional) – initial value of the process. default: 0.0
sigma_min (number, optional) – sigma_min in the sigma equation. default: None
n_steps_annealing (int) – number of steps for the sigma annealing. default: 1000
action_key (NestedKey, optional) – key of the action to be modified. default: “action”
is_init_key (NestedKey, optional) – key where to find the is_init flag used to reset the noise steps. default: “is_init”
spec (TensorSpec, optional) – if provided, the sampled action will be projected onto the valid action space once explored. If not provided, the exploration wrapper will attempt to recover it from the policy.
safe (bool) – if
True
, actions that are out of bounds given the action specs will be projected in the space given theTensorSpec.project
heuristic. default: True
Examples
>>> import torch >>> from tensordict import TensorDict >>> from torchrl.data import Bounded >>> from torchrl.modules import OrnsteinUhlenbeckProcessWrapper, Actor >>> torch.manual_seed(0) >>> spec = Bounded(-1, 1, torch.Size([4])) >>> module = torch.nn.Linear(4, 4, bias=False) >>> policy = Actor(module=module, spec=spec) >>> explorative_policy = OrnsteinUhlenbeckProcessWrapper(policy) >>> td = TensorDict({"observation": torch.zeros(10, 4)}, batch_size=[10]) >>> print(explorative_policy(td)) TensorDict( fields={ _ou_prev_noise: Tensor(torch.Size([10, 4]), dtype=torch.float32), _ou_steps: Tensor(torch.Size([10, 1]), dtype=torch.int64), action: Tensor(torch.Size([10, 4]), dtype=torch.float32), observation: Tensor(torch.Size([10, 4]), dtype=torch.float32)}, batch_size=torch.Size([10]), device=None, is_shared=False)
- forward(tensordict: TensorDictBase) TensorDictBase [source]¶
Define the computation performed at every call.
Should be overridden by all subclasses.
Note
Although the recipe for forward pass needs to be defined within this function, one should call the
Module
instance afterwards instead of this since the former takes care of running the registered hooks while the latter silently ignores them.