Overview
The definition of hybrid action policy network used in PPO, which is mainly composed of three parts: encoder, action_type head (discrete) and action_args head (continuous).
from typing import Dict
import torch
import torch.nn as nn
import treetensor.torch as ttorch
from torch.distributions import Normal, Independent
class HybridPolicyNetwork(nn.Module):
def __init__(self, obs_shape: int, action_shape: Dict[str, int]) -> None:
PyTorch necessary requirements for extending nn.Module . Our network should also subclass this class.
super(HybridPolicyNetwork, self).__init__()
Define encoder module, which maps raw state into embedding vector.
It could be different for various state, such as Convolution Neural Network (CNN) for image state and Multilayer perceptron (MLP) for vector state, respectively.
$$ y = max(W_2 max(W_1x+b_1, 0) + b_2, 0)$$
self.encoder = nn.Sequential(
nn.Linear(obs_shape, 16),
nn.ReLU(),
nn.Linear(16, 32),
nn.ReLU(),
)
Define action_type head module, which outputs discrete logit.
$$ y = Wx + b $$
self.action_type_shape = action_shape['action_type_shape']
self.action_type_head = nn.Linear(32, self.action_type_shape)
Define action_args head module, which outputs corresponding continuous action arguments.
$$ \mu = Wx + b $$
$$\sigma = e^w$$
self.action_args_shape = action_shape['action_args_shape']
self.action_args_mu = nn.Linear(32, self.action_args_shape)
self.action_args_log_sigma = nn.Parameter(torch.zeros(1, self.action_args_shape))
Overview
The computation graph of hybrid action policy network used in PPO.
x -> encoder -> action_type_head -> action_type_logit .
x -> encoder -> action_args_mu -> \mu .
action_args_log_sigma -> exp -> sigma .
def forward(self, x: torch.Tensor) -> Dict[str, torch.Tensor]:
Transform original state into embedding vector, i.e. $$(B, *) -> (B, N)$$
x = self.encoder(x)
Output discrete action logit.
logit = self.action_type_head(x)
Output the argument mu depending on the embedding vector.
mu = self.action_args_mu(x)
Utilize broadcast mechanism to make the same shape between log_sigma and mu.
zeros_like operation doesn't pass gradient.
Related Link
log_sigma = self.action_args_log_sigma + torch.zeros_like(mu)
Utilize exponential operation to produce the actual sigma.
$$\sigma = e^w$$
sigma = torch.exp(log_sigma)
Return treetensor-type output.
return ttorch.as_tensor({
'action_type': logit,
'action_args': {
'mu': mu,
'sigma': sigma
}
})
Overview
The function of sampling hybrid action, input is a treetensor with two keys action_type and action_args .
def sample_hybrid_action(logit: ttorch.Tensor) -> torch.Tensor:
Transform logit (raw output of discrete policy head, e.g. last fully connected layer) into probability.
$$\text{Softmax}(x_{i}) = \frac{\exp(x_i)}{\sum_j \exp(x_j)}$$
prob = torch.softmax(logit.action_type, dim=-1)
Construct categorical distribution. The probability mass function is: $$f(x=i|\boldsymbol{p})=p_i$$
Related Link
discrete_dist = torch.distributions.Categorical(probs=prob)
Sample one discrete action type per sample (state input).
action_type = discrete_dist.sample()
Construct gaussian distribution
$$X \sim \mathcal{N}(\mu,\,\sigma^{2})$$
Its probability density function is: $$f(x) = \frac{1}{\sigma\sqrt{2\pi}} \exp\left( -\frac{1}{2}\left(\frac{x-\mu}{\sigma}\right)^{\!2}\,\right)$$
Related Link
continuous_dist = Normal(logit.action_args.mu, logit.action_args.sigma)
Reinterpret action_shape gaussian distribution into a multivariate gaussian distribution with
diagonal convariance matrix.
Ensure each event is independent with each other.
Related Link
continuous_dist = Independent(continuous_dist, 1)
Sample one action args of the shape action_shape per sample (state input).
action_args = continuous_dist.sample()
Return the final parameterized action.
return ttorch.as_tensor({
'action_type': action_type,
'action_args': action_args,
})
Overview
The function of testing sampling hybrid action. Construct a hybrid action (parameterized action)
policy and sample a group of action.
def test_sample_hybrid_action():
Set batch_size = 4, obs_shape = 10, action_shape is a dict, including 3 possible discrete action types and 3 corresponding continuous arguments. The relationship between action_type and action_args are represented by the below mask .
B, obs_shape, action_shape = 4, 10, {'action_type_shape': 3, 'action_args_shape': 3}
mask = [[0, 1, 0], [1, 0, 0], [0, 0, 1]]
Generate state data from uniform distribution in [0, 1].
state = torch.rand(B, obs_shape)
Define hybrid action network with encoder, discrete head and continuous head.
policy_network = HybridPolicyNetwork(obs_shape, action_shape)
Policy network forward procedure, input state and output treetensor-type logit.
logit = policy_network(state)
assert isinstance(logit, ttorch.Tensor)
assert logit.action_type.shape == (B, action_shape['action_type_shape'])
assert logit.action_args.mu.shape == (B, action_shape['action_args_shape'])
assert logit.action_args.sigma.shape == (B, action_shape['action_args_shape'])
Sample action accoding to corresponding logit part.
action = sample_hybrid_action(logit)
assert action.action_type.shape == (B, )
assert action.action_args.shape == (B, action_shape['action_args_shape'])
Acquire each sample's mask by looking up in mask with action type。
data_mask = torch.as_tensor([mask[i] for i in action.action_type]).bool()
Filter corresponding action_args according to mask and re-assign it.
filtered_action_args = ttorch.masked_select(action.action_args, data_mask)
action.action_args = filtered_action_args
assert action.action_args.shape == (B, )
Select some samples with slicing (for example).
selected_action = action[1:3]
assert selected_action.action_type.shape == (2, )
If you have any questions or advices about this documation, you can raise issues in GitHub (https://github.com/opendilab/PPOxFamily) or email us (opendilab@pjlab.org.cn).