您最多选择25个主题
主题必须以中文或者字母或数字开头,可以包含连字符 (-),并且长度不得超过35个字符
183 行
6.9 KiB
183 行
6.9 KiB
from typing import Dict
|
|
import numpy as np
|
|
import torch
|
|
|
|
from mlagents.trainers.policy.torch_policy import TorchPolicy
|
|
from mlagents.trainers.demo_loader import demo_to_buffer
|
|
from mlagents.trainers.settings import BehavioralCloningSettings, ScheduleType
|
|
from mlagents.trainers.torch.utils import ModelUtils
|
|
|
|
|
|
class BCModule:
|
|
def __init__(
|
|
self,
|
|
policy: TorchPolicy,
|
|
settings: BehavioralCloningSettings,
|
|
policy_learning_rate: float,
|
|
default_batch_size: int,
|
|
default_num_epoch: int,
|
|
):
|
|
"""
|
|
A BC trainer that can be used inline with RL.
|
|
:param policy: The policy of the learning model
|
|
:param settings: The settings for BehavioralCloning including LR strength, batch_size,
|
|
num_epochs, samples_per_update and LR annealing steps.
|
|
:param policy_learning_rate: The initial Learning Rate of the policy. Used to set an appropriate learning rate
|
|
for the pretrainer.
|
|
"""
|
|
self.policy = policy
|
|
self._anneal_steps = settings.steps
|
|
self.current_lr = policy_learning_rate * settings.strength
|
|
|
|
learning_rate_schedule: ScheduleType = ScheduleType.LINEAR if self._anneal_steps > 0 else ScheduleType.CONSTANT
|
|
self.decay_learning_rate = ModelUtils.DecayedValue(
|
|
learning_rate_schedule, self.current_lr, 1e-10, self._anneal_steps
|
|
)
|
|
params = self.policy.actor_critic.parameters()
|
|
self.optimizer = torch.optim.Adam(params, lr=self.current_lr)
|
|
_, self.demonstration_buffer = demo_to_buffer(
|
|
settings.demo_path, policy.sequence_length, policy.behavior_spec
|
|
)
|
|
|
|
self.batch_size = (
|
|
settings.batch_size if settings.batch_size else default_batch_size
|
|
)
|
|
self.num_epoch = settings.num_epoch if settings.num_epoch else default_num_epoch
|
|
self.n_sequences = max(
|
|
min(self.batch_size, self.demonstration_buffer.num_experiences)
|
|
// policy.sequence_length,
|
|
1,
|
|
)
|
|
|
|
self.has_updated = False
|
|
self.use_recurrent = self.policy.use_recurrent
|
|
self.samples_per_update = settings.samples_per_update
|
|
|
|
def update(self) -> Dict[str, np.ndarray]:
|
|
"""
|
|
Updates model using buffer.
|
|
:param max_batches: The maximum number of batches to use per update.
|
|
:return: The loss of the update.
|
|
"""
|
|
# Don't continue training if the learning rate has reached 0, to reduce training time.
|
|
|
|
decay_lr = self.decay_learning_rate.get_value(self.policy.get_current_step())
|
|
if self.current_lr <= 0:
|
|
return {"Losses/Pretraining Loss": 0}
|
|
|
|
batch_losses = []
|
|
possible_demo_batches = (
|
|
self.demonstration_buffer.num_experiences // self.n_sequences
|
|
)
|
|
possible_batches = possible_demo_batches
|
|
|
|
max_batches = self.samples_per_update // self.n_sequences
|
|
|
|
n_epoch = self.num_epoch
|
|
for _ in range(n_epoch):
|
|
self.demonstration_buffer.shuffle(
|
|
sequence_length=self.policy.sequence_length
|
|
)
|
|
if max_batches == 0:
|
|
num_batches = possible_batches
|
|
else:
|
|
num_batches = min(possible_batches, max_batches)
|
|
for i in range(num_batches // self.policy.sequence_length):
|
|
demo_update_buffer = self.demonstration_buffer
|
|
start = i * self.n_sequences * self.policy.sequence_length
|
|
end = (i + 1) * self.n_sequences * self.policy.sequence_length
|
|
mini_batch_demo = demo_update_buffer.make_mini_batch(start, end)
|
|
run_out = self._update_batch(mini_batch_demo, self.n_sequences)
|
|
loss = run_out["loss"]
|
|
batch_losses.append(loss)
|
|
|
|
ModelUtils.update_learning_rate(self.optimizer, decay_lr)
|
|
self.current_lr = decay_lr
|
|
|
|
self.has_updated = True
|
|
update_stats = {"Losses/Pretraining Loss": np.mean(batch_losses)}
|
|
return update_stats
|
|
|
|
def _behavioral_cloning_loss(self, selected_actions, log_probs, expert_actions):
|
|
if self.policy.use_continuous_act:
|
|
bc_loss = torch.nn.functional.mse_loss(selected_actions, expert_actions)
|
|
else:
|
|
log_prob_branches = ModelUtils.break_into_branches(
|
|
log_probs, self.policy.act_size
|
|
)
|
|
bc_loss = torch.mean(
|
|
torch.stack(
|
|
[
|
|
torch.sum(
|
|
-torch.nn.functional.log_softmax(log_prob_branch, dim=1)
|
|
* expert_actions_branch,
|
|
dim=1,
|
|
)
|
|
for log_prob_branch, expert_actions_branch in zip(
|
|
log_prob_branches, expert_actions
|
|
)
|
|
]
|
|
)
|
|
)
|
|
return bc_loss
|
|
|
|
def _update_batch(
|
|
self, mini_batch_demo: Dict[str, np.ndarray], n_sequences: int
|
|
) -> Dict[str, float]:
|
|
"""
|
|
Helper function for update_batch.
|
|
"""
|
|
vec_obs = [ModelUtils.list_to_tensor(mini_batch_demo["vector_obs"])]
|
|
act_masks = None
|
|
if self.policy.use_continuous_act:
|
|
expert_actions = ModelUtils.list_to_tensor(mini_batch_demo["actions"])
|
|
else:
|
|
raw_expert_actions = ModelUtils.list_to_tensor(
|
|
mini_batch_demo["actions"], dtype=torch.long
|
|
)
|
|
expert_actions = ModelUtils.actions_to_onehot(
|
|
raw_expert_actions, self.policy.act_size
|
|
)
|
|
act_masks = ModelUtils.list_to_tensor(
|
|
np.ones(
|
|
(
|
|
self.n_sequences * self.policy.sequence_length,
|
|
sum(self.policy.behavior_spec.discrete_action_branches),
|
|
),
|
|
dtype=np.float32,
|
|
)
|
|
)
|
|
|
|
memories = []
|
|
if self.policy.use_recurrent:
|
|
memories = torch.zeros(1, self.n_sequences, self.policy.m_size)
|
|
|
|
if self.policy.use_vis_obs:
|
|
vis_obs = []
|
|
for idx, _ in enumerate(
|
|
self.policy.actor_critic.network_body.visual_encoders
|
|
):
|
|
vis_ob = ModelUtils.list_to_tensor(
|
|
mini_batch_demo["visual_obs%d" % idx]
|
|
)
|
|
vis_obs.append(vis_ob)
|
|
else:
|
|
vis_obs = []
|
|
|
|
selected_actions, all_log_probs, _, _, _ = self.policy.sample_actions(
|
|
vec_obs,
|
|
vis_obs,
|
|
masks=act_masks,
|
|
memories=memories,
|
|
seq_len=self.policy.sequence_length,
|
|
all_log_probs=True,
|
|
)
|
|
bc_loss = self._behavioral_cloning_loss(
|
|
selected_actions, all_log_probs, expert_actions
|
|
)
|
|
self.optimizer.zero_grad()
|
|
bc_loss.backward()
|
|
|
|
self.optimizer.step()
|
|
run_out = {"loss": bc_loss.detach().cpu().numpy()}
|
|
return run_out
|