浏览代码

fixed beta

/asymm-envs
Andrew Cohen 4 年前
当前提交
50e4585f
共有 1 个文件被更改,包括 12 次插入6 次删除
  1. 18
      ml-agents/mlagents/trainers/ppo/optimizer.py

18
ml-agents/mlagents/trainers/ppo/optimizer.py


"Losses/Value Loss": "value_loss",
"Losses/Policy Loss": "policy_loss",
"Policy/Learning Rate": "learning_rate",
"Policy/Beta": "beta",
"Policy/Epsilon": "epsilon",
}
if self.policy.use_recurrent:
self.m_size = self.policy.m_size

"policy_loss": self.abs_policy_loss,
"update_batch": self.update_batch,
"learning_rate": self.learning_rate,
"beta": self.decay_beta,
"epsilon": self.decay_epsilon,
}
)

# decay_beta = tf.train.polynomial_decay(
# beta, self.policy.global_step, max_step, 1e-5, power=1.0
# )
decay_epsilon = tf.Variable(epsilon)
decay_beta = tf.Variable(beta)
self.decay_epsilon = tf.constant(epsilon)
self.decay_beta = tf.constant(beta)
-decay_epsilon,
decay_epsilon,
-self.decay_epsilon,
self.decay_epsilon,
)
v_opt_a = tf.squared_difference(
self.returns_holders[name], tf.reduce_sum(head, axis=1)

r_theta = tf.exp(probs - old_probs)
p_opt_a = r_theta * advantage
p_opt_b = (
tf.clip_by_value(r_theta, 1.0 - decay_epsilon, 1.0 + decay_epsilon)
tf.clip_by_value(
r_theta, 1.0 - self.decay_epsilon, 1.0 + self.decay_epsilon
)
* advantage
)
self.policy_loss = -tf.reduce_mean(

self.loss = (
self.policy_loss
+ 0.5 * self.value_loss
- decay_beta
- self.decay_beta
* tf.reduce_mean(tf.dynamic_partition(entropy, self.policy.mask, 2)[1])
)

正在加载...
取消
保存