浏览代码

try reload all

/develop/bisim-review
Andrew Cohen 4 年前
当前提交
12eda929
共有 4 个文件被更改,包括 8 次插入9 次删除
  1. 1
      config/ppo_transfer/3DBall.yaml
  2. 1
      config/ppo_transfer/3DBallHard.yaml
  3. 5
      config/ppo_transfer/3DBallHardTransfer.yaml
  4. 10
      ml-agents/mlagents/trainers/policy/transfer_policy.py

1
config/ppo_transfer/3DBall.yaml


predict_return: true
use_bisim: false
separate_value_train: true
separate_value_net: true
in_batch_alter: true
network_settings:
normalize: true

1
config/ppo_transfer/3DBallHard.yaml


predict_return: true
use_bisim: false
separate_value_train: true
separate_value_net: true
in_batch_alter: true
network_settings:
normalize: true

5
config/ppo_transfer/3DBallHardTransfer.yaml


predict_return: true
use_bisim: false
separate_value_train: true
separate_value_net: true
load_policy: false
load_policy: true
load_value: false
load_value: true
transfer_path: "results/3d-tr-s1/3DBall"
use_transfer: true
in_batch_alter: true

10
ml-agents/mlagents/trainers/policy/transfer_policy.py


:param encoded_state: Tensor corresponding to encoded current state.
:param encoded_next_state: Tensor corresponding to encoded next state.
"""
if not self.transfer:
encoded_state = tf.stop_gradient(encoded_state)
#if not self.transfer:
# hidden = tf.stop_gradient(hidden)
for i in range(forward_layers):
hidden = tf.layers.dense(
hidden,

forward_layers: int,
separate_train: bool = False,
):
if not self.transfer:
encoded_state = tf.stop_gradient(encoded_state)
#if not self.transfer:
# hidden = tf.stop_gradient(hidden)

正在加载...
取消
保存