浏览代码

replaced actions with previous_actions in the BrainInfo object

/tag-0.2.0
vincentpierre 7 年前
当前提交
41ab078d
共有 3 个文件被更改,包括 190 次插入214 次删除
  1. 2
      docs/broadcast.md
  2. 400
      images/broadcast.png
  3. 2
      python/unityagents/brain.py

2
docs/broadcast.md


## How to use : Python
When you launch your Unity Environment from python, you can see what the agents connected to non-external brains are doing. When calling `step` or `reset` on your environment, you retrieve a dictionary from brain names to `BrainInfo` objects. Each `BrainInfo` the non-external brains set to broadcast.
Just like with an external brain, the `BrainInfo` object contains the fields for `observations`, `states`, `memories`,`rewards`, `local_done`, `agents` and `actions`. Note that `actions` corresponds to the actions that were taken by the agents at the previous step.
Just like with an external brain, the `BrainInfo` object contains the fields for `observations`, `states`, `memories`,`rewards`, `local_done`, `agents` and `previous_actions`. Note that `previous_actions` corresponds to the actions that were taken by the agents at the previous step, not the current one.
Note that when you do a `step` on the environment, you cannot provide actions for non-external brains. If there are no external brains in the scene, simply call `step()` with no arguments.
You can use the broadcast feature to collect data generated by Player, Heuristics or Internal brains game sessions. You can then use this data to train an agent in a supervised context.

400
images/broadcast.png

之前 之后
宽度: 550  |  高度: 550  |  大小: 64 KiB

2
python/unityagents/brain.py


self.rewards = reward
self.local_done = local_done
self.agents = agents
self.actions = action
self.previous_actions = action
class BrainParameters:

正在加载...
取消
保存