Unity 机器学习代理工具包 (ML-Agents) 是一个开源项目,它使游戏和模拟能够作为训练智能代理的环境。
您最多选择25个主题 主题必须以中文或者字母或数字开头,可以包含连字符 (-),并且长度不得超过35个字符
 
 
 
 
 

3.8 KiB

Imitation Learning

It is often more intuitive to simply demonstrate the behavior we want an agent to perform, rather than attempting to have it learn via trial-and-error methods. Consider our running example of training a medic NPC : instead of indirectly training a medic with the help of a reward function, we can give the medic real world examples of observations from the game and actions from a game controller to guide the medic's behavior. More specifically, in this mode, the Brain type during training is set to Player and all the actions performed with the controller (in addition to the agent observations) will be recorded and sent to the Python API. The imitation learning algorithm will then use these pairs of observations and actions from the human player to learn a policy. Video Link.

Using Behavioral Cloning

There are a variety of possible imitation learning algorithms which can be used, the simplest one of them is Behavioral Cloning. It works by collecting training data from a teacher, and then simply uses it to directly learn a policy, in the same way the supervised learning for image classification or other traditional Machine Learning tasks work.

  1. In order to use imitation learning in a scene, the first thing you will need is to create two Brains, one which will be the "Teacher," and the other which will be the "Student." We will assume that the names of the brain GameObjects are "Teacher" and "Student" respectively.
  2. Set the "Teacher" brain to Player mode, and properly configure the inputs to map to the corresponding actions. Ensure that "Broadcast" is checked within the Brain inspector window.
  3. Set the "Student" brain to External mode.
  4. Link the brains to the desired agents (one agent as the teacher and at least one agent as a student).
  5. In trainer_config.yaml, add an entry for the "Student" brain. Set the trainer parameter of this entry to imitation, and the brain_to_imitate parameter to the name of the teacher brain: "Teacher". Additionally, set batches_per_epoch, which controls how much training to do each moment. Increase the max_steps option if you'd like to keep training the agents for a longer period of time.
  6. Launch the training process with python3 python/learn.py --train --slow, and press the ▶️ button in Unity when the message "Ready to connect with the Editor" is displayed on the screen
  7. From the Unity window, control the agent with the Teacher brain by providing "teacher demonstrations" of the behavior you would like to see.
  8. Watch as the agent(s) with the student brain attached begin to behave similarly to the demonstrations.
  9. Once the Student agents are exhibiting the desired behavior, end the training process with CTL+C from the command line.
  10. Move the resulting *.bytes file into the TFModels subdirectory of the Assets folder (or a subdirectory within Assets of your choosing) , and use with Internal brain.

BC Teacher Helper

We provide a convenience utility, BC Teacher Helper component that you can add to the Teacher Agent.

BC Teacher Helper

This utility enables you to use keyboard shortcuts to do the following:

  1. To start and stop recording experiences. This is useful in case you'd like to interact with the game but not have the agents learn from these interactions. The default command to toggle this is to press R on the keyboard.

  2. Reset the training buffer. This enables you to instruct the agents to forget their buffer of recent experiences. This is useful if you'd like to get them to quickly learn a new behavior. The default command to reset the buffer is to press C on the keyboard.