Unity 机器学习代理工具包 (ML-Agents) 是一个开源项目,它使游戏和模拟能够作为训练智能代理的环境。
您最多选择25个主题 主题必须以中文或者字母或数字开头,可以包含连字符 (-),并且长度不得超过35个字符
 
 
 
 
 

4.4 KiB

Training with Imitation Learning

It is often more intuitive to simply demonstrate the behavior we want an agent to perform, rather than attempting to have it learn via trial-and-error methods. Consider our running example of training a medic NPC. Instead of indirectly training a medic with the help of a reward function, we can give the medic real world examples of observations from the game and actions from a game controller to guide the medic's behavior. Imitation Learning uses pairs of observations and actions from a demonstration to learn a policy. Video Link.

Imitation learning can also be used to help reinforcement learning. Especially in environments with sparse (i.e., infrequent or rare) rewards, the agent may never see the reward and thus not learn from it. Curiosity (which is available in the toolkit) helps the agent explore, but in some cases it is easier to show the agent how to achieve the reward. In these cases, imitation learning combined with reinforcement learning can dramatically reduce the time the agent takes to solve the environment. For instance, on the Pyramids environment, using 6 episodes of demonstrations can reduce training steps by more than 4 times. See Behavioral Cloning + GAIL + Curiosity + RL below.

Using Demonstrations with Reinforcement Learning

The ML-Agents toolkit provides two features that enable your agent to learn from demonstrations. In most scenarios, you should combine these two features

  • GAIL (Generative Adversarial Imitation Learning) uses an adversarial approach to reward your Agent for behaving similar to a set of demonstrations. To use GAIL, you can add the GAIL reward signal. GAIL can be used with or without environment rewards, and works well when there are a limited number of demonstrations.
  • Behavioral Cloning (BC) trains the Agent's neural network to exactly mimic the actions shown in a set of demonstrations. The BC feature can be enabled on the PPO or SAC trainer. BC tends to work best when there are a lot of demonstrations, or in conjunction with GAIL and/or an extrinsic reward.

How to Choose

If you want to help your agents learn (especially with environments that have sparse rewards) using pre-recorded demonstrations, you can generally enable both GAIL and Behavioral Cloning at low strengths in addition to having an extrinsic reward. An example of this is provided for the Pyramids example environment under PyramidsLearning in config/gail_config.yaml.

If you want to train purely from demonstrations, GAIL and BC without an extrinsic reward signal is the preferred approach. An example of this is provided for the Crawler example environment under CrawlerStaticLearning in config/gail_config.yaml.

Recording Demonstrations

It is possible to record demonstrations of agent behavior from the Unity Editor, and save them as assets. These demonstrations contain information on the observations, actions, and rewards for a given agent during the recording session. They can be managed from the Editor, as well as used for training with BC and GAIL.

In order to record demonstrations from an agent, add the Demonstration Recorder component to a GameObject in the scene which contains an Agent component. Once added, it is possible to name the demonstration that will be recorded from the agent.

BC Teacher Helper

When Record is checked, a demonstration will be created whenever the scene is played from the Editor. Depending on the complexity of the task, anywhere from a few minutes or a few hours of demonstration data may be necessary to be useful for imitation learning. When you have recorded enough data, end the Editor play session, and a .demo file will be created in the Assets/Demonstrations folder (by default). This file contains the demonstrations. Clicking on the file will provide metadata about the demonstration in the inspector.

BC Teacher Helper