Unity 机器学习代理工具包 (ML-Agents) 是一个开源项目,它使游戏和模拟能够作为训练智能代理的环境。
您最多选择25个主题 主题必须以中文或者字母或数字开头,可以包含连字符 (-),并且长度不得超过35个字符
 
 
 
 
 

185 行
6.7 KiB

import threading
from mlagents.torch_utils import torch
from mlagents_envs.logging_util import get_logger
from mlagents.trainers.settings import SerializationSettings
logger = get_logger(__name__)
class exporting_to_onnx:
"""
Set this context by calling
```
with exporting_to_onnx():
```
Within this context, the variable exporting_to_onnx.is_exporting() will be true.
This implementation is thread safe.
"""
# local is_exporting flag for each thread
_local_data = threading.local()
_local_data._is_exporting = False
# global lock shared among all threads, to make sure only one thread is exporting at a time
_lock = threading.Lock()
def __enter__(self):
self._lock.acquire()
self._local_data._is_exporting = True
def __exit__(self, *args):
self._local_data._is_exporting = False
self._lock.release()
@staticmethod
def is_exporting():
if not hasattr(exporting_to_onnx._local_data, "_is_exporting"):
return False
return exporting_to_onnx._local_data._is_exporting
class TensorNames:
BatchSizePlaceholder = "batch_size"
SequenceLengthPlaceholder = "sequence_length"
VectorObservationPlaceholder = "vector_observation"
RecurrentInPlaceholder = "recurrent_in"
recurrentInPlaceholderH = "recurrent_in_h"
recurrentInPlaceholderC = "recurrent_in_c"
VisualObservationPlaceholderPrefix = "visual_observation_"
ObservationPlaceholderPrefix = "obs_"
PreviousActionPlaceholder = "prev_action"
ActionMaskPlaceholder = "action_masks"
RandomNormalEpsilonPlaceholder = "epsilon"
ValueEstimateOutput = "value_estimate"
RecurrentOutput = "recurrent_out"
recurrentOutputH = "recurrent_out_h"
recurrentOutputC = "recurrent_out_c"
MemorySize = "memory_size"
VersionNumber = "version_number"
ContinuousActionOutputShape = "continuous_action_output_shape"
DiscreteActionOutputShape = "discrete_action_output_shape"
ContinuousActionOutput = "continuous_actions"
DiscreteActionOutput = "discrete_actions"
# Deprecated TensorNames entries for backward compatibility
IsContinuousControlDeprecated = "is_continuous_control"
ActionOutputDeprecated = "action"
ActionOutputShapeDeprecated = "action_output_shape"
class ModelSerializer:
def __init__(self, policy):
# ONNX only support input in NCHW (channel first) format.
# Barracuda also expect to get data in NCHW.
# Any multi-dimentional input should follow that otherwise will
# cause problem to barracuda import.
self.policy = policy
observation_specs = self.policy.behavior_spec.observation_specs
batch_dim = [1]
seq_len_dim = [1]
vec_obs_size = 0
for obs_spec in observation_specs:
if len(obs_spec.shape) == 1:
vec_obs_size += obs_spec.shape[0]
num_vis_obs = sum(
1 for obs_spec in observation_specs if len(obs_spec.shape) == 3
)
dummy_vec_obs = [torch.zeros(batch_dim + [vec_obs_size])]
# create input shape of NCHW
# (It's NHWC in observation_specs.shape)
dummy_vis_obs = [
torch.zeros(
batch_dim + [obs_spec.shape[2], obs_spec.shape[0], obs_spec.shape[1]]
)
for obs_spec in observation_specs
if len(obs_spec.shape) == 3
]
dummy_var_len_obs = [
torch.zeros(batch_dim + [obs_spec.shape[0], obs_spec.shape[1]])
for obs_spec in observation_specs
if len(obs_spec.shape) == 2
]
dummy_masks = torch.ones(
batch_dim + [sum(self.policy.behavior_spec.action_spec.discrete_branches)]
)
dummy_memories = torch.zeros(
batch_dim + seq_len_dim + [self.policy.export_memory_size]
)
self.dummy_input = (
dummy_vec_obs,
dummy_vis_obs,
dummy_var_len_obs,
dummy_masks,
dummy_memories,
)
self.input_names = [TensorNames.VectorObservationPlaceholder]
for i in range(num_vis_obs):
self.input_names.append(
TensorNames.VisualObservationPlaceholderPrefix + str(i)
)
for i, obs_spec in enumerate(observation_specs):
if len(obs_spec.shape) == 2:
self.input_names.append(
TensorNames.ObservationPlaceholderPrefix + str(i)
)
self.input_names += [
TensorNames.ActionMaskPlaceholder,
TensorNames.RecurrentInPlaceholder,
]
self.dynamic_axes = {name: {0: "batch"} for name in self.input_names}
self.output_names = [TensorNames.VersionNumber, TensorNames.MemorySize]
if self.policy.behavior_spec.action_spec.continuous_size > 0:
self.output_names += [
TensorNames.ContinuousActionOutput,
TensorNames.ContinuousActionOutputShape,
]
self.dynamic_axes.update({TensorNames.ContinuousActionOutput: {0: "batch"}})
if self.policy.behavior_spec.action_spec.discrete_size > 0:
self.output_names += [
TensorNames.DiscreteActionOutput,
TensorNames.DiscreteActionOutputShape,
]
self.dynamic_axes.update({TensorNames.DiscreteActionOutput: {0: "batch"}})
if (
self.policy.behavior_spec.action_spec.continuous_size == 0
or self.policy.behavior_spec.action_spec.discrete_size == 0
):
self.output_names += [
TensorNames.ActionOutputDeprecated,
TensorNames.IsContinuousControlDeprecated,
TensorNames.ActionOutputShapeDeprecated,
]
self.dynamic_axes.update({TensorNames.ActionOutputDeprecated: {0: "batch"}})
if self.policy.export_memory_size > 0:
self.output_names += [TensorNames.RecurrentOutput]
def export_policy_model(self, output_filepath: str) -> None:
"""
Exports a Torch model for a Policy to .onnx format for Unity embedding.
:param output_filepath: file path to output the model (without file suffix)
"""
onnx_output_path = f"{output_filepath}.onnx"
logger.info(f"Converting to {onnx_output_path}")
with exporting_to_onnx():
torch.onnx.export(
self.policy.actor,
self.dummy_input,
onnx_output_path,
opset_version=SerializationSettings.onnx_opset,
input_names=self.input_names,
output_names=self.output_names,
dynamic_axes=self.dynamic_axes,
)
logger.info(f"Exported {onnx_output_path}")