Unity 机器学习代理工具包 (ML-Agents) 是一个开源项目,它使游戏和模拟能够作为训练智能代理的环境。
您最多选择25个主题 主题必须以中文或者字母或数字开头,可以包含连字符 (-),并且长度不得超过35个字符
 
 
 
 
 
Marwan Mattar 19a8c654 Added a note to the Translation section 7 年前
docs included baidu pan for the TensorflowSharp plugin 7 年前
python Merge pull request #547 from Unity-Technologies/develop-feature-docker-improvements 7 年前
unity-environment Merge pull request #536 from Unity-Technologies/master 7 年前
unity-volume [containerization] CPU based containerization to support all environments that don't use observations 7 年前
.gitattributes added the localized zh docs 7 年前
.gitignore Dev docs (#361) 7 年前
CODE_OF_CONDUCT.md Adds code of conduct to the repo 7 年前
CONTRIBUTING.md Fixed lingering broken links 7 年前
Dockerfile [containers] Enables container support for scenes that use visual observations 7 年前
LICENSE Initial commit 7 年前
README.md Added a note to the Translation section 7 年前

README.md

Unity ML-Agents (Beta)

Unity Machine Learning Agents (ML-Agents) is an open-source Unity plugin that enables games and simulations to serve as environments for training intelligent agents. Agents can be trained using reinforcement learning, imitation learning, neuroevolution, or other machine learning methods through a simple-to-use Python API. We also provide implementations (based on TensorFlow) of state-of-the-art algorithms to enable game developers and hobbyists to easily train intelligent agents for 2D, 3D and VR/AR games. These trained agents can be used for multiple purposes, including controlling NPC behavior (in a variety of settings such as multi-agent and adversarial), automated testing of game builds and evaluating different game design decisions pre-release. ML-Agents is mutually beneficial for both game developers and AI researchers as it provides a central platform where advances in AI can be evaluated on Unity’s rich environments and then made accessible to the wider research and game developer communities.

Features

  • Unity environment control from Python
  • 10+ sample Unity environments
  • Support for multiple environment configurations and training scenarios
  • Train memory-enhanced Agents using deep reinforcement learning
  • Easily definable Curriculum Learning scenarios
  • Broadcasting of Agent behavior for supervised learning
  • Built-in support for Imitation Learning
  • Flexible Agent control with On Demand Decision Making
  • Visualizing network outputs within the environment
  • Simplified set-up with Docker

Documentation and References

For more information, in addition to installation and usage instructions, see our documentation home. If you have used a version of ML-Agents prior to v0.3, we strongly recommend our guide on migrating to v0.3.

We have also published a series of blog posts that are relevant for ML-Agents:

In addition to our own documentation, here are some additional, relevant articles:

Community and Feedback

ML-Agents is an open-source project and we encourage and welcome contributions. If you wish to contribute, be sure to review our contribution guidelines and code of conduct.

You can connect with us and the broader community through Unity Connect and GitHub:

  • Join our Unity Machine Learning Channel to connect with others using ML-Agents and Unity developers enthusiastic about machine learning. We use that channel to surface updates regarding ML-Agents (and, more broadly, machine learning in games).
  • If you run into any problems using ML-Agents, submit an issue and make sure to include as much detail as possible.

For any other questions or feedback, connect directly with the ML-Agents team at ml-agents@unity3d.com.

Translations

To make Unity ML-Agents accessible to the global research and Unity developer communities, we're attempting to create and maintain translations of our documentation. We've started with translating a subset of the documentation to one language (Chinese), but we hope to continue translating more pages and to other languages. Consequently, we welcome any enhancements and improvements from the community.

License

Apache License 2.0