GitHub
582859b6
New Crawler Variable Speed Scenes ( #4382 )
* init
* updating prefabs
* spawn a target
* add brains
* update static prefabs
* enable enhanced determinism
* reset manifest
* add nn files. update to 15M steps
* update prefabs
* increase max speed to 15
* add new local model for 15 speed
* update prefabs
* add configs
* update configs/prefabs
* cleanup
* added final nn models
* add new demos and do more cleanup.
* add meta files
* add RigidbodySensor
* update prefab. about to retrain
* remove body pen
* add fixed crawler & retrained nn file, new demos
* train 10M steps
* Update Crawler Docs
* more prefab cleanup
* add meta files
* Update Project/Assets/ML-Agents/Examples/Crawler/Scripts/CrawlerAgent.cs
Co-authored-by: Vincent-Pierre BERGES <vincentpierre@unity3d.com>
* remove unused prefab
* update comment
* add summary tags
* cleanup and add more comments
* remove unused prefab
* Update P...
4 年前
GitHub
06f788a4
Fix a few out-of-date things in CONTRIBUTING.md ( #4428 )
* Fix a few out-of-date things in CONTRIBUTING.md
* Update CONTRIBUTING.md
4 年前
GitHub
7b4d0865
[Bug fix] Fix bug in GAIL gradient penalty ( #4425 )
4 年前
GitHub
a117c932
Grid Sensor ( #4399 )
Co-authored-by: Chris Elion <chris.elion@unity3d.com>
4 年前
GitHub
6db9520c
enable enhanceddeterminism in physics settings ( #4423 )
4 年前
GitHub
bfda9576
Replace brain_name with behavior_name ( #4419 )
brain_name -> behavior_name
some prob -> log_prob in comments
rename files optimizer -> optimizer_tf for tensorflow
4 年前
Ervin Teng
1dca75d8
Move linear encoding to NetworkBody
4 年前
Ruo-Ping Dong
27fb4270
brain_name to behavior_name
4 年前
GitHub
0e590a49
[Feature] use extra_requirement to add torch as optional dependency to mlagents ( #4417 )
* Feature use extra_requirement to add torch as optional dependency to mlagents
- todo: documentation
* Update ml-agents/setup.py
Co-authored-by: Chris Elion <chris.elion@unity3d.com>
Co-authored-by: Chris Elion <chris.elion@unity3d.com>
4 年前
GitHub
1076d275
Remove unused methods in trainer_controller.py ( #4418 )
4 年前
GitHub
3379734a
Run pytest on GitHub Actions ( #4416 )
4 年前
GitHub
498934f9
Replace torch.detach().cpu().numpy() with a utils method ( #4406 )
* Replace torch.detach().cpu().numpy() with a utils method
* Using item() in place of to_numpy()
* more use of item() and additional tests
4 年前
GitHub
4775e23c
ignore code coverage and api scrape output ( #4412 )
4 年前
GitHub
ffcb00d2
Update Background-Machine-Learning.md ( #4415 )
Added missing word in text.
4 年前
GitHub
9bf1c03f
precommit github action ( #4410 )
4 年前
GitHub
12e15e29
Fix on GAIL Torch when using actions ( #4407 )
4 年前
GitHub
d08bad06
Increase min supported tensorflow to 1.14.0 ( #4411 )
4 年前
GitHub
48f217b9
Rename Saver to ModelSaver ( #4402 )
Rename Saver to ModelSaver to avoid confusion with tf.Saver
4 年前
Ruo-Ping Dong
56feb8af
update test_saver_reward_providers.py
4 年前
Ruo-Ping Dong
88eff042
Merge branch 'master' into develop-saver-name
4 年前
Ruo-Ping Dong
e60c7038
Merge branch 'master' into develop-saver-name
4 年前
GitHub
328353bc
Torch : Saving/Loading of the reward providers ( #4405 )
* Saving the reward providers
* adding tests
* Moved the tests around
* Update ml-agents/mlagents/trainers/tests/torch/saver/test_saver_reward_providers.py
* Update ml-agents/mlagents/trainers/tests/torch/saver/test_saver_reward_providers.py
* Update ml-agents/mlagents/trainers/tests/torch/saver/test_saver_reward_providers.py
Co-authored-by: Ruo-Ping (Rachel) Dong <ruoping.dong@unity3d.com>
* Update ml-agents/mlagents/trainers/tests/torch/saver/test_saver_reward_providers.py
Co-authored-by: Ruo-Ping (Rachel) Dong <ruoping.dong@unity3d.com>
Co-authored-by: Ruo-Ping (Rachel) Dong <ruoping.dong@unity3d.com>
4 年前
Ruo-Ping Dong
07e82899
update torch saver test
4 年前
GitHub
38e9387b
Fix NNCheckpointManager for Torch
Fix NNCheckpointManager for Torch
4 年前
Ruo-Ping Dong
a74c904a
Merge branch 'master' into develop-saver-name
4 年前
GitHub
df685184
Make --torch use torch even without config ( #4400 )
* Make --torch use torch even without config
* Update ml-agents/mlagents/trainers/trainer_util.py
Co-authored-by: andrewcoh <54679309+andrewcoh@users.noreply.github.com>
* Update ml-agents/mlagents/trainers/trainer_util.py
Co-authored-by: andrewcoh <54679309+andrewcoh@users.noreply.github.com>
* Update ml-agents/mlagents/trainers/trainer_util.py
Co-authored-by: andrewcoh <54679309+andrewcoh@users.noreply.github.com>
* renaming use_torch to force_torch
Co-authored-by: andrewcoh <54679309+andrewcoh@users.noreply.github.com>
4 年前
GitHub
70197342
Add torch saver test
Add torch saver test
4 年前
Ruo-Ping Dong
f2a8c421
add torch saver test
4 年前
Ruo-Ping Dong
09c22679
fix NNCheckpointManager for Torch
4 年前
Ruo-Ping Dong
c47ffc20
Rename saver
4 年前
GitHub
a79aa854
[ci] Shorten max steps for strikergoalie ( #4394 )
4 年前
GitHub
1955af9e
[feature] Add experimental PyTorch support ( #4335 )
* Begin porting work
* Add ResNet and distributions
* Dynamically construct actor and critic
* Initial optimizer port
* Refactoring policy and optimizer
* Resolving a few bugs
* Share more code between tf and torch policies
* Slightly closer to running model
* Training runs, but doesn’t actually work
* Fix a couple additional bugs
* Add conditional sigma for distribution
* Fix normalization
* Support discrete actions as well
* Continuous and discrete now train
* Mulkti-discrete now working
* Visual observations now train as well
* GRU in-progress and dynamic cnns
* Fix for memories
* Remove unused arg
* Combine actor and critic classes. Initial export.
* Support tf and pytorch alongside one another
* Prepare model for onnx export
* Use LSTM and fix a few merge errors
* Fix bug in probs calculation
* Optimize np -> tensor operations
* Time action sample funct...
4 年前
GitHub
ea3944fc
DemonstrationRecorder: Add numStepsToRecord feature. ( #4381 )
* add numSteps property and update recorder script
* add changes based on PR suggestions
* make zero default. update logic
* updated comments & tooltip
4 年前
GitHub
a796b0f1
Merge pull request #4387 from Unity-Technologies/master-pin-xdist
Pin xdist version.
4 年前
Christopher Goy
c448eb65
Pin xdist version on verfided branch.
4 年前
GitHub
b337e62b
add build-in module dependencies ( #4384 )
* add build-in module dependencies
* changelog
4 年前
GitHub
abfadb3d
Reduce max steps for striker vs. goalie ( #4377 )
4 年前
GitHub
31919e08
[MLA-1267] Account for actuators in training and inference. ( #4371 )
4 年前
GitHub
5457ff38
Update README.md ( #4376 )
Adjusted readme to reflect new release
4 年前
GitHub
9cc4e0b2
update demonstration meta files ( #4367 )
4 年前
GitHub
99ccb7b9
Merge pull request #4369 from Unity-Technologies/release_6-to-master
Merge release 6 back into master
4 年前
GitHub
e7916b08
add pre-commit hook for dotnet-format ( #4362 )
4 年前
GitHub
cf570f8c
Update barracuda in the hopes that our burst crashes go away. ( #4359 ) ( #4365 )
4 年前
GitHub
8bbcfc28
fix deleted observations ( #4366 )
4 年前
GitHub
bb9417f7
Update example environments to use the Actuator API ( #4363 )
4 年前
GitHub
134f548c
doc fix ( #4354 )
4 年前
GitHub
19a80b1c
Fix null reference error when passing a null array to an action segment. ( #4358 )
4 年前
GitHub
1f00faa5
Remove whitespace in markdown ( #4357 )
4 年前
GitHub
e3bc3352
[pytorch] Add decoders, distributions, encoders, layers, networks, and utils ( #4349 )
4 年前
GitHub
b51347ac
New Variable Speed Walker Environments ( #4301 )
* init
* Add reward manager and hurryUpReward
* fix hurry reward/ add awful first training
* Turn off head height and hurry rew
* changed max speed to 15. added small hh rew
* add NaN check for reward manager. start vel penalty
* add bpVel pen
* add new BPVelPen nn file
* remove outdated nn file
* add randomize speed bool
* try rewad product
* change coeff to 1
* try avg vel of all bp for reward
* move outside loop
* try linear inverselerp for vel
* add avg rew matchspeed15 nn file. looks much better
* save scene
* no hand penalty, random walk speed
* fix inverse lerp
* try new reward falloff
* cleanup
* added new nn file. don't allow hand contact
* update obsv
* remove hh rew. add trained no-hh model
* add new nn file
* new curve
* add new models. try no reset
* add hh rew
* clamp hh
* zero rewards if ground contact
* switch to approved with movi...
4 年前