Skip to content

Release v0.4 #884

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 86 commits into from
Jun 16, 2018
Merged

Release v0.4 #884

merged 86 commits into from
Jun 16, 2018

Conversation

awjuliani
Copy link
Contributor

No description provided.

sterlingcrispin and others added 30 commits April 19, 2018 11:20
No game object is named "Platform" therefore the legs never make contact with anything
* Report means instead of totals for losses.
* Report absolute loss for policy.
* [Fix] Use the stored agent info instead of the previous agent info when bootstraping the value

* [Bug Fix] Addressed #643

* [Added Line Break]
Quick start integrated into Installation
This PR makes the following changes:

* Moves clipping of continuous control model into model itself. Output is now always [-1, 1].
* Internal model values are now clipped between [-3, 3] before being rescaled to [-1, 1] for output.  * This improves training performance by providing a wider range of values within which the pdf of the gaussian can fall. Output of [-1, 1] is used to be more environment-creator friendly.
* Fixes issue where epsilon was erroneously being used to reconstruct old probabilities during PPO update, leading to reduced learning performance.
* Introduce ScaleAction() function within python to easily rescale values from [-1, 1] to arbitrary range.
* Re-train all CC models using improved algorithm. All performance levels are equal or improved. In the case of Crawler, improvement is drastic.
* Update documentation appropriately.
* Made miscellaneous minor code style and optimization improvements within environments.
* [Cold Fix] Split the way cummulative rewards and episode length are counted
The reward is appended at each step to the cummulative reward
The episode count is ONLY incremented when d_t+1 is false
* Add `Walker` example environment and documentation.
* First draft of Azure support docs

* Correcting links to other docs

* Adding additional links and cleaning instructions

* Adding references to Azure docs in other appropriate places
Fixes the issue raised by @hsaikia in #552
Added the memory_size variable to the BC model
Added memory_size and recurrent_out to the output nodes of the graph when using BC with LSTM
- Indent the section about providing actions to multiple brains to be in line with the rest of the step() docs.
 - Move the line about what step() returns closer to the top of the docs so it's harder to overlook.
 - Add a small code snippet about how to get BrainInfo belonging to a specific brain and how to get data from that BrainInfo object.
* [Refactor] Fixed line indentation
* Removed the library Newtonsoft.Json from the monitor
* Replaced calls to JSON converstion with manual conversion
* [Modified] The Monitor now has multiple
* Log methods that take different object types
* [Added Ascii art on learn.py]
Note : This is by far the best feature of 0.4
Some suggestions to avoid ambiguity
Shouldn't Done(); be placed after the rewards are given?
Replaced UNITY ML AGENTS with the unity logo
awjuliani and others added 19 commits June 15, 2018 14:50
* Fix BananaIL frozen agent material

* [Fix] Added texture on the imitation learning scene and linked the models to the internal brains
Also made the look code into Update
…l-bugfix

[Hotfix] Remove code in PushAgentBasic that presumes that there is only one brain
* Added missing declaration to docs sample code.

* Added pretrained model as default graph in Internal brain of Tennis scene

* Disabled PlayerBrain in Tennis by default.

* Removed accidental config.
…model-fix

some hack to make windows save the model when do ctrl+c
* Remove extra bouncer brain hyperparameters

* Add error when using curiosity+odd
@eshvk
Copy link
Contributor

eshvk commented Jun 16, 2018

:+1

Copy link
Contributor

@xiaomaogy xiaomaogy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

+1

@xiaomaogy xiaomaogy merged commit 20569f9 into master Jun 16, 2018
@awjuliani awjuliani deleted the release-v0.4 branch June 29, 2018 18:04
@github-actions github-actions bot locked as resolved and limited conversation to collaborators May 19, 2021
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

10 participants