-
Updated
Jan 29, 2020 - Python
deep-reinforcement-learning
Here are 1,265 public repositories matching this topic...
In the Ray Design Agents documentation the Ray Perception parameter Ray Layer Mask
is not mentioned. I am a bit confused about what does it do and if it interacts with the Detectable Tags
parameter.
. I like the work and I respect every project that goes deep down on neural network
Our default placeholders sometimes cause confusion, like in this thread. Perhaps different placeholders, like "[email protected]"
and "YourTokenHere"
will help with that.
Another thread: https://www.coursera.org/learn/practical-rl/discussions/all/threads/jxK8jdhzQ6iSvI3Yc-Oouw
Please specify in the the Configuration section of the documentation that Windows users need to remove the SSH entry pertaining to 10.0.0.2 from C:\Users\CurrentUser.ssh\known_hosts every time they update pwnagotchi, since the newly generated keys won't match the one in the known_hosts file.
The OpenAI Gym installation instructions are missing reference to the "Build Tools for Visual Studio 2019" from the following site.
https://visualstudio.microsoft.com/downloads/
I also found this by reading the following article.
https://towardsdatascience.com/how-to-install-openai-gym-in-a-windows-environment-338969e24d30
Even though this is an issue in the OpenAI gym, a note in this RE
-
Updated
Apr 21, 2020 - Jupyter Notebook
-
Updated
Jun 22, 2020 - Python
I was surprised to see this loss function because it is generally used when the target is a distribution (i.e. sums to 1). This is not the case for the advantage estimate. However, I worked out the math and it does appear to be doing the right thing which is neat!
I think this trick should be mentioned in the code.
-
Updated
Feb 7, 2020 - Jupyter Notebook
-
Updated
Jun 23, 2020 - Jupyter Notebook
-
Updated
Oct 1, 2018
-
Updated
Jun 23, 2020 - Python
-
Updated
Mar 3, 2020 - Python
-
Updated
Jun 4, 2018 - Python
-
Updated
Jan 28, 2020 - Python
-
Updated
Jan 8, 2019 - JavaScript
-
Updated
May 10, 2020 - Python
-
Updated
Jan 28, 2020 - Python
-
Updated
Feb 2, 2020 - Python
-
Updated
May 22, 2020 - Python
-
Updated
Jan 17, 2020 - Python
-
Updated
Mar 18, 2020 - Python
-
Updated
Apr 25, 2020 - HTML
Improve this page
Add a description, image, and links to the deep-reinforcement-learning topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the deep-reinforcement-learning topic, visit your repo's landing page and select "manage topics."
There seems to be conflicting data on what the "OriginGeopoint" is. In the documentation it's referenced as the location of the PlayerStart while in code it's commented as the coordinates of Unreal level at the coordinates 0,0,0.
Code:
https://github.com/microsoft/AirSim/blob/e24a11d5985c1d6f573735f28a3844bc634559db/AirLib/include/common/AirSimSettings.hpp#L350
Documentation:
https://micro