From ec836c9b8b24f09b0f6eae1b1f54adb4a2cd048c Mon Sep 17 00:00:00 2001 From: Ruo-Ping Dong Date: Sat, 22 May 2021 13:12:02 -0700 Subject: [PATCH 1/2] remove duplicated Policy in glossary --- docs/Glossary.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/Glossary.md b/docs/Glossary.md index 61e0891fa7..d4c709ff03 100644 --- a/docs/Glossary.md +++ b/docs/Glossary.md @@ -7,7 +7,8 @@ - **Agent** - Unity Component which produces observations and takes actions in the environment. Agents actions are determined by decisions produced by a Policy. -- **Policy** - The decision making mechanism, typically a neural network model. +- **Policy** - The decision making mechanism for producing decisions from + observations, typically a neural network model. - **Decision** - The specification produced by a Policy for an action to be carried out given an observation. - **Editor** - The Unity Editor, which may include any pane (e.g. Hierarchy, @@ -19,7 +20,6 @@ Corresponds to each `Update` call of the game engine. - **Observation** - Partial information describing the state of the environment available to a given agent. (e.g. Vector, Visual) -- **Policy** - Function for producing decisions from observations. - **Reward** - Signal provided at every step used to indicate desirability of an agent’s action within the current state of the environment. - **State** - The underlying properties of the environment (including all agents From a66b0acdea83fbe3c3ce3c1249d61fb3bd0e5b4d Mon Sep 17 00:00:00 2001 From: Ruo-Ping Dong Date: Tue, 25 May 2021 11:11:20 -0700 Subject: [PATCH 2/2] sort alphabetically --- docs/Glossary.md | 16 ++++++++-------- 1 file changed, 8 insertions(+), 8 deletions(-) diff --git a/docs/Glossary.md b/docs/Glossary.md index d4c709ff03..7d912fe296 100644 --- a/docs/Glossary.md +++ b/docs/Glossary.md @@ -7,30 +7,30 @@ - **Agent** - Unity Component which produces observations and takes actions in the environment. Agents actions are determined by decisions produced by a Policy. -- **Policy** - The decision making mechanism for producing decisions from - observations, typically a neural network model. - **Decision** - The specification produced by a Policy for an action to be carried out given an observation. - **Editor** - The Unity Editor, which may include any pane (e.g. Hierarchy, Scene, Inspector). - **Environment** - The Unity scene which contains Agents. +- **Experience** - Corresponds to a tuple of [Agent observations, actions, + rewards] of a single Agent obtained after a Step. +- **External Coordinator** - ML-Agents class responsible for communication with + outside processes (in this case, the Python API). - **FixedUpdate** - Unity method called each time the game engine is stepped. ML-Agents logic should be placed here. - **Frame** - An instance of rendering the main camera for the display. Corresponds to each `Update` call of the game engine. - **Observation** - Partial information describing the state of the environment available to a given agent. (e.g. Vector, Visual) +- **Policy** - The decision making mechanism for producing decisions from + observations, typically a neural network model. - **Reward** - Signal provided at every step used to indicate desirability of an agent’s action within the current state of the environment. - **State** - The underlying properties of the environment (including all agents within it) at a given time. - **Step** - Corresponds to an atomic change of the engine that happens between Agent decisions. -- **Experience** - Corresponds to a tuple of [Agent observations, actions, - rewards] of a single Agent obtained after a Step. -- **Update** - Unity function called each time a frame is rendered. ML-Agents - logic should not be placed here. -- **External Coordinator** - ML-Agents class responsible for communication with - outside processes (in this case, the Python API). - **Trainer** - Python class which is responsible for training a given group of Agents. +- **Update** - Unity function called each time a frame is rendered. ML-Agents + logic should not be placed here.