# Table of Contents

* [mlagents\_envs.envs.unity\_gym\_env](#mlagents_envs.envs.unity_gym_env)
  * [UnityGymException](#mlagents_envs.envs.unity_gym_env.UnityGymException)
  * [UnityToGymWrapper](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper)
    * [\_\_init\_\_](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.__init__)
    * [reset](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.reset)
    * [step](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.step)
    * [render](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.render)
    * [close](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.close)
    * [seed](#mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.seed)
  * [ActionFlattener](#mlagents_envs.envs.unity_gym_env.ActionFlattener)
    * [\_\_init\_\_](#mlagents_envs.envs.unity_gym_env.ActionFlattener.__init__)
    * [lookup\_action](#mlagents_envs.envs.unity_gym_env.ActionFlattener.lookup_action)

<a name="mlagents_envs.envs.unity_gym_env"></a>
# mlagents\_envs.envs.unity\_gym\_env

<a name="mlagents_envs.envs.unity_gym_env.UnityGymException"></a>
## UnityGymException Objects

```python
class UnityGymException(error.Error)
```

Any error related to the gym wrapper of ml-agents.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper"></a>
## UnityToGymWrapper Objects

```python
class UnityToGymWrapper(gym.Env)
```

Provides Gym wrapper for Unity Learning Environments.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.__init__"></a>
#### \_\_init\_\_

```python
 | __init__(unity_env: BaseEnv, uint8_visual: bool = False, flatten_branched: bool = False, allow_multiple_obs: bool = False, action_space_seed: Optional[int] = None)
```

Environment initialization

**Arguments**:

- `unity_env`: The Unity BaseEnv to be wrapped in the gym. Will be closed when the UnityToGymWrapper closes.
- `uint8_visual`: Return visual observations as uint8 (0-255) matrices instead of float (0.0-1.0).
- `flatten_branched`: If True, turn branched discrete action spaces into a Discrete space rather than
    MultiDiscrete.
- `allow_multiple_obs`: If True, return a list of np.ndarrays as observations with the first elements
    containing the visual observations and the last element containing the array of vector observations.
    If False, returns a single np.ndarray containing either only a single visual observation or the array of
    vector observations.
- `action_space_seed`: If non-None, will be used to set the random seed on created gym.Space instances.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.reset"></a>
#### reset

```python
 | reset() -> Union[List[np.ndarray], np.ndarray]
```

Resets the state of the environment and returns an initial observation.
Returns: observation (object/list): the initial observation of the
space.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.step"></a>
#### step

```python
 | step(action: List[Any]) -> GymStepResult
```

Run one timestep of the environment's dynamics. When end of
episode is reached, you are responsible for calling `reset()`
to reset this environment's state.
Accepts an action and returns a tuple (observation, reward, done, info).

**Arguments**:

- `action` _object/list_ - an action provided by the environment

**Returns**:

- `observation` _object/list_ - agent's observation of the current environment
  reward (float/list) : amount of reward returned after previous action
- `done` _boolean/list_ - whether the episode has ended.
- `info` _dict_ - contains auxiliary diagnostic information.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.render"></a>
#### render

```python
 | render(mode="rgb_array")
```

Return the latest visual observations.
Note that it will not render a new frame of the environment.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.close"></a>
#### close

```python
 | close() -> None
```

Override _close in your subclass to perform any necessary cleanup.
Environments will automatically close() themselves when
garbage collected or when the program exits.

<a name="mlagents_envs.envs.unity_gym_env.UnityToGymWrapper.seed"></a>
#### seed

```python
 | seed(seed: Any = None) -> None
```

Sets the seed for this env's random number generator(s).
Currently not implemented.

<a name="mlagents_envs.envs.unity_gym_env.ActionFlattener"></a>
## ActionFlattener Objects

```python
class ActionFlattener()
```

Flattens branched discrete action spaces into single-branch discrete action spaces.

<a name="mlagents_envs.envs.unity_gym_env.ActionFlattener.__init__"></a>
#### \_\_init\_\_

```python
 | __init__(branched_action_space)
```

Initialize the flattener.

**Arguments**:

- `branched_action_space`: A List containing the sizes of each branch of the action
space, e.g. [2,3,3] for three branches with size 2, 3, and 3 respectively.

<a name="mlagents_envs.envs.unity_gym_env.ActionFlattener.lookup_action"></a>
#### lookup\_action

```python
 | lookup_action(action)
```

Convert a scalar discrete action into a unique set of branched actions.

**Arguments**:

- `action`: A scalar value representing one of the discrete actions.

**Returns**:

The List containing the branched actions.