Skip to content
Branch: master
Find file Copy path
Find file Copy path
Fetching contributors…
Cannot retrieve contributors at this time
119 lines (96 sloc) 4.38 KB
#!/usr/bin/env python3
import gym
import numpy as np
from gym.envs.mujoco.mujoco_env import MujocoEnv
from learn2learn.gym.envs.meta_env import MetaEnv
class AntDirectionEnv(MetaEnv, MujocoEnv, gym.utils.EzPickle):
This environment requires the Ant to learn to run in a random direction in the
XY plane. At each time step the ant receives a signal composed of a
control cost and a reward equal to its average velocity in the direction
of the plane. The tasks are 2d-arrays sampled uniformly along the unit circle.
The target direction is indicated by the vector from the origin to the sampled point.
The velocity is calculated as the distance (in the target direction) of the ant's torso
position before and after taking the specified action divided by a small value dt.
As noted in [1], a small positive bonus is added to the reward to stop the ant from
prematurely ending the episode.
Adapted from Jonas Rothfuss' implementation.
1. Finn et al. 2017. "Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks." arXiv [cs.LG].
2. Rothfuss et al. 2018. "ProMP: Proximal Meta-Policy Search." arXiv [cs.LG].
def __init__(self, task=None):
MetaEnv.__init__(self, task)
MujocoEnv.__init__(self, 'ant.xml', 5)
# -------- MetaEnv Methods --------
def set_task(self, task):
MetaEnv.set_task(self, task)
self.goal_direction = task['direction']
def sample_tasks(self, num_tasks):
directions = np.random.normal(size=(num_tasks, 2))
directions /= np.linalg.norm(directions, axis=1)[..., np.newaxis]
tasks = [{'direction': direction} for direction in directions]
return tasks
# -------- Mujoco Methods --------
def _get_obs(self):
return np.concatenate([[2:],,
np.clip(, -1, 1).flat,
def viewer_setup(self):
camera_id = self.model.camera_name2id('track') = 2 = camera_id = self.model.stat.extent * 0.5
# Hide the overlay
self.viewer._hide_overlay = True
def reset_model(self):
qpos = self.init_qpos + self.np_random.uniform(size=self.model.nq, low=-.1, high=.1)
qvel = self.init_qvel + self.np_random.randn(self.model.nv) * .1
self.set_state(qpos, qvel)
# -------- Gym Methods --------
def step(self, action):
posbefore = np.copy(self.get_body_com("torso")[:2])
self.do_simulation(action, self.frame_skip)
posafter = self.get_body_com("torso")[:2]
forward_reward = np.sum(self.goal_direction * (posafter - posbefore)) / self.dt
ctrl_cost = .5 * np.square(action).sum()
contact_cost = 0.5 * 1e-3 * np.sum(
np.square(np.clip(, -1, 1)))
survive_reward = 1.0
reward = forward_reward - ctrl_cost - contact_cost + survive_reward
state = self.state_vector()
notdone = np.isfinite(state).all() and 1.0 >= state[2] >= 0.
done = not notdone
ob = self._get_obs()
return ob, reward, done, dict(
def reset(self, *args, **kwargs):
MujocoEnv.reset(self, *args, **kwargs)
return self._get_obs()
def render(self, mode='human'):
if mode == 'rgb_array':
# window size used for old mujoco-py:
width, height = 500, 500
data = self._get_viewer(mode).read_pixels(width,
return data
elif mode == 'human':
if __name__ == '__main__':
env = AntDirectionEnv()
for task in [env.get_task(), env.sample_tasks(1)[0]]:
action = env.action_space.sample()
You can’t perform that action at this time.