Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Translate augmentation diverges #9

Closed
HosseinSheikhi opened this issue Oct 28, 2020 · 7 comments
Closed

Translate augmentation diverges #9

HosseinSheikhi opened this issue Oct 28, 2020 · 7 comments

Comments

@HosseinSheikhi
Copy link

Hello,
I wonder if I have to do fine-tunings to get results from Translate augmentation. It always diverges! I have tested for Cartpole, Walker, and Cheetah.
In the following figures, the diverged one is Translate.
rsz_screenshot_from_2020-10-27_20-26-19
rsz_screenshot_from_2020-10-27_20-25-48

@MishaLaskin
Copy link
Owner

Hi @HosseinSheikhi thanks for bringing this to our attention, that shouldn't be the case so there may have been an issue with the translate PR. @WendyShang can you check if your PR was incorporated correctly? Maybe it's using different hyperparams?

@WendyShang
Copy link

Hi @HosseinSheikhi, the plot of my run from scripts/cheetah_test.sh is attached. Could you share your version of pytorch as well as your scripts in producing those diverged plots?

Screenshot from 2020-10-28 21-47-37

@HosseinSheikhi
Copy link
Author

Thank you both for your prompt reply. The PyTorch version is 1.6.0 .
Here it is the scripts (I run Cartpole pretty much like Walker, just changing the action repeat to 8):

CUDA_VISIBLE_DEVICES=0 python train.py
--domain_name cheetah
--task_name run
--encoder_type pixel --work_dir ./tmp
--action_repeat 4 --num_eval_episodes 10
--pre_transform_image_size 100 --image_size 84
--agent rad_sac --frame_stack 3 --data_augs translate
--seed 23 --critic_lr 2e-4 --actor_lr 1e-3 --eval_freq 10000 --batch_size 128 --num_train_steps 500000

UDA_VISIBLE_DEVICES=0 python train.py
--domain_name walker
--task_name walk
--encoder_type pixel --work_dir ./tmp
--action_repeat 2 --num_eval_episodes 10
--pre_transform_image_size 100 --image_size 84
--agent rad_sac --frame_stack 3 --data_augs translate
--seed 23 --critic_lr 1e-3 --actor_lr 1e-3 --eval_freq 10000 --batch_size 128 --num_train_steps 500000

@WendyShang
Copy link

WendyShang commented Oct 29, 2020

@HosseinSheikhi For Cheetah Run, could you please try and let me know how the training curves look:
CUDA_VISIBLE_DEVICES=0 python train.py
--domain_name cheetah
--task_name run
--encoder_type pixel --work_dir ./tmp
--action_repeat 4 --num_eval_episodes 10
--pre_transform_image_size 100 --image_size 108
--agent rad_sac --frame_stack 3 --data_augs translate
--seed 23 --critic_lr 2e-4 --actor_lr 2e-4 --eval_freq 10000
--batch_size 128 --num_train_steps 600000 --init_steps 10000
--num_filters 32 --encoder_feature_dim 64 --replay_buffer_capacity 100000 \

@HosseinSheikhi
Copy link
Author

I will update you, but just in case, pre_transform_image_size should not be greater than image_size?

@MishaLaskin
Copy link
Owner

MishaLaskin commented Oct 30, 2020 via email

@HosseinSheikhi
Copy link
Author

Yes, that was the reason, now its converging. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants