Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does this work for any model with activation function as relu? #34

Open
kailashg26 opened this issue Mar 3, 2022 · 4 comments
Open

Comments

@kailashg26
Copy link

Hello, I'm trying to use the actnn with maddpg (an MARL algorithm). The model just has 3 layers with activation function relu. If so, can you let us know if this mechanism will give the results with smaller models.

Thank you.

Link of maddpg https://github.com/marlbenchmark/off-policy/tree/release/offpolicy/algorithms/maddpg

@merrymercy
Copy link
Member

It should work with relu activation function, but we didn't test any RL tasks. Did you meet memory issues even with this small model?

@kailashg26
Copy link
Author

I mean, when I train the maddpg, almost 10GB of memory gets used. So, I wanted to try some compression functions. It would be a great help if you can provide some insights on how to test it with the maddpg if you have any idea.

Thanks

@merrymercy
Copy link
Member

merrymercy commented Mar 10, 2022

You can try to follow the usage and replace the layers in your model with actnn layers. You can start with higher bits and see whether the lossy compression hurts reward.

@kailashg26
Copy link
Author

Thank you. I'll look into it and post here if I have any doubts

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants