-
Notifications
You must be signed in to change notification settings - Fork 75.3k
LeakyReLU uses up too much memory. #4079
Copy link
Copy link
Closed
Description
Right now I am implementing leaky relus like this tf.maximum(0.1 * x, x). This works fine except when it comes to memory usage. Networks which will fit on my GPU when using tf.nn.relu or tf.nn.elu fail when I am using my leaky relu implementation. I think this is because it needs to store both the intermediate 0.1 * x and x values of the activations to compute the gradients which essentially does the memory usage. However, I do not think this would be an issue if there were a dedicated tf.nn.leaky_relu. Can someone consider adding this to a future tensorflow release.
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels