You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm trying to correctly express a single GP model applied independently along the batch dimension (as if by a for-loop with traditional_batch_length iterations), but have failed. Reading code, documentation, and examples, I'm finding myself confused by the meaning of batch_shape in GPyTorch (and how it relates to the traditional meaning of the term batch in deep learning), how it relates to "multitask", and by the different variational strategies. I'd appreciate any help. Please excuse the venue.
More specifically, I'm doing so in the context of DUE. The following is a snippet from within DUE, slightly adapted for brevity.
Applying this as follows (eventually, considering lazy eval. here and there) produces a CUDA OOM error, due to a dense covariance matrix with side traditional_batch_length.
Here, by traditional_batch_length I mean that this is a traditional batch in the sense common in deep learning.
Needless to say, this also fails the same way with num_outputs > 1.
The text was updated successfully, but these errors were encountered:
I'm trying to correctly express a single GP model applied independently along the batch dimension (as if by a for-loop with
traditional_batch_length
iterations), but have failed. Reading code, documentation, and examples, I'm finding myself confused by the meaning ofbatch_shape
in GPyTorch (and how it relates to the traditional meaning of the term batch in deep learning), how it relates to "multitask", and by the different variational strategies. I'd appreciate any help. Please excuse the venue.More specifically, I'm doing so in the context of DUE. The following is a snippet from within DUE, slightly adapted for brevity.
Applying this as follows (eventually, considering lazy eval. here and there) produces a CUDA OOM error, due to a dense covariance matrix with side
traditional_batch_length
.Here, by
traditional_batch_length
I mean that this is a traditional batch in the sense common in deep learning.Needless to say, this also fails the same way with
num_outputs
> 1.The text was updated successfully, but these errors were encountered: