-
Notifications
You must be signed in to change notification settings - Fork 46
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Run v0.1 embeddings to store raw encoder output #127
Comments
Thanks. I.e.
Note: These grouping do NOT mean that the model has parallel track for each group. When training we calculate the self-attention (QKV) individually for each layer. Groups, are more akin to sentences, groups of words. This means that the RGB group also has information about what SAR has, and vice versa. @srmsoumya to confirm this. If this is the case, I don't understand the value of the grouping instead of making one embedding per self-attention patch. What's the value of grouping the embeddings this way? TLDR for @MaceGrim. The semantic resolution at the self-attention path before the average is |
Thanks. I.e.
Note: These grouping do NOT mean that the model has parallel track for each group. When training we calculate the self-attention (QKV) individually for each layer. Groups, are more akin to sentences, groups of words. This means that the RGB group also has information about that SAR has, and vice versa. @srmsoumya to confirm this. If this is the case, I don't understand the value of the grouping instead of making one embedding per self-attention patch across all bands. What's the value of grouping the embeddings this way? Would not make sense to reduce the semantics across all bands into one? TLDR for @MaceGrim the semantic resolution at the self-attention path is |
We are already running on the previous version that only stores average embeddings. So creating the raw embeddings might be scheduled later in tandem with other model updates. |
The option to output this has been implemented in #133 . We have multiple people running patch embeddings for specific use cases. So we can close this high level issue here. |
To enable fast downstream applications, we could store the raw encoder output and not only the average embedding that we are already creating for similarity search.
Refs
The text was updated successfully, but these errors were encountered: