New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clarifications of the method #144
Comments
Hi Alex, thank you for your interest in the work. embedding layer
smooth cutoff
normalization
intermediate node embedding
equivariant message
scalar product
Hope this helps, feel free to ask if you have further questions. |
Hello, after reading the paper, I had several questions regarding your approach. Thanks a lot in advance for taking the time to answer them.
Your embedding layer is more complex than usual: your initial node representation already seems to depend on its neighbour’s representation.
Graph construction: you use a smooth cutoff function and describe some benefits. You describe a Transformers but still use a cutoff value.
You say the feature vector are passed through a normalization layer.
An intermediate node embedding (y_i) utilising attention scores is created and impact final x_i and v_i embeddings. This step weights a projection of each neighbor’s representation ~$a_{ij} (W \cdot RBF(d_{ij}) \cdot \vec{V}_j)$ by the attention score.
The equivariant message m_ij (component of sum to obtain w_i) is obtained by multiplying s_ij^2 (i.e. v_j scaled by RBF(d_ij)) by the directional info r_ij; then adding to it s_ij^1 (i.e. v_j scaled by RBF(d_ij)) re-multiplied by v_j.
Do you think that multiplying the message sequentially by distance info and directional info is the best choice to embed both info. type ? Why not concatenate r_ij (r_i - r_j) and d_ij (norm of r_ij = distance) info and have a single operation for instance ?
Is multiplying s_ij^1 by v_j (again) necessary ? (first in s_ij then by multiplying element-wise s_ij to v_j)
IMPORTANT. r_ij has dimension 3 while s_ij^2 has dimension F. In Eq (11), how can you apply an element-wise multiplication ? Is it a typo ? How exactly do you combine these two quantities ? What’s your take on the best way to combine 3D info (directional vector) with existing embedding ? This is a true question I am interested in, if you have references or insights on this bit…
Invariant representation involves the scalar product of the equivariant vector v_i, projected with matrix U1 by (U2 v_i).
The text was updated successfully, but these errors were encountered: