You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The attention layer we use computes attention for a certain dimension in a 4D vector.
If you want to compute attention on other dimensions of a 4D vector, you can use the transpose or permute operations to change the dimension order of the vector.
If the input is a 3D vector, I think you can consider simplifying the calculation equation for attention.
For other layers, you may need to redesign the network structure according to your specific situation.
Hi there,
is there a way how to adjust the attention based layers to train the network with differently shaped graph data?
thanks.
The text was updated successfully, but these errors were encountered: