-
Notifications
You must be signed in to change notification settings - Fork 330
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Channel/Spatial/Element wise Attention Modules #44
Comments
I have not extensively checked about these in the model gardens components, but just picking the First One in the list: |
Currently we are prioritizing components that are required to achieve state of the art results on specific tasks. I.e., imagenet1k classification, COCO object detection, etc. Any chance I could get some guidance as to where these components excel? Thanks |
@LukeWood
It can be considered as a to-do list, an interested contributor can get references from here in the coming days. Let me know if I miss something. Also, I think a few components should be already on the current priority list, for example, Squeeze-and-Excitation or CBAM. |
II think that in this specific field you need to always find a balance between popupalirity, state of the art, sedimentation and the available human and computing resources you have at a specific point on the time dimension in your community. E.g. we are still investing resoruces for resnext-rs and waiting for the citation/popularity threshold for the next STOA: Often an high popupalirity component/model is needed to be maintained just cause it is a reccurent baseline for new accademic work. Another addition dimension/tension Is the computing resources required by a model and its own components. |
@innat @LukeWood
Could you post these separately so people could be assigned if interested and maybe some of the model could be prioritized. |
Ah, also can SWIN transformer be added to the list? |
@old-school-kid I didn't make a separate post regarding the above modules because I was not sure (and still) whether it's welcomed or we may need some discussion before approaching. The discussion thread wasn't created when I posted it. Maybe that's the right place. But I think, if it's welcomed, the interested contributor may pick up their interest from the above list and send PR. For Swin-Transformer, it's been asked, #15545 |
This issue was moved to a discussion.
You can continue the conversation there. Go to discussion →
In the timm package, it provides some soft attention modules to building network blocks and I think it's a good fit here, for example:
and many others.
The text was updated successfully, but these errors were encountered: