Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why the Gated CNN Blocks are not 24 layers? #5

Open
David-Ttao opened this issue May 14, 2024 · 9 comments
Open

Why the Gated CNN Blocks are not 24 layers? #5

David-Ttao opened this issue May 14, 2024 · 9 comments

Comments

@David-Ttao
Copy link

David-Ttao commented May 14, 2024

I think its necessary to set 24 layers of MambaOut in memory of Kobe Bryant.

@David-Ttao David-Ttao changed the title 什么罐头我说? Why the Gated CNN Blocks are not 24 layers? May 14, 2024
@Celestial-Bai
Copy link

They can also set 8 layers, but they did not either. Man!

@David-Ttao
Copy link
Author

They can also set 8 layers, but they did not either. Man!

its worth discussing and i think its necessary to reproduce the code and change the layers to test result.

@pvbvcv
Copy link

pvbvcv commented May 14, 2024

It will be a meaningful work!

@ChaohuanDeng123
Copy link

就你这个issue显得格格不入。伟大无需多言!

@LightwishWONG
Copy link

《changed the title 什么罐头我说? Why the Gated CNN Blocks are not 24 layers?》hhhhhh

@yuweihao
Copy link
Owner

yuweihao commented May 20, 2024

Thank you so much for your suggestion. We released MambaOut-Kobe model, a Kobe Memorial version with 24 Gated CNN blocks. MambaOut-Kobe achieves really competitive performance, surpassing ResNet-50 and ViT-S with much fewer parameters and FLOPs. For example, MambaOut-Kobe outperforms ViT-S by 0.2% accuracy with only 41% parameters and 33% FLOPs.

Model Resolution Params MACs Top1 Acc
ResNet-50
(ResNet strikes back)
224 25.5M 4.1G 79.8
ViT-S 224 22.1M 4.6G 79.8
MambaOut-Kobe 224 9.1M 1.5G 80.0

@Celestial-Bai
Copy link

Thank you so much for your suggestion. We released MambaOut-Kobe model, a Kobe Memorial version with 24 Gated CNN blocks. MambaOut-Kobe achieves really competitive performance, surpassing ResNet-50 and ViT-S with much fewer parameters and FLOPs. For example, MambaOut-Kobe outperforms ResNet-50 by 0.2% accuracy with only 36% parameters and MACs.

Model Resolution Params MACs Top1 Acc
ResNet-50 224 25.5M 4.1G 79.8*
ViT-S 224 22.1M 4.6G 79.8
MambaOut-Kobe 224 9.1M 1.5G 80.0

  • The result is cited from "ResNet strikes back" paper, a very strong version of ResNet trained for 300 epochs.

Man! Hahahaha

@David-Ttao
Copy link
Author

David-Ttao commented May 20, 2024

Thank you so much for your suggestion. We released MambaOut-Kobe model, a Kobe Memorial version with 24 Gated CNN blocks. MambaOut-Kobe achieves really competitive performance, surpassing ResNet-50 and ViT-S with much fewer parameters and FLOPs. For example, MambaOut-Kobe outperforms ResNet-50 by 0.2% accuracy with only 36% parameters and MACs.

Model Resolution Params MACs Top1 Acc
ResNet-50 224 25.5M 4.1G 79.8*
ViT-S 224 22.1M 4.6G 79.8
MambaOut-Kobe 224 9.1M 1.5G 80.0

  • The result is cited from "ResNet strikes back" paper, a very strong version of ResNet trained for 300 epochs.

its a meaningful work, what can i say?

@WindZh03
Copy link

What a great suggestion!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants