Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Reproducing classification results of EfficientFormer_L1] #34

Open
Amshaker opened this issue Dec 16, 2022 · 3 comments
Open

[Reproducing classification results of EfficientFormer_L1] #34

Amshaker opened this issue Dec 16, 2022 · 3 comments

Comments

@Amshaker
Copy link

Hi,

Thank you for sharing your research work. I am trying to reproduce the classification results of efficientformer_l1, I followed the instruction in your repo by running the following command:

sh dist_train.sh efficientformer_l1 8

However, the reproduced accuracy after training for the network on the 50000 test images: is 78.0% for 300 epochs. The result you mentioned in the paper is 79.2%. Could you please comment on that and let me know the reason?

I hope to hear from you soon,
Abdelrahman.

@liyy201912
Copy link
Collaborator

Hi,

In most of our experiments, we use 16gpu x 128bs to 16gpu x 256bs for training. We do notice that it is not easy to achieve best results when batch size is small (say 8x128). So there are two possible solutions,

  1. In our latest EfficientFormerV2 release, we add syncronized normalization, which may help to achieve more stable training results with small batch size.
  2. Use a larger batch size if there's sufficient VRAM.

@Amshaker
Copy link
Author

@liyy201912 Thank you for your reply and your clarification, I will check that.

I have another question please, can you provide more details about the Fast Latency-Driven Slimming algorithm? How many GPUs do you need to train it with as well as what is the running time for that? Is this algorithm open source or do you plan to release it soon?

@leonid-pishchulin
Copy link

+1 on that request

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants