-
Notifications
You must be signed in to change notification settings - Fork 539
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hyperparameters for ImageNet training? #120
Comments
Hi @InugYoon, I assume you mean the implementation with the momentum encoder trick. If so, I adopt most hyper-parameters from MoCo-v2. IIRC, the differences are that: (1) I used a queue size of 8192; (2) temperature is 0.07; (3) batch size of 1024. If you want to be faithful to the paper (not using the momentum encoder trick), please consider following the paper. |
Hello @HobbitLong, thank you for quick reply. First of all, I wanted to re-implement the results based on code here, without using Moco-trick, following the paper. However for the ImageNet, I couldn't find the listed hyperparameters both on github or the paper. Now with your kind reply, I noticed that you used the hyperparameters from Moco-v2 IIRC. About the moco-v2 hyperparameters, did you used from here? |
Hello @HobbitLong, I have incorporated most of the hyper-parameters from MoCo-v2, which align with the ones you previously mentioned. However, there appears to be a gap in accuracy, and I postulate that this may be due to the number of epochs. Thank you for your time and consideration. |
Hello @kiimmm, Would you please release your MoCo version code? I'm trying to reimplement it but got stuck in the MoCo code. Thanks. |
Hi, I am trying to reproduce the results.
May I get the hyperparameters for ImageNet experiment?
The text was updated successfully, but these errors were encountered: