Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does reducing batch size affects convergence? #1

Open
pinakinathc opened this issue Dec 20, 2020 · 1 comment
Open

Does reducing batch size affects convergence? #1

pinakinathc opened this issue Dec 20, 2020 · 1 comment

Comments

@pinakinathc
Copy link

Due to a limit of 11GB GPU resource, I am forced to reduce the batch size for image2text to 8 instead of 128.
After training for a while, I was unable to get a converged trained model that replicated results in the paper.

Hence, my question is:

  1. How much GPU memory did you use?
  2. Does using smaller batch size leads to failure in training?
@pinakinathc
Copy link
Author

@s-mahajan can you upload the trained checkpoints as mentioned in the README.md file?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant