-
Notifications
You must be signed in to change notification settings - Fork 981
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Speed up model.generate() with coca? #475
Comments
@Pclanglais Hi, I will get to work on 1 as soon as I can, as it is not possible right away. For 2 did you try setting |
hello @gpucce Thanks a lot. For 1. I just wanted to be sure that I hadn't missed any option but I could fork it on my side. It's a very good idea for 2 : I'm going to test it right away. |
Duplicate of #409 This is an important issue to fix for usability |
@Pclanglais Maybe a bit late, but if you aren't batching yet you can try #498. When I try replicating your findings, assuming GPU, I'm getting around a 100 images processed in around 40 seconds when batch size is 1. You can already batch with model.generate(), however I hoped to make a easier for future use in the PR. |
I am building an image classification workflow on top of coca captions and embeddings. The only downside is that this is slow (about 100/images per minute on a google colab).
So two related questions:
The text was updated successfully, but these errors were encountered: