-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
change batch_size
in sparsify
#639
Comments
hey @giovp @MUCDK, I ran some quick benchmark to capture the peak memory. From what I checked maximum memory allocated for If
however as in the linked issue For the benchmark setup.I ran with
Here is the code for benchmark (I ran it with |
Great, thanks. what is |
So seems like |
No, solve doesn't exceed 500mb. Here are the results with batch_size_sp=1, its similar to solve
|
yeah but seems like All in all, this means that we require |
Yes, that is correct. |
okay, now the question is whether it's faster when we decrease the batch size in ott-jax (i.e. PointCloud), and hence to increase the batch size in Any chance you could benchmark this? |
yep, here are the results
|
Hence, it's faster id we have a large batch size in Thus, if we use Maybe one last thing @selmanozleyen : did you convert the output to a |
yes for these values of m,n,d.
For the whole sparsify method, I am not sure if
No this was just for apply. |
Btw I tried this comparison as you told me. The difference is so much and the one with cpu begining takes so much that it times out of 6 minutes while other is done in 5 seconds. I just think that the slowness of the computation on cpu exceeds the cost copying from gpu to cpu. I can share more details once our clusters are faster, I did this quickly on colab
@giovp @MUCDK |
with the
I set these values for comparison with last year. The results don't seem that different except the peak memory might stablize as the |
that sounds good then, thanks a lot! |
Needed for keeping solver fast (high batch size), but prevent OOM in
sparsify
(need low batch size), see theislab/cellrank#1146 (comment)The text was updated successfully, but these errors were encountered: