Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

release gpu memory #324

Open
Myson850 opened this issue Jul 25, 2024 · 1 comment
Open

release gpu memory #324

Myson850 opened this issue Jul 25, 2024 · 1 comment
Labels
wontfix This will not be worked on

Comments

@Myson850
Copy link

Feature request

Release gpu memory after a certain number of calls

Motivation

After setting the --batch-size of the embed model to 100, I tried to call the data with a batch size of 80. It succeeded, but then I called the data with a batch size of 10 many times. The GPU occupied the video memory and did not release or lower

Your contribution

.

Repository owner deleted a comment from greptile-apps bot Jul 25, 2024
@michaelfeil
Copy link
Owner

Set batch_size to e.g. 32 (multiple of 8 and pow 2 encouraged) for better usage. Sorry, but your request does not make much sense. YOu also have computation graphs in torch.compile that make your proposed feature very unattractive.

@michaelfeil michaelfeil added the wontfix This will not be worked on label Jul 25, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
wontfix This will not be worked on
Projects
None yet
Development

No branches or pull requests

2 participants