High memory consumption for inference #9528
Unanswered
PushpakBhoge
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have noticed that during inference memory consumption shoots up
take a look at this code
ideally, the
diff
should be approximately zero (a 10-30 MB difference is acceptable)but I found out that this difference is coming out very high varying from 400 MB to 1500 MB, and this consumption goes up to a few inferences then it settles
why is this happening? This actually causes issues when the model is deployed using the Torchserve the memory usage shoots up
Beta Was this translation helpful? Give feedback.
All reactions