Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Missing GPU kernels when using @profile and -b flag #312

Open
dwchang79 opened this issue Oct 2, 2023 · 2 comments
Open

Missing GPU kernels when using @profile and -b flag #312

dwchang79 opened this issue Oct 2, 2023 · 2 comments

Comments

@dwchang79
Copy link

dwchang79 commented Oct 2, 2023

I am using the @Profile and -b flag to try and remove the initial training section of a ML workload so that I can only profile the inference part. That is working, but the problem is the GPU kernels and information are now missing. The call stack shows the functions, but they do not link to the GPU and no GPU devices are shown nor is anything shown running on them.

I have attached two screenshots. One with the entire run (without the profile flags) where the GPU section is shown at the bottom as "HIP Activity Device 2, Queue 0" and a second screenshot where only the inference part is profiled, but the GPU information is now gone.

Thank you.
Complete
Inference

@jrmadsen
Copy link
Collaborator

Try prefixing the command with omnitrace-run -- python3 -m omnitrace -b -- <script> <script-args>. I suspect the later initialization of omnitrace due to the @profile is causing in omnitrace getting initialized after the hip runtime, resulting in omnitrace not getting registered as profiling tool for the HIP runtime.

@ppanchad-amd
Copy link

Hi @dwchang79. Has your issue been resolved? If so, please close the ticket. Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants