Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[NDArray] Map NDArray ops to GPUs #974

Merged
merged 5 commits into from
Dec 17, 2024
Merged

Conversation

tkarna
Copy link
Contributor

@tkarna tkarna commented Dec 3, 2024

Adds machinery to support GPU pipeline in Sharded Array for Python.

  • InsertGPUAllocsPass: add host-shared option, defaults to true.
  • Mark ndarray.copy ops with a special region environment ("gpu_copy_op") if either operand has a GPU environment.
  • Adds InsertGPUCopyPass to convert memref.copy ops within GPU env with gpu.memcpy ops.
  • Adds TileLoops pass to tile loops ops with tileUsingSCF. Apply twice to map to blocks and threads.
  • Adds RegionParallelLoopToGpu to convert parallel loops to GPU only if within a gpu env region.

Please review these guidelines to help with the review process:

  • Have you provided a meaningful PR description?
  • Have you added a test, a reproducer, or a reference to an issue with a reproducer?
  • Have you tested your changes locally for CPU and GPU devices?
  • Have you made sure that new changes do not introduce compiler warnings?
  • If this PR is a work in progress, are you filing the PR as a draft?
  • Have you organized your commits logically and ensured each can be built by itself?

@tkarna tkarna force-pushed the ndarray-gpu-upstream branch from b1d1811 to 412577e Compare December 17, 2024 09:23
@fschlimb fschlimb merged commit 67bf46f into intel:main Dec 17, 2024
2 checks passed
@tkarna tkarna deleted the ndarray-gpu-upstream branch December 17, 2024 12:27
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants