Skip to content

SYCL: Refactor and enable FP16 in binary broadcast OPs #12975

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 5 commits into from
Apr 18, 2025
Merged

Conversation

qnixsynapse
Copy link
Collaborator

@qnixsynapse qnixsynapse commented Apr 16, 2025

Somehow the original dispatcher logic was incorrect. Refactored and enabled FP16 on these operations.

I have structured this branch in a way so that each commit can be inspected properly by reviewers.

test-backend-ops is passing with this change.

@github-actions github-actions bot added ggml changes relating to the ggml tensor library for machine learning SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language labels Apr 16, 2025
@qnixsynapse qnixsynapse requested a review from Rbiessy April 18, 2025 08:01
Copy link
Collaborator

@Rbiessy Rbiessy left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM thanks for making this easy to review!

@Rbiessy Rbiessy merged commit 8d66005 into master Apr 18, 2025
48 of 51 checks passed
@Rbiessy Rbiessy deleted the binary-ops branch April 18, 2025 13:58
colout pushed a commit to colout/llama.cpp that referenced this pull request Apr 21, 2025
* SYCL: refactor move to a separate file

* Fix binbcast

* Remove duplicates

* fix include formatting

* fix typo
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ggml changes relating to the ggml tensor library for machine learning SYCL https://en.wikipedia.org/wiki/SYCL - GPU programming language
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants