Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Custom PA perf improvements #222

Merged
merged 13 commits into from
Oct 8, 2024
Merged

Custom PA perf improvements #222

merged 13 commits into from
Oct 8, 2024

Conversation

sanyalington
Copy link

@sanyalington sanyalington commented Oct 4, 2024

Enable 128K context length in custom PA.
Enable custom PA to write fp8 output with scaling, enabled this perf optimization for LLama. This optimization is only enabled on rocm custom PA when chunked prefill is disabled and environment variable VLLM_USE_ROCM_CUSTOM_PAGED_ATTN_FP8_OUT=1

Copy link
Collaborator

@shajrawi shajrawi left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

ship it

@shajrawi shajrawi merged commit b51fe69 into main Oct 8, 2024
16 of 17 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants