Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[BF16]For LayoutLMForSequenceClassification model on stock pytorch, gelu cost time on pvc-1100 worse than A100 * ratio #800

Open
xiaowangintel opened this issue Aug 22, 2024 · 2 comments
Assignees
Milestone

Comments

@xiaowangintel
Copy link
Contributor

xiaowangintel commented Aug 22, 2024

🐛 Describe the bug

For more details, please refer to https://jira.devtools.intel.com/browse/PYTORCHDGQ-5064.

For more details, please refer to https://jira.devtools.intel.com/browse/PYTORCHDGQ-5089?filter=-2.

Versions

pytorch commit:03480213dea1f60f6d12e7348904d2f3ef7314d0
torch-xpu-ops commit:718bc42c667539977e5eadb11ea4dec602544bf2
driver : hotfix_agama-ci-devel-881.19
pti : l_intel-pti-dev_p_0.9.0.38_offline.sh
basekit : l_BaseKit_p_2024.2.1.100_offline.sh

@xytintel xytintel added the loops_kernel Loops Kernel Backbone label Sep 10, 2024
@chuanqi129 chuanqi129 added this to the PT2.6 milestone Oct 14, 2024
@retonym
Copy link
Contributor

retonym commented Nov 19, 2024

xpu performance is not targeted to PT 2.6

@retonym retonym modified the milestones: PT2.6, PT2.7 Nov 19, 2024
@weishi-deng
Copy link
Contributor

rerun this test and the perf for gelu is reasonable.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

5 participants