-
Notifications
You must be signed in to change notification settings - Fork 97
Issues: pytorch/ao
[RFC] Which low bit CUDA kernels should we merge or write?
#697
opened Aug 17, 2024 by
msaroufim
Open
11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
More fine-grained documentation needed for
torchao.autoquant()
#907
opened Sep 19, 2024 by
suvadityamuk
we should ensure activation checkpointing with Float8Linear behaves optimally
#893
opened Sep 16, 2024 by
vkuzo
Expected Tensor argument scales to have dtype torch.bfloat16, but got torch.float32 instead
#876
opened Sep 11, 2024 by
agunapal
NotImplementedError: aten.linear.default not implemented when using MXTensor
mx
#796
opened Sep 3, 2024 by
Ali-Flt
[Question] Is there any plans to support fp8 batched matmul (
_scaled_bmm
)
float8
#781
opened Aug 30, 2024 by
leeeizhang
How does this work with ONNX export and quantization?
export
inference
#777
opened Aug 29, 2024 by
ogencoglu
All code snippets on README should be self contained or link to single file runnable code
#765
opened Aug 27, 2024 by
msaroufim
2 tasks
Previous Next
ProTip!
Add no:assignee to see everything that’s not assigned.