[Quantization] Channel-wise Output Activation Quantization for Attention QKV Modules + KV-cache channel quantization#1233
Draft
horheynm wants to merge 9 commits intomainfrom attn_quant
+39-6
Commits
Commits on Feb 11, 2025
- committed
Commits on Mar 6, 2025
Commits on Mar 7, 2025
- committed
- committed
- authored
- committed
- committed
- committed