Skip to content

Commit

Permalink
pre-commit
Browse files Browse the repository at this point in the history
  • Loading branch information
dan-garvey committed Nov 26, 2024
1 parent 6257639 commit 14e67f6
Showing 1 changed file with 3 additions and 3 deletions.
6 changes: 3 additions & 3 deletions sharktank/sharktank/layers/linear.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,8 @@ class LinearLayer(ThetaLayer):
x = x * premul_input
matmul(x, weight.T) + bias
fake quant only exists in order to allow for q_input to act as qdq.
when fake quant is false, q_input will quantize normally.
fake quant only exists in order to allow for q_input to act as qdq.
when fake quant is false, q_input will quantize normally.
```
"""

Expand Down Expand Up @@ -80,7 +80,7 @@ def forward(self, x):
y = ops.linear(x, weight, bias)

# Unconditionally dequantize.
if isinstance(y, QuantizedTensor):
if isinstance(y, QuantizedTensor):
y = y.unpack().dequant()
# Note that f8_e4m3fnuz types on AMD GPUs accumulate to fp32.
# We can truncate to fp16 in iree, so we do a cast here
Expand Down

0 comments on commit 14e67f6

Please sign in to comment.