Skip to content

Commit

Permalink
nit
Browse files Browse the repository at this point in the history
  • Loading branch information
robertgshaw2-redhat committed Jul 8, 2024
1 parent 3868193 commit ccbe6db
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion docs/conceptual_guides/inference_acceleration.md
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ With weight-only quantization, weights are quantized to low precision (typically

### How Can We Speed Up Weight-Only Quantization?

Roughly speaking, the time required to execute a matrix multiplication on a GPU equals:
Roughly speaking, the time required to execute a matrix multiplication on a GPU equals the sum of:
* Latency of moving the weights from main memory (DRAM) to the compute (SRAM)
* Latency of the tensor-core compute operations

Expand Down

0 comments on commit ccbe6db

Please sign in to comment.