We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 7b65989 commit 08e5e20Copy full SHA for 08e5e20
torchao/quantization/README.md
@@ -6,7 +6,7 @@ Typically quantization algorithms will have different schemes for how the activa
6
All the following benchmarks are for `meta-llama/Llama-3-8.1B` using `lm-eval` measured on an H100 GPU.
7
8
| weight | activation | wikitext-perplexity | winogrande | checkpoint size (GB) |
9
-| --------- | ------------------- | ---------- | -------------------- |
+| --------- | ------------------- | ---------- | -------------------- | -------- |
10
| bfloat16 | bfloat16 | 7.3315 | 0.7380 | 16.1 |
11
| float8_rowwise | float8_rowwise | 7.4197 | 0.7388 | 9.1 |
12
| int8_rowwise | bfloat16 | 7.3451 | 0.7340 | 9.1 |
0 commit comments