Skip to content

Commit a7d3636

Browse files
Jack-Khuumalfet
authored andcommitted
Fix quantization.md Table (#876)
1 parent 3b97eb8 commit a7d3636

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

docs/quantization.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ While quantization can potentially degrade the model's performance, the methods
1515
## Supported Quantization Schemes
1616
### Weight Quantization
1717
| compression | bitwidth| group size | dynamic activation quantization | Eager | AOTI | ExecuTorch |
18-
|--|--|--|--|--|--|--|--|
18+
|--|--|--|--|--|--|--|
1919
| linear (asymmetric) | [8, 4]* | [32, 64, 128, 256]** | ||| 🚧 |
2020
| linear with dynamic activations (symmetric) | | [32, 64, 128, 256]* | a8w4dq | 🚧 |🚧 ||
2121

@@ -26,7 +26,7 @@ quantizing the embeddings to further reduce the model size for
2626
on-device usecases.
2727

2828
| compression | weight quantization (bitwidth)| weight quantization (group size) | dynamic activation quantization | Eager | AOTI | ExecuTorch |
29-
|--|--|--|--|--|--|--|--|
29+
|--|--|--|--|--|--|--|
3030
| embedding (symmetric) | [8, 4]* | [32, 64, 128, 256]+ | ||||
3131

3232

0 commit comments

Comments
 (0)