Skip to content

Commit df587a3

Browse files
authored
Update quantization.md
Update quantization.md
1 parent 571841e commit df587a3

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

docs/quantization.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ While quantization can potentially degrade the model's performance, the methods
1515
## Supported Quantization Schemes
1616
### Weight Quantization
1717
| compression | bitwidth| group size | dynamic activation quantization | Eager | AOTI | ExecuTorch |
18-
|--|--|--|--|--|--|--|--|
18+
|--|--|--|--|--|--|--|
1919
| linear (asymmetric) | [8, 4]* | [32, 64, 128, 256]** | ||| 🚧 |
2020
| linear with GPTQ*** (asymmetric) | |[32, 64, 128, 256]** | ||||
2121
| linear with HQQ*** (asymmetric) | |[32, 64, 128, 256]** | ||||
@@ -28,7 +28,7 @@ quantizing the embeddings to further reduce the model size for
2828
on-device usecases.
2929

3030
| compression | weight quantization (bitwidth)| weight quantization (group size) | dynamic activation quantization | Eager | AOTI | ExecuTorch |
31-
|--|--|--|--|--|--|--|--|
31+
|--|--|--|--|--|--|--|
3232
| embedding (symmetric) | [8, 4]* | [32, 64, 128, 256]+ | ||||
3333

3434

0 commit comments

Comments
 (0)