Skip to content

Commit 0ce221e

Browse files
mcremon-metafacebook-github-bot
authored andcommitted
Allow int8 type in quantized_linear and quantized_fully_connected (#5900)
Summary: Pull Request resolved: #5900 As titled. Reviewed By: zonglinpeng Differential Revision: D63659950 fbshipit-source-id: 1c28a797ef2a02ba86a2d860bf6ad44194f38ca0
1 parent f5f6969 commit 0ce221e

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

backends/cadence/aot/ops_registrations.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ def quantized_linear_meta(
111111
weight_size = list(weight.size())
112112
assert len(weight_size) == 2
113113
out_size[-1] = weight_size[0]
114-
return src.new_empty(out_size, dtype=torch.uint8)
114+
return src.new_empty(out_size, dtype=src.dtype)
115115

116116

117117
@register_fake("cadence::quantized_conv")

0 commit comments

Comments
 (0)