Skip to content

Commit 1342c71

Browse files
mcremon-metafacebook-github-bot
authored andcommitted
Allow int8 type in quantized_linear and quantized_fully_connected (#5900)
Summary: As titled. Reviewed By: zonglinpeng Differential Revision: D63659950
1 parent 12b9819 commit 1342c71

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

backends/cadence/aot/ops_registrations.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ def quantized_linear_meta(
111111
weight_size = list(weight.size())
112112
assert len(weight_size) == 2
113113
out_size[-1] = weight_size[0]
114-
return src.new_empty(out_size, dtype=torch.uint8)
114+
return src.new_empty(out_size, dtype=src.dtype)
115115

116116

117117
@register_fake("cadence::quantized_conv")

0 commit comments

Comments
 (0)