Skip to content

Commit 3291cb7

Browse files
committed
Update on "Add quantized op support to llama runner"
Summary: Test Plan: Reviewers: Subscribers: Tasks: Tags: Differential Revision: [D56197863](https://our.internmc.facebook.com/intern/diff/D56197863) [ghstack-poisoned]
2 parents 38cfb8d + 56f02f1 commit 3291cb7

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

.ci/scripts/test_llama.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -141,7 +141,7 @@ if [[ "${CUSTOM}" == "ON" ]]; then
141141
EXPORT_ARGS="${EXPORT_ARGS} --use_sdpa_with_kv_cache"
142142
fi
143143
if [[ "${QE}" == "ON" ]]; then
144-
EXPORT_ARGS="${EXPORT_ARGS} --embedding-quantize '8,1024'"
144+
EXPORT_ARGS="${EXPORT_ARGS} --embedding-quantize 8,1024"
145145
fi
146146
# Add dynamically linked library location
147147
$PYTHON_EXECUTABLE -m examples.models.llama2.export_llama ${EXPORT_ARGS}

0 commit comments

Comments
 (0)