Skip to content

Commit 05c5a8b

Browse files
authored
test: Limiting multi-gpu tests to use Ray as distributed_executor_backend (#47)
1 parent db3d794 commit 05c5a8b

File tree

2 files changed

+11
-3
lines changed

2 files changed

+11
-3
lines changed

ci/L0_multi_gpu/multi_lora/test.sh

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -62,7 +62,8 @@ model_json=$(cat <<EOF
6262
"enforce_eager": "true",
6363
"enable_lora": "true",
6464
"max_lora_rank": 32,
65-
"lora_extra_vocab_size": 256
65+
"lora_extra_vocab_size": 256,
66+
"distributed_executor_backend":"ray"
6667
}
6768
EOF
6869
)
@@ -120,7 +121,8 @@ model_json=$(cat <<EOF
120121
"block_size": 16,
121122
"enforce_eager": "true",
122123
"enable_lora": "false",
123-
"lora_extra_vocab_size": 256
124+
"lora_extra_vocab_size": 256,
125+
"distributed_executor_backend":"ray"
124126
}
125127
EOF
126128
)

ci/L0_multi_gpu/vllm_backend/test.sh

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -63,6 +63,7 @@ function run_multi_gpu_test() {
6363
export KIND="${1}"
6464
export TENSOR_PARALLELISM="${2}"
6565
export INSTANCE_COUNT="${3}"
66+
export DISTRIBUTED_EXECUTOR_BACKEND="${4}"
6667

6768
# Setup a clean model repository
6869
export TEST_MODEL="vllm_opt_${KIND}_tp${TENSOR_PARALLELISM}_count${INSTANCE_COUNT}"
@@ -73,6 +74,10 @@ function run_multi_gpu_test() {
7374
cp -r "${SAMPLE_MODELS_REPO}/vllm_model" "models/${TEST_MODEL}"
7475
sed -i "s/KIND_MODEL/${KIND}/" "${TEST_MODEL_TRITON_CONFIG}"
7576
sed -i "3s/^/ \"tensor_parallel_size\": ${TENSOR_PARALLELISM},\n/" "${TEST_MODEL_VLLM_CONFIG}"
77+
if [ $TENSOR_PARALLELISM -ne "1" ]; then
78+
jq --arg backend $DISTRIBUTED_EXECUTOR_BACKEND '. += {"distributed_executor_backend":$backend}' "${TEST_MODEL_VLLM_CONFIG}" > "temp.json"
79+
mv temp.json "${TEST_MODEL_VLLM_CONFIG}"
80+
fi
7681
# Assert the correct kind is set in case the template config changes in the future
7782
validate_file_contains "${KIND}" "${TEST_MODEL_TRITON_CONFIG}"
7883

@@ -119,10 +124,11 @@ RET=0
119124
KINDS="KIND_MODEL KIND_GPU"
120125
TPS="1 2"
121126
INSTANCE_COUNTS="1 2"
127+
DISTRIBUTED_EXECUTOR_BACKEND="ray"
122128
for kind in ${KINDS}; do
123129
for tp in ${TPS}; do
124130
for count in ${INSTANCE_COUNTS}; do
125-
run_multi_gpu_test "${kind}" "${tp}" "${count}"
131+
run_multi_gpu_test "${kind}" "${tp}" "${count}" "${DISTRIBUTED_EXECUTOR_BACKEND}"
126132
done
127133
done
128134
done

0 commit comments

Comments
 (0)