Skip to content

Commit e25c9c1

Browse files
committed
llama : print devices used on model load
1 parent a9b39f4 commit e25c9c1

File tree

1 file changed

+8
-3
lines changed

1 file changed

+8
-3
lines changed

src/llama.cpp

Lines changed: 8 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -19115,10 +19115,15 @@ struct llama_model * llama_load_model_from_file(
1911519115
// TODO: rework API to give user more control over device selection
1911619116
for (size_t i = 0; i < ggml_backend_dev_count(); ++i) {
1911719117
ggml_backend_dev_t dev = ggml_backend_dev_get(i);
19118-
// skip the CPU backend since it is handled separately
19119-
if (ggml_backend_dev_type(dev) != GGML_BACKEND_DEVICE_TYPE_CPU_FULL) {
19120-
model->devices.push_back(dev);
19118+
if (ggml_backend_dev_type(dev) == GGML_BACKEND_DEVICE_TYPE_CPU_FULL) {
19119+
// skip the CPU backend since it is handled separately
19120+
continue;
1912119121
}
19122+
19123+
size_t free, total; // NOLINT
19124+
ggml_backend_dev_memory(dev, &free, &total);
19125+
LLAMA_LOG_INFO("%s: using device %s (%s) - %zu MiB free\n", __func__, ggml_backend_dev_name(dev), ggml_backend_dev_description(dev), free/1024/1024);
19126+
model->devices.push_back(dev);
1912219127
}
1912319128

1912419129
int status = llama_model_load(path_model, *model, params);

0 commit comments

Comments
 (0)