Skip to content

Commit 20ca856

Browse files
committed
llama : print devices used on model load
1 parent 5f4e30d commit 20ca856

File tree

2 files changed

+6
-2
lines changed

2 files changed

+6
-2
lines changed

ggml/src/ggml-backend.cpp

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -555,7 +555,6 @@ struct ggml_backend_registry {
555555
register_backend(ggml_backend_cuda_reg());
556556
#endif
557557
#ifdef GGML_USE_METAL
558-
559558
register_backend(ggml_backend_metal_reg());
560559
#endif
561560
#ifdef GGML_USE_VULKAN
@@ -565,7 +564,7 @@ struct ggml_backend_registry {
565564
register_backend(ggml_backend_blas_reg());
566565
#endif
567566

568-
// TODO: sycl, vulkan, kompute, cann
567+
// TODO: sycl, kompute, cann
569568

570569
register_backend(ggml_backend_cpu_reg());
571570
}

src/llama.cpp

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19100,8 +19100,13 @@ struct llama_model * llama_load_model_from_file(
1910019100

1910119101
case GGML_BACKEND_DEVICE_TYPE_GPU:
1910219102
case GGML_BACKEND_DEVICE_TYPE_GPU_FULL:
19103+
{
19104+
size_t free, total; // NOLINT
19105+
ggml_backend_dev_memory(dev, &free, &total);
19106+
LLAMA_LOG_INFO("%s: using device %s (%s) - %zu MiB free\n", __func__, ggml_backend_dev_name(dev), ggml_backend_dev_description(dev), free/1024/1024);
1910319107
model->devices.push_back(dev);
1910419108
break;
19109+
}
1910519110
}
1910619111
}
1910719112

0 commit comments

Comments
 (0)