We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 6619832 commit 95a9f8bCopy full SHA for 95a9f8b
src/llama-kv-cache.cpp
@@ -86,7 +86,7 @@ llama_kv_cache_unified::llama_kv_cache_unified(
86
87
ggml_backend_buffer_type_t buft = ggml_backend_cpu_buffer_type();
88
89
- if (!offload) {
+ if (offload) {
90
auto * dev = model.dev_layer(i);
91
buft = ggml_backend_dev_buffer_type(dev);
92
@@ -1445,7 +1445,7 @@ llama_kv_cache_recurrent::llama_kv_cache_recurrent(
1445
1446
1447
1448
1449
1450
1451
0 commit comments