We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent eb623f2 commit 17809cfCopy full SHA for 17809cf
src/llama-kv-cache.cpp
@@ -86,7 +86,7 @@ llama_kv_cache_unified::llama_kv_cache_unified(
86
87
ggml_backend_buffer_type_t buft = ggml_backend_cpu_buffer_type();
88
89
- if (!offload) {
+ if (offload) {
90
auto * dev = model.dev_layer(i);
91
buft = ggml_backend_dev_buffer_type(dev);
92
@@ -1445,7 +1445,7 @@ llama_kv_cache_recurrent::llama_kv_cache_recurrent(
1445
1446
1447
1448
1449
1450
1451
0 commit comments