File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -3986,7 +3986,7 @@ struct llama_model_loader {
3986
3986
return std::make_pair(cur, ggml_validate_row_data(cur->type, data, n_size));
3987
3987
}));
3988
3988
}
3989
- // TODO LORA allocation of base tensors
3989
+
3990
3990
GGML_ASSERT(buf_mmap || cur->data); // either we have a buffer to allocate the tensor in, or it is already allocated
3991
3991
if (buf_mmap && cur->data == nullptr) {
3992
3992
ggml_backend_tensor_alloc(buf_mmap, cur, data);
@@ -5427,7 +5427,7 @@ static bool llm_load_tensors(
5427
5427
auto ctx_for_layer_split = [&](int i) { return ctx_map.at(model.buft_layer[i].buft_matrix); };
5428
5428
5429
5429
model.layers.resize(n_layer);
5430
- // main players model, ml, ctx_input/output, tn (gets name?)
5430
+
5431
5431
const auto tn = LLM_TN(model.arch);
5432
5432
switch (model.arch) {
5433
5433
case LLM_ARCH_LLAMA:
@@ -6701,7 +6701,7 @@ static bool llm_load_tensors(
6701
6701
#endif
6702
6702
}
6703
6703
}
6704
- #ifdef GGML_USE_METAL // LORA Use metal on base tensors
6704
+ #ifdef GGML_USE_METAL
6705
6705
else if (ml.use_mmap && use_mmap_buffer && buft == ggml_backend_metal_buffer_type()) {
6706
6706
for (uint32_t idx = 0; idx < ml.files.size(); idx++) {
6707
6707
const size_t max_size = ggml_get_max_tensor_size(ctx);
You can’t perform that action at this time.
0 commit comments