@@ -2253,10 +2253,10 @@ std::vector<std::pair<std::string, struct ggml_tensor *>>& llama_internal_get_te
2253
2253
// Returns the size of the state
2254
2254
size_t llama_get_state_size (struct llama_context * ctx) {
2255
2255
const size_t s_bool = sizeof (int32_t );
2256
- // we don't know size of rng until we actually serialize it. so reserve more than enough memory for its serialized state.
2256
+ // we don't know size of rng until we actually serialize it. so reserve more than enough memory for its serialized state.
2257
2257
// for reference, std::mt19937(1337) serializes to 6701 bytes.
2258
- const size_t s_rng_size = sizeof (size_t );
2259
- const size_t s_rng = 64 *1024 ;
2258
+ const size_t s_rng_size = sizeof (size_t );
2259
+ const size_t s_rng = 64 *1024 ;
2260
2260
const size_t s_logits_capacity = sizeof (size_t );
2261
2261
const size_t s_logits_size = sizeof (size_t );
2262
2262
const size_t s_logits = ctx->logits .capacity () * sizeof (float );
@@ -2300,7 +2300,7 @@ size_t llama_copy_state_data(struct llama_context * ctx, uint8_t * dest) {
2300
2300
memcpy (out, &logits_capacity, sizeof (size_t )); out += sizeof (size_t );
2301
2301
memcpy (out, &logits_size, sizeof (size_t )); out += sizeof (size_t );
2302
2302
if (logits_size) {
2303
- memcpy (out, ctx->logits .data (), logits_size * sizeof (float ));
2303
+ memcpy (out, ctx->logits .data (), logits_size * sizeof (float ));
2304
2304
}
2305
2305
out += logits_capacity * sizeof (float );
2306
2306
memcpy (out, &embedding_size, sizeof (size_t )); out += sizeof (size_t );
@@ -2342,13 +2342,13 @@ size_t llama_set_state_data(struct llama_context * ctx, const uint8_t * src) {
2342
2342
LLAMA_ASSERT (ctx->logits .capacity () == logits_capacity);
2343
2343
if (logits_size) {
2344
2344
ctx->logits .resize (logits_size);
2345
- memcpy (ctx->logits .data (), in, logits_size * sizeof (float ));
2345
+ memcpy (ctx->logits .data (), in, logits_size * sizeof (float ));
2346
2346
}
2347
2347
in += logits_capacity * sizeof (float );
2348
2348
memcpy (&embedding_size, in, sizeof (size_t )); in += sizeof (size_t );
2349
2349
LLAMA_ASSERT (ctx->embedding .capacity () == embedding_size);
2350
2350
if (embedding_size) {
2351
- memcpy (ctx->embedding .data (), in, embedding_size * sizeof (float ));
2351
+ memcpy (ctx->embedding .data (), in, embedding_size * sizeof (float ));
2352
2352
in += embedding_size * sizeof (float );
2353
2353
}
2354
2354
memcpy (&kv_size, in, sizeof (size_t )); in += sizeof (size_t );
@@ -2357,7 +2357,7 @@ size_t llama_set_state_data(struct llama_context * ctx, const uint8_t * src) {
2357
2357
LLAMA_ASSERT (ctx->model .kv_self .buf .size == kv_size);
2358
2358
void * k_data = ctx->model .kv_self .k ->data ; // remember data pointers
2359
2359
void * v_data = ctx->model .kv_self .v ->data ; // because their value is stored in buf and overwritten by memcpy
2360
- memcpy (ctx->model .kv_self .buf .addr , in, kv_size);
2360
+ memcpy (ctx->model .kv_self .buf .addr , in, kv_size);
2361
2361
ctx->model .kv_self .k ->data = k_data; // restore correct data pointers
2362
2362
ctx->model .kv_self .v ->data = v_data;
2363
2363
in += kv_size;
0 commit comments