Skip to content

Commit 93daad7

Browse files
committed
Prefer west const.
1 parent e46a8b5 commit 93daad7

File tree

3 files changed

+28
-28
lines changed

3 files changed

+28
-28
lines changed

examples/beam_search/beam_search.cpp

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@ struct ostream_beam_view {
3232
llama_context * ctx;
3333
llama_beam_view beam_view;
3434
};
35-
std::ostream& operator<<(std::ostream& os, ostream_beam_view const & obv) {
35+
std::ostream& operator<<(std::ostream& os, const ostream_beam_view & obv) {
3636
os << "p(" << obv.beam_view.p << ") eos(" << std::boolalpha << obv.beam_view.eos << ") tokens(";
3737
for (size_t i = 0 ; i < obv.beam_view.n_tokens ; ++i) {
3838
os << llama_token_to_str(obv.ctx, obv.beam_view.tokens[i]);
@@ -46,7 +46,7 @@ struct beam_search_callback_data {
4646
std::vector<llama_token> response;
4747
};
4848

49-
bool is_at_eos(beam_search_callback_data const & callback_data, llama_token const * tokens, size_t const n_tokens) {
49+
bool is_at_eos(const beam_search_callback_data & callback_data, const llama_token * tokens, const size_t n_tokens) {
5050
return n_tokens && tokens[n_tokens-1] == llama_token_eos(callback_data.ctx);
5151
}
5252

@@ -66,10 +66,10 @@ void beam_search_callback(void * callback_data_ptr, llama_beams_state beams_stat
6666
}
6767
}
6868
printf(","); // Show progress
69-
if (size_t const n = beams_state.common_prefix_length) {
69+
if (const size_t n = beams_state.common_prefix_length) {
7070
callback_data.response.resize(callback_data.response.size() + n);
7171
assert(0u < beams_state.n_beams);
72-
llama_token const * tokens = beams_state.beam_views[0].tokens;
72+
const llama_token * tokens = beams_state.beam_views[0].tokens;
7373
std::copy(tokens, tokens + n, callback_data.response.end() - n);
7474
printf("%lu", n);
7575
}

examples/server/server.cpp

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1209,7 +1209,7 @@ static void log_server_request(const Request &req, const Response &res)
12091209
});
12101210
}
12111211

1212-
bool is_at_eos(llama_server_context & server_context, llama_token const * tokens, size_t const n_tokens) {
1212+
bool is_at_eos(llama_server_context & server_context, const llama_token * tokens, const size_t n_tokens) {
12131213
return n_tokens && tokens[n_tokens-1] == llama_token_eos(server_context.ctx);
12141214
}
12151215

@@ -1229,11 +1229,11 @@ void beam_search_callback(void * callback_data, llama_beams_state beams_state) {
12291229
}
12301230
}
12311231
printf(","); // Show progress
1232-
if (size_t const n = beams_state.common_prefix_length) {
1232+
if (const size_t n = beams_state.common_prefix_length) {
12331233
llama.generated_token_probs.resize(llama.generated_token_probs.size() + n);
12341234
assert(0u < beams_state.n_beams);
1235-
llama_token const * tokens = beams_state.beam_views[0].tokens;
1236-
auto const map = [](llama_token tok) { return completion_token_output{{},tok}; };
1235+
const llama_token * tokens = beams_state.beam_views[0].tokens;
1236+
const auto map = [](llama_token tok) { return completion_token_output{{},tok}; };
12371237
std::transform(tokens, tokens + n, llama.generated_token_probs.end() - n, map);
12381238
printf("%lu", n);
12391239
}
@@ -1255,12 +1255,12 @@ struct token_translator {
12551255
void append_to_generated_text_from_generated_token_probs(llama_server_context & llama) {
12561256
auto & gtps = llama.generated_token_probs;
12571257
auto translator = token_translator{llama.ctx};
1258-
auto add_strlen = [=](size_t sum, completion_token_output const & cto) { return sum + translator(cto).size(); };
1259-
size_t const len = std::accumulate(gtps.begin(), gtps.end(), size_t(0), add_strlen);
1258+
auto add_strlen = [=](size_t sum, const completion_token_output & cto) { return sum + translator(cto).size(); };
1259+
const size_t len = std::accumulate(gtps.begin(), gtps.end(), size_t(0), add_strlen);
12601260
if (llama.generated_text.capacity() < llama.generated_text.size() + len) {
12611261
llama.generated_text.reserve(llama.generated_text.size() + len);
12621262
}
1263-
for (completion_token_output const & cto : gtps) {
1263+
for (const completion_token_output & cto : gtps) {
12641264
llama.generated_text += translator(cto);
12651265
}
12661266
}

llama.cpp

Lines changed: 17 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -4335,11 +4335,11 @@ struct llama_beam {
43354335
float p; // Cumulative beam probability (renormalized relative to all beams)
43364336
bool eos; // Initialize end-of-sentence to false. Callback sets this to true.
43374337
// Sort beams by probability. In case of ties, prefer beams at eos.
4338-
bool operator<(llama_beam const & rhs) const {
4338+
bool operator<(const llama_beam & rhs) const {
43394339
return std::make_tuple(p, eos) < std::make_tuple(rhs.p, rhs.eos);
43404340
}
43414341
// Shift off first n tokens and discard them.
4342-
void shift_tokens(size_t const n) {
4342+
void shift_tokens(const size_t n) {
43434343
if (n) {
43444344
std::copy(tokens.begin() + n, tokens.end(), tokens.begin());
43454345
tokens.resize(tokens.size() - n);
@@ -4350,10 +4350,10 @@ struct llama_beam {
43504350

43514351
// A struct for calculating logit-related info.
43524352
struct logit_info {
4353-
float const * const logits;
4354-
int const n_vocab;
4355-
float const max_l;
4356-
float const normalizer;
4353+
const float * const logits;
4354+
const int n_vocab;
4355+
const float max_l;
4356+
const float normalizer;
43574357
struct sum_exp {
43584358
float max_l;
43594359
float operator()(float sum, float l) const { return sum + std::exp(l - max_l); }
@@ -4364,19 +4364,19 @@ struct logit_info {
43644364
, max_l(*std::max_element(logits, logits + n_vocab))
43654365
, normalizer(1.0f / std::accumulate(logits, logits + n_vocab, 0.0f, sum_exp{max_l}))
43664366
{ }
4367-
llama_token_data get_token_data(llama_token const token_id) const {
4367+
llama_token_data get_token_data(const llama_token token_id) const {
43684368
constexpr auto p = std::numeric_limits<float>::quiet_NaN(); // never used
43694369
return {token_id, logits[token_id], p};
43704370
}
43714371
// Return top k token_data by logit.
43724372
std::vector<llama_token_data> top_k(size_t k) {
43734373
std::vector<llama_token_data> min_heap; // min-heap by logit
4374-
llama_token const k_min = std::min(static_cast<llama_token>(k), n_vocab);
4374+
const llama_token k_min = std::min(static_cast<llama_token>(k), n_vocab);
43754375
min_heap.reserve(k_min);
43764376
for (llama_token token_id = 0 ; token_id < k_min ; ++token_id) {
43774377
min_heap.push_back(get_token_data(token_id));
43784378
}
4379-
auto comp = [](llama_token_data const & a, llama_token_data const & b) { return a.logit > b.logit; };
4379+
auto comp = [](const llama_token_data & a, const llama_token_data & b) { return a.logit > b.logit; };
43804380
std::make_heap(min_heap.begin(), min_heap.end(), comp);
43814381
for (llama_token token_id = k_min ; token_id < n_vocab ; ++token_id) {
43824382
if (min_heap.front().logit < logits[token_id]) {
@@ -4420,7 +4420,7 @@ struct beam_search {
44204420
}
44214421

44224422
// Collapse beams to a single beam given by index.
4423-
void collapse_beams(size_t const beam_idx) {
4423+
void collapse_beams(const size_t beam_idx) {
44244424
if (0u < beam_idx) {
44254425
std::swap(beams[0], beams[beam_idx]);
44264426
}
@@ -4434,7 +4434,7 @@ struct beam_search {
44344434
// least element to the back(), replace it with the new, then push it into the heap.
44354435
void fill_next_beams_by_top_probabilities(llama_beam & beam) {
44364436
// Min-heaps use a greater-than comparator.
4437-
auto const comp = [](llama_beam const & a, llama_beam const & b) { return a.p > b.p; };
4437+
const auto comp = [](const llama_beam & a, const llama_beam & b) { return a.p > b.p; };
44384438
if (beam.eos) {
44394439
// beam is at end-of-sentence, so just copy it to next_beams if its probability is high enough.
44404440
if (next_beams.size() < n_beams) {
@@ -4473,7 +4473,7 @@ struct beam_search {
44734473
}
44744474
}
44754475
for (; i < n_beams ; ++i) {
4476-
float const next_p = beam.p * logit_info.probability_from_logit(next_tokens[i].logit);
4476+
const float next_p = beam.p * logit_info.probability_from_logit(next_tokens[i].logit);
44774477
if (next_beams.front().p < next_p) {
44784478
std::pop_heap(next_beams.begin(), next_beams.end(), comp);
44794479
next_beams.back() = beam;
@@ -4503,7 +4503,7 @@ struct beam_search {
45034503

45044504
// Construct beams_state to send back to caller via the callback function.
45054505
// Side effect: set common_prefix_length = find_common_prefix_length();
4506-
llama_beams_state get_beams_state(bool const last_call) {
4506+
llama_beams_state get_beams_state(const bool last_call) {
45074507
for (size_t i = 0 ; i < beams.size() ; ++i) {
45084508
beam_views[i] = beams[i].view();
45094509
}
@@ -4516,9 +4516,9 @@ struct beam_search {
45164516
// * any of the beams have not yet reached end-of-sentence, AND
45174517
// * the highest probability beam(s) (plural in case of ties) are not at end-of-sentence
45184518
// (since all other beam probabilities can only decrease)
4519-
void loop(llama_beam_search_callback_fn_t const callback, void * const callback_data) {
4519+
void loop(const llama_beam_search_callback_fn_t callback, void * const callback_data) {
45204520
beams.push_back({{}, 1.0f, false}); // Start with one empty beam w/ probability = 1.0 and !eos.
4521-
auto const not_eos = [](llama_beam const & beam) { return !beam.eos; };
4521+
const auto not_eos = [](const llama_beam & beam) { return !beam.eos; };
45224522
for (int i = 0 ; i < n_predict && std::any_of(beams.begin(),beams.end(),not_eos) &&
45234523
!beams[top_beam_index()].eos ; ++i) {
45244524
callback(callback_data, get_beams_state(false)); // Sets common_prefix_length
@@ -4544,8 +4544,8 @@ struct beam_search {
45444544
// As beams grow, the cumulative probabilities decrease.
45454545
// Renormalize them to avoid floating point underflow.
45464546
static void renormalize_beam_probabilities(std::vector<llama_beam> & beams) {
4547-
auto const sum_p = [](float sum, llama_beam & beam) { return sum + beam.p; };
4548-
float const inv_sum = 1.0f / std::accumulate(beams.begin(), beams.end(), 0.0f, sum_p);
4547+
const auto sum_p = [](float sum, llama_beam & beam) { return sum + beam.p; };
4548+
const float inv_sum = 1.0f / std::accumulate(beams.begin(), beams.end(), 0.0f, sum_p);
45494549
std::for_each(beams.begin(), beams.end(), [=](llama_beam & beam) { beam.p *= inv_sum; });
45504550
}
45514551

0 commit comments

Comments
 (0)