We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 24c15be commit 69ebccfCopy full SHA for 69ebccf
src/llama-context.cpp
@@ -1058,7 +1058,10 @@ int llama_context::decode(llama_batch & inp_batch) {
1058
1059
if (!res) {
1060
// the last ubatch failed or was aborted -> remove all positions of that ubatch from the KV cache
1061
- llama_pos pos_min[LLAMA_MAX_PARALLEL_SEQUENCES] = { std::numeric_limits<llama_pos>::max() };
+ llama_pos pos_min[LLAMA_MAX_PARALLEL_SEQUENCES];
1062
+ for (int s = 0; s < LLAMA_MAX_PARALLEL_SEQUENCES; ++s) {
1063
+ pos_min[s] = std::numeric_limits<llama_pos>::max();
1064
+ }
1065
1066
for (uint32_t i = 0; i < ubatch.n_tokens; ++i) {
1067
const auto & seq_id = ubatch.seq_id[i][0];
0 commit comments