Skip to content

Commit 226251e

Browse files
authored
embeddings : fix batch sizes (#13076)
ggml-ci
1 parent 87616f0 commit 226251e

File tree

1 file changed

+7
-1
lines changed

1 file changed

+7
-1
lines changed

examples/embedding/embedding.cpp

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -89,6 +89,13 @@ int main(int argc, char ** argv) {
8989
common_init();
9090

9191
params.embedding = true;
92+
93+
// utilize the full context
94+
if (params.n_batch < params.n_ctx) {
95+
LOG_WRN("%s: setting batch size to %d\n", __func__, params.n_ctx);
96+
params.n_batch = params.n_ctx;
97+
}
98+
9299
// For non-causal models, batch size must be equal to ubatch size
93100
params.n_ubatch = params.n_batch;
94101

@@ -134,7 +141,6 @@ int main(int argc, char ** argv) {
134141

135142
// max batch size
136143
const uint64_t n_batch = params.n_batch;
137-
GGML_ASSERT(params.n_batch >= params.n_ctx);
138144

139145
// tokenize the prompts and trim
140146
std::vector<std::vector<int32_t>> inputs;

0 commit comments

Comments
 (0)