Skip to content

Commit 87f4102

Browse files
committed
llama : revert n_threads_batch logic
ggml-ci
1 parent e9b7a5c commit 87f4102

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

llama.cpp

+1-1
Original file line numberDiff line numberDiff line change
@@ -5433,7 +5433,7 @@ static int llama_decode_internal(
54335433

54345434
GGML_ASSERT(n_tokens <= n_batch);
54355435

5436-
int n_threads = n_tokens < 32 ? cparams.n_threads : cparams.n_threads_batch;
5436+
int n_threads = n_tokens == 1 ? cparams.n_threads : cparams.n_threads_batch;
54375437
GGML_ASSERT((!batch.token && batch.embd) || (batch.token && !batch.embd)); // NOLINT
54385438

54395439
const int64_t t_start_us = ggml_time_us();

0 commit comments

Comments
 (0)