Skip to content

Commit 3720c73

Browse files
committed
Update llama.cpp
1 parent 3bca770 commit 3720c73

File tree

2 files changed

+6
-6
lines changed

2 files changed

+6
-6
lines changed

llama_cpp/llama_cpp.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -301,19 +301,19 @@ class llama_model_params(Structure):
301301

302302
# struct llama_context_params {
303303
# uint32_t seed; // RNG seed, -1 for random
304-
# uint32_t n_ctx; // text context
305-
# uint32_t n_batch; // prompt processing batch size
304+
# uint32_t n_ctx; // text context, 0 = from model
305+
# uint32_t n_batch; // prompt processing maximum batch size
306306
# uint32_t n_threads; // number of threads to use for generation
307307
# uint32_t n_threads_batch; // number of threads to use for batch processing
308308

309309
# // ref: https://github.com/ggerganov/llama.cpp/pull/2054
310-
# float rope_freq_base; // RoPE base frequency
311-
# float rope_freq_scale; // RoPE frequency scaling factor
310+
# float rope_freq_base; // RoPE base frequency, 0 = from model
311+
# float rope_freq_scale; // RoPE frequency scaling factor, 0 = from model
312312

313313

314314
# // Keep the booleans together to avoid misalignment during copy-by-value.
315315
# bool mul_mat_q; // if true, use experimental mul_mat_q kernels
316-
# bool f16_kv; // use fp16 for KV cache
316+
# bool f16_kv; // use fp16 for KV cache, fp32 otherwise
317317
# bool logits_all; // the llama_eval() call computes all logits, not just the last one
318318
# bool embedding; // embedding mode only
319319
# };

vendor/llama.cpp

0 commit comments

Comments
 (0)