File tree Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -301,19 +301,19 @@ class llama_model_params(Structure):
301
301
302
302
# struct llama_context_params {
303
303
# uint32_t seed; // RNG seed, -1 for random
304
- # uint32_t n_ctx; // text context
305
- # uint32_t n_batch; // prompt processing batch size
304
+ # uint32_t n_ctx; // text context, 0 = from model
305
+ # uint32_t n_batch; // prompt processing maximum batch size
306
306
# uint32_t n_threads; // number of threads to use for generation
307
307
# uint32_t n_threads_batch; // number of threads to use for batch processing
308
308
309
309
# // ref: https://github.com/ggerganov/llama.cpp/pull/2054
310
- # float rope_freq_base; // RoPE base frequency
311
- # float rope_freq_scale; // RoPE frequency scaling factor
310
+ # float rope_freq_base; // RoPE base frequency, 0 = from model
311
+ # float rope_freq_scale; // RoPE frequency scaling factor, 0 = from model
312
312
313
313
314
314
# // Keep the booleans together to avoid misalignment during copy-by-value.
315
315
# bool mul_mat_q; // if true, use experimental mul_mat_q kernels
316
- # bool f16_kv; // use fp16 for KV cache
316
+ # bool f16_kv; // use fp16 for KV cache, fp32 otherwise
317
317
# bool logits_all; // the llama_eval() call computes all logits, not just the last one
318
318
# bool embedding; // embedding mode only
319
319
# };
You can’t perform that action at this time.
0 commit comments