Skip to content

Commit c0fc0a1

Browse files
committed
Update llama.cpp
1 parent e9bc4c4 commit c0fc0a1

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

llama_cpp/llama_cpp.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -405,7 +405,7 @@ class llama_model_params(Structure):
405405

406406
# // Keep the booleans together to avoid misalignment during copy-by-value.
407407
# bool mul_mat_q; // if true, use experimental mul_mat_q kernels (DEPRECATED - always true)
408-
# bool logits_all; // the llama_eval() call computes all logits, not just the last one
408+
# bool logits_all; // the llama_eval() call computes all logits, not just the last one (DEPRECATED - set llama_batch.logits instead)
409409
# bool embedding; // embedding mode only
410410
# bool offload_kqv; // whether to offload the KQV ops (including the KV cache) to GPU
411411
# };
@@ -430,7 +430,7 @@ class llama_context_params(Structure):
430430
type_v (int): data type for V cache
431431
mul_mat_q (bool): if true, use experimental mul_mat_q kernels (DEPRECATED - always true)
432432
f16_kv (bool): use fp16 for KV cache, fp32 otherwise
433-
logits_all (bool): the llama_eval() call computes all logits, not just the last one
433+
logits_all (bool): the llama_eval() call computes all logits, not just the last one (DEPRECATED - set llama_batch.logits instead)
434434
embedding (bool): embedding mode only"""
435435
_fields_ = [
436436
("seed", c_uint32),

vendor/llama.cpp

0 commit comments

Comments
 (0)