update llama.cpp

This commit is contained in:
Michael Yang
2023-08-01 16:22:03 -07:00
parent da52f5bfdd
commit 7a1c3e62dc
18 changed files with 2603 additions and 493 deletions

View File

@@ -1,5 +1,5 @@
/**
* llama.cpp - git d91f3f0c55663719ea03b76311e8c36ed55eb0e2
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
*
* MIT License
*
@@ -134,6 +134,7 @@ extern "C" {
// Keep the booleans together to avoid misalignment during copy-by-value.
bool low_vram; // if true, reduce VRAM usage at the cost of performance
bool mul_mat_q; // if true, use experimental mul_mat_q kernels
bool f16_kv; // use fp16 for KV cache
bool logits_all; // the llama_eval() call computes all logits, not just the last one
bool vocab_only; // only load the vocabulary, no weights