mirror of
https://github.com/likelovewant/ollama-for-amd.git
synced 2025-12-23 23:18:26 +00:00
update llama.cpp
This commit is contained in:
@@ -1,5 +1,5 @@
|
||||
/**
|
||||
* llama.cpp - git d91f3f0c55663719ea03b76311e8c36ed55eb0e2
|
||||
* llama.cpp - git c574bddb368424b5996cbee2ec45ec050967d404
|
||||
*
|
||||
* MIT License
|
||||
*
|
||||
@@ -134,6 +134,7 @@ extern "C" {
|
||||
|
||||
// Keep the booleans together to avoid misalignment during copy-by-value.
|
||||
bool low_vram; // if true, reduce VRAM usage at the cost of performance
|
||||
bool mul_mat_q; // if true, use experimental mul_mat_q kernels
|
||||
bool f16_kv; // use fp16 for KV cache
|
||||
bool logits_all; // the llama_eval() call computes all logits, not just the last one
|
||||
bool vocab_only; // only load the vocabulary, no weights
|
||||
|
||||
Reference in New Issue
Block a user