Namespace: LLama.Native
public struct LLamaContextParams
Inheritance Object → ValueType → LLamaContextParams
text context
public int n_ctx;
number of layers to store in VRAM
public int n_gpu_layers;
RNG seed, -1 for random
public int seed;
use fp16 for KV cache
public bool f16_kv;
the llama_eval() call computes all logits, not just the last one
public bool logits_all;
only load the vocabulary, no weights
public bool vocab_only;
use mmap if possible
public bool use_mmap;
force system to keep model in RAM
public bool use_mlock;
embedding mode only
public bool embedding;
called with a progress value between 0 and 1, pass NULL to disable
public IntPtr progress_callback;
context pointer passed to the progress callback
public IntPtr progress_callback_user_data;