using LLama.Abstractions; using System; using System.Collections.Generic; using LLama.Native; using LLama.Sampling; namespace LLama.Common { /// /// The parameters used for inference. /// public record InferenceParams : IInferenceParams { /// /// number of tokens to keep from initial prompt /// public int TokensKeep { get; set; } = 0; /// /// how many new tokens to predict (n_predict), set to -1 to inifinitely generate response /// until it complete. /// public int MaxTokens { get; set; } = -1; /// /// logit bias for specific tokens /// public Dictionary? LogitBias { get; set; } = null; /// /// Sequences where the model will stop generating further tokens. /// public IReadOnlyList AntiPrompts { get; set; } = Array.Empty(); /// public int TopK { get; set; } = 40; /// public float TopP { get; set; } = 0.95f; /// public float MinP { get; set; } = 0.05f; /// public float TfsZ { get; set; } = 1.0f; /// public float TypicalP { get; set; } = 1.0f; /// public float Temperature { get; set; } = 0.8f; /// public float RepeatPenalty { get; set; } = 1.1f; /// public int RepeatLastTokensCount { get; set; } = 64; /// public float FrequencyPenalty { get; set; } = .0f; /// public float PresencePenalty { get; set; } = .0f; /// public MirostatType Mirostat { get; set; } = MirostatType.Disable; /// public float MirostatTau { get; set; } = 5.0f; /// public float MirostatEta { get; set; } = 0.1f; /// public bool PenalizeNL { get; set; } = true; /// public SafeLLamaGrammarHandle? Grammar { get; set; } /// public ISamplingPipeline? SamplingPipeline { get; set; } } /// /// Type of "mirostat" sampling to use. /// https://github.com/basusourya/mirostat /// public enum MirostatType { /// /// Disable Mirostat sampling /// Disable = 0, /// /// Original mirostat algorithm /// Mirostat = 1, /// /// Mirostat 2.0 algorithm /// Mirostat2 = 2 } }