using LLama.Abstractions;
using System;
using System.Collections.Generic;
using LLama.Native;
using LLama.Sampling;
namespace LLama.Common
{
///
/// The parameters used for inference.
///
public record InferenceParams
: IInferenceParams
{
///
/// number of tokens to keep from initial prompt
///
public int TokensKeep { get; set; } = 0;
///
/// how many new tokens to predict (n_predict), set to -1 to inifinitely generate response
/// until it complete.
///
public int MaxTokens { get; set; } = -1;
///
/// logit bias for specific tokens
///
public Dictionary? LogitBias { get; set; } = null;
///
/// Sequences where the model will stop generating further tokens.
///
public IReadOnlyList AntiPrompts { get; set; } = Array.Empty();
///
public int TopK { get; set; } = 40;
///
public float TopP { get; set; } = 0.95f;
///
public float MinP { get; set; } = 0.05f;
///
public float TfsZ { get; set; } = 1.0f;
///
public float TypicalP { get; set; } = 1.0f;
///
public float Temperature { get; set; } = 0.8f;
///
public float RepeatPenalty { get; set; } = 1.1f;
///
public int RepeatLastTokensCount { get; set; } = 64;
///
public float FrequencyPenalty { get; set; } = .0f;
///
public float PresencePenalty { get; set; } = .0f;
///
public MirostatType Mirostat { get; set; } = MirostatType.Disable;
///
public float MirostatTau { get; set; } = 5.0f;
///
public float MirostatEta { get; set; } = 0.1f;
///
public bool PenalizeNL { get; set; } = true;
///
public SafeLLamaGrammarHandle? Grammar { get; set; }
///
public ISamplingPipeline? SamplingPipeline { get; set; }
}
///
/// Type of "mirostat" sampling to use.
/// https://github.com/basusourya/mirostat
///
public enum MirostatType
{
///
/// Disable Mirostat sampling
///
Disable = 0,
///
/// Original mirostat algorithm
///
Mirostat = 1,
///
/// Mirostat 2.0 algorithm
///
Mirostat2 = 2
}
}