using System.Collections.Generic; using System.Threading; namespace LLama.Abstractions { /// /// A high level interface for LLama models. /// public interface ILLamaExecutor { /// /// The loaded context for this executor. /// public LLamaContext Context { get; } /// /// Asynchronously infers a response from the model. /// /// Your prompt /// Any additional parameters /// A cancellation token. /// IAsyncEnumerable InferAsync(string text, IInferenceParams? inferenceParams = null, CancellationToken cancellationToken = default); } }