using System.Collections.Generic;
using System.Threading;
namespace LLama.Abstractions
{
///
/// A high level interface for LLama models.
///
public interface ILLamaExecutor
{
///
/// The loaded context for this executor.
///
public LLamaContext Context { get; }
///
/// Asynchronously infers a response from the model.
///
/// Your prompt
/// Any additional parameters
/// A cancellation token.
///
IAsyncEnumerable InferAsync(string text, IInferenceParams? inferenceParams = null, CancellationToken cancellationToken = default);
}
}