diff --git a/0.5/404.html b/0.5/404.html new file mode 100755 index 00000000..8a544c66 --- /dev/null +++ b/0.5/404.html @@ -0,0 +1,2009 @@ + + + + + + + + + + + + + + + + + + LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ +

404 - Not found

+ +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Architecture/index.html b/0.5/Architecture/index.html new file mode 100755 index 00000000..e0ae20ab --- /dev/null +++ b/0.5/Architecture/index.html @@ -0,0 +1,2112 @@ + + + + + + + + + + + + + + + + + + + + + + Architecture - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Architecture

+

Architecture of main functions

+

The figure below shows the core framework structure, which is separated to four levels.

+
    +
  • LLamaContext: The holder of a model which directly interact with native library and provide some basic APIs such as tokenization and embedding. Currently it includes three classes: LLamaContext, LLamaEmbedder and LLamaQuantizer.
  • +
  • LLamaExecutors: Executors which define the way to run the LLama model. It provides text-to-text APIs to make it easy to use. Currently we provide three kinds of executors: InteractiveExecutor, InstructuExecutor and StatelessExecutor.
  • +
  • ChatSession: A wrapping for InteractiveExecutor and LLamaContext, which supports interactive tasks and saving/re-loading sessions. It also provides a flexible way to customize the text process by IHistoryTransform, ITextTransform and ITextStreamTransform.
  • +
  • High-level Applications: Some applications that provides higher-level integration. For example, BotSharp provides integration for vector search, Chatbot UI and Web APIs. semantic-kernel provides various APIs for manipulations related with LLM. If you've made an integration, please tell us and add it to the doc!
  • +
+

structure_image

+ +

Since LLamaContext interact with native library, it's not recommended to use the methods of it directly unless you know what you are doing. So does the NativeApi, which is not included in the architecture figure above.

+

ChatSession is recommended to be used when you want to build an application similar to ChatGPT, or the ChatBot, because it works best with InteractiveExecutor. Though other executors are also allowed to passed as a parameter to initialize a ChatSession, it's not encouraged if you are new to LLamaSharp and LLM.

+

High-level applications, such as BotSharp, are supposed to be used when you concentrate on the part not related with LLM. For example, if you want to deploy a chat bot to help you remember your schedules, using BotSharp may be a good choice.

+

Note that the APIs of the high-level applications may not be stable now. Please take it into account when using them.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/ChatSession/basic-usages/index.html b/0.5/ChatSession/basic-usages/index.html new file mode 100755 index 00000000..c71a889b --- /dev/null +++ b/0.5/ChatSession/basic-usages/index.html @@ -0,0 +1,2136 @@ + + + + + + + + + + + + + + + + + + + + + + Basic Usages - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Basic usages of ChatSession

+

ChatSession is a higher-level abstraction than the executors. In the context of a chat application like ChatGPT, a "chat session" refers to an interactive conversation or exchange of messages between the user and the chatbot. It represents a continuous flow of communication where the user enters input or asks questions, and the chatbot responds accordingly. A chat session typically starts when the user initiates a conversation with the chatbot and continues until the interaction comes to a natural end or is explicitly terminated by either the user or the system. During a chat session, the chatbot maintains the context of the conversation, remembers previous messages, and generates appropriate responses based on the user's inputs and the ongoing dialogue.

+

Initialize a session

+

Currently, the only parameter that is accepted is an ILLamaExecutor, because this is the only parameter that we're sure to exist in all the future versions. Since it's the high-level abstraction, we're conservative to the API designs. In the future, there may be more kinds of constructors added.

+
InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath)));
+ChatSession session = new ChatSession(ex);
+
+

Chat with the bot

+

There'll be two kinds of input accepted by the Chat API, which are ChatHistory and String. The API with string is quite similar to that of the executors. Meanwhile, the API with ChatHistory is aimed to provide more flexible usages. For example, you have had a chat with the bot in session A before you open the session B. Now session B has no memory for what you said before. Therefore, you can feed the history of A to B.

+
string prompt = "What is C#?";
+
+foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } })) // the inference params should be changed depending on your statement
+{
+    Console.Write(text);
+}
+
+

Get the history

+

Currently History is a property of ChatSession.

+
foreach(var rec in session.History.Messages)
+{
+    Console.WriteLine($"{rec.AuthorRole}: {rec.Content}");
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/ChatSession/save-load-session/index.html b/0.5/ChatSession/save-load-session/index.html new file mode 100755 index 00000000..fa11b905 --- /dev/null +++ b/0.5/ChatSession/save-load-session/index.html @@ -0,0 +1,2048 @@ + + + + + + + + + + + + + + + + + + + + + + Save/Load Session - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Save/Load Chat Session

+

Generally, the chat session could be switched, which requires the ability of loading and saving session.

+

When building a chat bot app, it's NOT encouraged to initialize many chat sessions and keep them in memory to wait for being switched, because the memory consumption of both CPU and GPU is expensive. It's recommended to save the current session before switching to a new session, and load the file when switching back to the session.

+

The API is also quite simple, the files will be saved into a directory you specified. If the path does not exist, a new directory will be created.

+
string savePath = "<save dir>";
+session.SaveSession(savePath);
+
+session.LoadSession(savePath);
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/ChatSession/transforms/index.html b/0.5/ChatSession/transforms/index.html new file mode 100755 index 00000000..0aac75ad --- /dev/null +++ b/0.5/ChatSession/transforms/index.html @@ -0,0 +1,2328 @@ + + + + + + + + + + + + + + + + + + + + + + Transoforms - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Transforms in Chat Session

+

There's three important elements in ChatSession, which are input, output and history. Besides, there're some conversions between them. Since the process of them under different conditions varies, LLamaSharp hands over this part of the power to the users.

+

Currently, there're three kinds of process that could be customized, as introduced below.

+

Input transform

+

In general, the input of the chat API is a text (without stream), therefore ChatSession processes it in a pipeline. If you want to use your customized transform, you need to define a transform that implements ITextTransform and add it to the pipeline of ChatSession.

+
public interface ITextTransform
+{
+    string Transform(string text);
+}
+
+
public class MyInputTransform1 : ITextTransform
+{
+    public string Transform(string text)
+    {
+        return $"Question: {text}\n";
+    }
+}
+
+public class MyInputTransform2 : ITextTransform
+{
+    public string Transform(string text)
+    {
+        return text + "Answer: ";
+    }
+}
+
+session.AddInputTransform(new MyInputTransform1()).AddInputTransform(new MyInputTransform2());
+
+

Output transform

+

Different from the input, the output of chat API is a text stream. Therefore you need to process it word by word, instead of getting the full text at once.

+

The interface of it has an IEnumerable<string> as input, which is actually a yield sequence.

+
public interface ITextStreamTransform
+{
+    IEnumerable<string> Transform(IEnumerable<string> tokens);
+    IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens);
+}
+
+

When implementing it, you could throw a not-implemented exception in one of them if you only need to use the chat API in synchronously or asynchronously.

+

Different from the input transform pipeline, the output transform only supports one transform.

+
session.WithOutputTransform(new MyOutputTransform());
+
+

Here's an example of how to implement the interface. In this example, the transform detects whether there's some keywords in the response and removes them.

+
/// <summary>
+/// A text output transform that removes the keywords from the response.
+/// </summary>
+public class KeywordTextOutputStreamTransform : ITextStreamTransform
+{
+    HashSet<string> _keywords;
+    int _maxKeywordLength;
+    bool _removeAllMatchedTokens;
+
+    /// <summary>
+    /// 
+    /// </summary>
+    /// <param name="keywords">Keywords that you want to remove from the response.</param>
+    /// <param name="redundancyLength">The extra length when searching for the keyword. For example, if your only keyword is "highlight", 
+    /// maybe the token you get is "\r\nhighligt". In this condition, if redundancyLength=0, the token cannot be successfully matched because the length of "\r\nhighligt" (10)
+    /// has already exceeded the maximum length of the keywords (8). On the contrary, setting redundancyLengyh >= 2 leads to successful match.
+    /// The larger the redundancyLength is, the lower the processing speed. But as an experience, it won't introduce too much performance impact when redundancyLength <= 5 </param>
+    /// <param name="removeAllMatchedTokens">If set to true, when getting a matched keyword, all the related tokens will be removed. Otherwise only the part of keyword will be removed.</param>
+    public KeywordTextOutputStreamTransform(IEnumerable<string> keywords, int redundancyLength = 3, bool removeAllMatchedTokens = false)
+    {
+        _keywords = new(keywords);
+        _maxKeywordLength = keywords.Select(x => x.Length).Max() + redundancyLength;
+        _removeAllMatchedTokens = removeAllMatchedTokens;
+    }
+    /// <inheritdoc />
+    public IEnumerable<string> Transform(IEnumerable<string> tokens)
+    {
+        var window = new Queue<string>();
+
+        foreach (var s in tokens)
+        {
+            window.Enqueue(s);
+            var current = string.Join("", window);
+            if (_keywords.Any(x => current.Contains(x)))
+            {
+                var matchedKeyword = _keywords.First(x => current.Contains(x));
+                int total = window.Count;
+                for (int i = 0; i < total; i++)
+                {
+                    window.Dequeue();
+                }
+                if (!_removeAllMatchedTokens)
+                {
+                    yield return current.Replace(matchedKeyword, "");
+                }
+            }
+            if (current.Length >= _maxKeywordLength)
+            {
+                if (_keywords.Any(x => current.Contains(x)))
+                {
+                    var matchedKeyword = _keywords.First(x => current.Contains(x));
+                    int total = window.Count;
+                    for (int i = 0; i < total; i++)
+                    {
+                        window.Dequeue();
+                    }
+                    if (!_removeAllMatchedTokens)
+                    {
+                        yield return current.Replace(matchedKeyword, "");
+                    }
+                }
+                else
+                {
+                    int total = window.Count;
+                    for (int i = 0; i < total; i++)
+                    {
+                        yield return window.Dequeue();
+                    }
+                }
+            }
+        }
+        int totalCount = window.Count;
+        for (int i = 0; i < totalCount; i++)
+        {
+            yield return window.Dequeue();
+        }
+    }
+    /// <inheritdoc />
+    public async IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens)
+    {
+        throw new NotImplementedException(); // This is implemented in `LLamaTransforms` but we ignore it here.
+    }
+}
+
+

History transform

+

The chat history could be converted to or from a text, which is exactly what the interface of it.

+
public interface IHistoryTransform
+{
+    string HistoryToText(ChatHistory history);
+    ChatHistory TextToHistory(AuthorRole role, string text);
+}
+
+

Similar to the output transform, the history transform is added in the following way:

+
session.WithHistoryTransform(new MyHistoryTransform());
+
+

The implementation is quite flexible, depending on what you want the history message to be like. Here's an example, which is the default history transform in LLamaSharp.

+
/// <summary>
+/// The default history transform.
+/// Uses plain text with the following format:
+/// [Author]: [Message]
+/// </summary>
+public class DefaultHistoryTransform : IHistoryTransform
+{
+    private readonly string defaultUserName = "User";
+    private readonly string defaultAssistantName = "Assistant";
+    private readonly string defaultSystemName = "System";
+    private readonly string defaultUnknownName = "??";
+
+    string _userName;
+    string _assistantName;
+    string _systemName;
+    string _unknownName;
+    bool _isInstructMode;
+    public DefaultHistoryTransform(string? userName = null, string? assistantName = null, 
+        string? systemName = null, string? unknownName = null, bool isInstructMode = false)
+    {
+        _userName = userName ?? defaultUserName;
+        _assistantName = assistantName ?? defaultAssistantName;
+        _systemName = systemName ?? defaultSystemName;
+        _unknownName = unknownName ?? defaultUnknownName;
+        _isInstructMode = isInstructMode;
+    }
+
+    public virtual string HistoryToText(ChatHistory history)
+    {
+        StringBuilder sb = new();
+        foreach (var message in history.Messages)
+        {
+            if (message.AuthorRole == AuthorRole.User)
+            {
+                sb.AppendLine($"{_userName}: {message.Content}");
+            }
+            else if (message.AuthorRole == AuthorRole.System)
+            {
+                sb.AppendLine($"{_systemName}: {message.Content}");
+            }
+            else if (message.AuthorRole == AuthorRole.Unknown)
+            {
+                sb.AppendLine($"{_unknownName}: {message.Content}");
+            }
+            else if (message.AuthorRole == AuthorRole.Assistant)
+            {
+                sb.AppendLine($"{_assistantName}: {message.Content}");
+            }
+        }
+        return sb.ToString();
+    }
+
+    public virtual ChatHistory TextToHistory(AuthorRole role, string text)
+    {
+        ChatHistory history = new ChatHistory();
+        history.AddMessage(role, TrimNamesFromText(text, role));
+        return history;
+    }
+
+    public virtual string TrimNamesFromText(string text, AuthorRole role)
+    {
+        if (role == AuthorRole.User && text.StartsWith($"{_userName}:"))
+        {
+            text = text.Substring($"{_userName}:".Length).TrimStart();
+        }
+        else if (role == AuthorRole.Assistant && text.EndsWith($"{_assistantName}:"))
+        {
+            text = text.Substring(0, text.Length - $"{_assistantName}:".Length).TrimEnd();
+        }
+        if (_isInstructMode && role == AuthorRole.Assistant && text.EndsWith("\n> "))
+        {
+            text = text.Substring(0, text.Length - "\n> ".Length).TrimEnd();
+        }
+        return text;
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/ContributingGuide/index.html b/0.5/ContributingGuide/index.html new file mode 100755 index 00000000..77ce1c21 --- /dev/null +++ b/0.5/ContributingGuide/index.html @@ -0,0 +1,2194 @@ + + + + + + + + + + + + + + + + + + + + + + Contributing Guide - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+ +
+ + + +
+
+ + + + +

LLamaSharp Contributing Guide

+

Hi, welcome to develop LLamaSharp with us together! We are always open for every contributor and any format of contributions! If you want to maintain this library actively together, please contact us to get the write access after some PRs. (Email: AsakusaRinne@gmail.com)

+

In this page, we'd like to introduce how to make contributions here easily. 😊

+

Compile the native library from source

+

Firstly, please clone the llama.cpp repository and following the instructions in llama.cpp readme to configure your local environment.

+

If you want to support cublas in the compilation, please make sure that you've installed the cuda.

+

When building from source, please add -DBUILD_SHARED_LIBS=ON to the cmake instruction. For example, when building with cublas but without openblas, use the following instruction:

+
cmake .. -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS=ON
+
+

After running cmake --build . --config Release, you could find the llama.dll, llama.so or llama.dylib in your build directory. After pasting it to LLamaSharp/LLama/runtimes and renaming it to libllama.dll, libllama.so or libllama.dylib, you can use it as the native library in LLamaSharp.

+

Add a new feature to LLamaSharp

+

After refactoring the framework in v0.4.0, LLamaSharp will try to maintain the backward compatibility. However, in the following cases a breaking change will be required:

+
    +
  1. Due to some break changes in llama.cpp, making a breaking change will help to maintain the good abstraction and friendly user APIs.
  2. +
  3. A very important feature cannot be implemented unless refactoring some parts.
  4. +
  5. After some discussions, an agreement was reached that making the break change is reasonable.
  6. +
+

If a new feature could be added without introducing any break change, please open a PR rather than open an issue first. We will never refuse the PR but help to improve it, unless it's malicious.

+

When adding the feature, please take care of the namespace and the naming convention. For example, if you are adding an integration for WPF, please put the code under namespace LLama.WPF or LLama.Integration.WPF instead of putting it under the root namespace. The naming convention of LLamaSharp follows the pascal naming convention, but in some parts that are invisible to users, you can do whatever you want.

+

Find the problem and fix the BUG

+

If the issue is related to the LLM internal behaviour, such as endless generating the response, the best way to find the problem is to do comparison test between llama.cpp and LLamaSharp.

+

You could use exactly the same prompt, the same model and the same parameters to run the inference in llama.cpp and LLamaSharp respectively to see if it's really a problem caused by the implementation in LLamaSharp.

+

If the experiment showed that it worked well in llama.cpp but didn't in LLamaSharp, a search for the problem could be started. While the reason of the problem could be various, the best way I think is to add log-print in the code of llama.cpp and use it in LLamaSharp after compilation. Thus, when running LLamaSharp, you could see what happened in the native library.

+

After finding out the reason, a painful but happy process comes. When working on the BUG fix, there's only one rule to follow, that is keeping the examples working well. If the modification fixed the BUG but impact on other functions, it would not be a good fix.

+

During the BUG fix process, please don't hesitate to discuss together when you stuck on something.

+

Add integrations

+

All kinds of integration are welcomed here! Currently the following integrations are under work or on our schedule:

+
    +
  1. BotSharp
  2. +
  3. semantic-kernel
  4. +
  5. Unity
  6. +
+

Besides, for some other integrations, like ASP.NET core, SQL, Blazor and so on, we'll appreciate it if you could help with that. If the time is limited for you, providing an example for it also means a lot!

+

Add examples

+

There're mainly two ways to add an example:

+
    +
  1. Add the example to LLama.Examples of the repository.
  2. +
  3. Put the example in another repository and add the link to the readme or docs of LLamaSharp.
  4. +
+

Add documents

+

LLamaSharp uses mkdocs to build the documentation, please follow the tutorial of mkdocs to add or modify documents in LLamaSharp.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/ChatSessionStripRoleName/index.html b/0.5/Examples/ChatSessionStripRoleName/index.html new file mode 100755 index 00000000..f74e0de1 --- /dev/null +++ b/0.5/Examples/ChatSessionStripRoleName/index.html @@ -0,0 +1,2075 @@ + + + + + + + + + + + + + + + + + + + + + + Chat session 1 - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use chat session and strip role names

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class ChatSessionStripRoleName
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/chat-with-bob.txt").Trim();
+        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));
+        ChatSession session = new ChatSession(ex).WithOutputTransform(new LLamaTransforms.KeywordTextOutputStreamTransform(new string[] { "User:", "Bob:" }, redundancyLength: 8));
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The chat session has started. The role names won't be printed.");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        while (true)
+        {
+            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } }))
+            {
+                Console.Write(text);
+            }
+
+            Console.ForegroundColor = ConsoleColor.Green;
+            prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/ChatSessionWithRoleName/index.html b/0.5/Examples/ChatSessionWithRoleName/index.html new file mode 100755 index 00000000..2a6d4ea2 --- /dev/null +++ b/0.5/Examples/ChatSessionWithRoleName/index.html @@ -0,0 +1,2077 @@ + + + + + + + + + + + + + + + + + + + + + + Chat session 2 - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use chat session without removing role names

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class ChatSessionWithRoleName
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/chat-with-bob.txt").Trim();
+        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));
+        ChatSession session = new ChatSession(ex); // The only change is to remove the transform for the output text stream.
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The chat session has started. In this example, the prompt is printed for better visual result.");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        // show the prompt
+        Console.Write(prompt);
+        while (true)
+        {
+            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } }))
+            {
+                Console.Write(text);
+            }
+
+            Console.ForegroundColor = ConsoleColor.Green;
+            prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/GetEmbeddings/index.html b/0.5/Examples/GetEmbeddings/index.html new file mode 100755 index 00000000..aed9b245 --- /dev/null +++ b/0.5/Examples/GetEmbeddings/index.html @@ -0,0 +1,2068 @@ + + + + + + + + + + + + + + + + + + + + + + Get embeddings - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Get embeddings

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class GetEmbeddings
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var embedder = new LLamaEmbedder(new ModelParams(modelPath));
+
+        while (true)
+        {
+            Console.Write("Please input your text: ");
+            Console.ForegroundColor = ConsoleColor.Green;
+            var text = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+
+            Console.WriteLine(string.Join(", ", embedder.GetEmbeddings(text)));
+            Console.WriteLine();
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/InstructModeExecute/index.html b/0.5/Examples/InstructModeExecute/index.html new file mode 100755 index 00000000..e5a2e93a --- /dev/null +++ b/0.5/Examples/InstructModeExecute/index.html @@ -0,0 +1,2077 @@ + + + + + + + + + + + + + + + + + + + + + + Instruct executor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use instruct executor

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class InstructModeExecute
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/dan.txt").Trim();
+
+        InstructExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024)));
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The executor has been enabled. In this example, the LLM will follow your instructions. For example, you can input \"Write a story about a fox who want to " +
+            "make friend with human, no less than 200 words.\"");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        var inferenceParams = new InferenceParams() { Temperature = 0.8f, MaxTokens = 300 };
+
+        while (true)
+        {
+            foreach (var text in ex.Infer(prompt, inferenceParams))
+            {
+                Console.Write(text);
+            }
+            Console.ForegroundColor = ConsoleColor.Green;
+            prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/InteractiveModeExecute/index.html b/0.5/Examples/InteractiveModeExecute/index.html new file mode 100755 index 00000000..e0e83c2e --- /dev/null +++ b/0.5/Examples/InteractiveModeExecute/index.html @@ -0,0 +1,2078 @@ + + + + + + + + + + + + + + + + + + + + + + Interactive executor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use interactive executor

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class InteractiveModeExecute
+{
+    public async static Task Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/chat-with-bob.txt").Trim();
+
+        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to 64 and the context size is 256. (an example for small scale usage)");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        Console.Write(prompt);
+
+        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" }, MaxTokens = 64 };
+
+        while (true)
+        {
+            await foreach (var text in ex.InferAsync(prompt, inferenceParams))
+            {
+                Console.Write(text);
+            }
+            Console.ForegroundColor = ConsoleColor.Green;
+            prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/LoadAndSaveSession/index.html b/0.5/Examples/LoadAndSaveSession/index.html new file mode 100755 index 00000000..993762c0 --- /dev/null +++ b/0.5/Examples/LoadAndSaveSession/index.html @@ -0,0 +1,2103 @@ + + + + + + + + + + + + + + + + + + + + + + Load/Save session - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Load and save chat session

+
using LLama.Common;
+using LLama.OldVersion;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class SaveAndLoadSession
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/chat-with-bob.txt").Trim();
+        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));
+        ChatSession session = new ChatSession(ex); // The only change is to remove the transform for the output text stream.
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The chat session has started. In this example, the prompt is printed for better visual result. Input \"save\" to save and reload the session.");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        // show the prompt
+        Console.Write(prompt);
+        while (true)
+        {
+            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } }))
+            {
+                Console.Write(text);
+            }
+
+            Console.ForegroundColor = ConsoleColor.Green;
+            prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White;
+            if (prompt == "save")
+            {
+                Console.Write("Preparing to save the state, please input the path you want to save it: ");
+                Console.ForegroundColor = ConsoleColor.Green;
+                var statePath = Console.ReadLine();
+                session.SaveSession(statePath);
+                Console.ForegroundColor = ConsoleColor.White;
+                Console.ForegroundColor = ConsoleColor.Yellow;
+                Console.WriteLine("Saved session!");
+                Console.ForegroundColor = ConsoleColor.White;
+
+                ex.Model.Dispose();
+                ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));
+                session = new ChatSession(ex).WithOutputTransform(new LLamaTransforms.KeywordTextOutputStreamTransform(new string[] { "User:", "Bob:" }, redundancyLength: 8));
+                session.LoadSession(statePath);
+
+                Console.ForegroundColor = ConsoleColor.Yellow;
+                Console.WriteLine("Loaded session!");
+                Console.ForegroundColor = ConsoleColor.White;
+
+                Console.Write("Now you can continue your session: ");
+                Console.ForegroundColor = ConsoleColor.Green;
+                prompt = Console.ReadLine();
+                Console.ForegroundColor = ConsoleColor.White;
+            }
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/LoadAndSaveState/index.html b/0.5/Examples/LoadAndSaveState/index.html new file mode 100755 index 00000000..7cebe736 --- /dev/null +++ b/0.5/Examples/LoadAndSaveState/index.html @@ -0,0 +1,2104 @@ + + + + + + + + + + + + + + + + + + + + + + Load/Save state - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Load and save model/executor state

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class LoadAndSaveState
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+        var prompt = File.ReadAllText("Assets/chat-with-bob.txt").Trim();
+
+        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to 64 and the context size is 256. (an example for small scale usage)");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        Console.Write(prompt);
+
+        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } };
+
+        while (true)
+        {
+            foreach (var text in ex.Infer(prompt, inferenceParams))
+            {
+                Console.Write(text);
+            }
+
+            prompt = Console.ReadLine();
+            if (prompt == "save")
+            {
+                Console.Write("Your path to save model state: ");
+                string modelStatePath = Console.ReadLine();
+                ex.Model.SaveState(modelStatePath);
+
+                Console.Write("Your path to save executor state: ");
+                string executorStatePath = Console.ReadLine();
+                ex.SaveState(executorStatePath);
+
+                Console.ForegroundColor = ConsoleColor.Yellow;
+                Console.WriteLine("All states saved!");
+                Console.ForegroundColor = ConsoleColor.White;
+
+                var model = ex.Model;
+                model.LoadState(modelStatePath);
+                ex = new InteractiveExecutor(model);
+                ex.LoadState(executorStatePath);
+                Console.ForegroundColor = ConsoleColor.Yellow;
+                Console.WriteLine("Loaded state!");
+                Console.ForegroundColor = ConsoleColor.White;
+
+                Console.Write("Now you can continue your session: ");
+                Console.ForegroundColor = ConsoleColor.Green;
+                prompt = Console.ReadLine();
+                Console.ForegroundColor = ConsoleColor.White;
+            }
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/QuantizeModel/index.html b/0.5/Examples/QuantizeModel/index.html new file mode 100755 index 00000000..d17ec820 --- /dev/null +++ b/0.5/Examples/QuantizeModel/index.html @@ -0,0 +1,2068 @@ + + + + + + + + + + + + + + + + + + + + + + Quantize model - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Quantize model

+
using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading;
+using System.Threading.Tasks;
+
+public class QuantizeModel
+{
+    public static void Run()
+    {
+        Console.Write("Please input your original model path: ");
+        var inputPath = Console.ReadLine();
+        Console.Write("Please input your output model path: ");
+        var outputPath = Console.ReadLine();
+        Console.Write("Please input the quantize type (one of q4_0, q4_1, q5_0, q5_1, q8_0): ");
+        var quantizeType = Console.ReadLine();
+        if (LLamaQuantizer.Quantize(inputPath, outputPath, quantizeType))
+        {
+            Console.WriteLine("Quantization succeed!");
+        }
+        else
+        {
+            Console.WriteLine("Quantization failed!");
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Examples/StatelessModeExecute/index.html b/0.5/Examples/StatelessModeExecute/index.html new file mode 100755 index 00000000..0b57b441 --- /dev/null +++ b/0.5/Examples/StatelessModeExecute/index.html @@ -0,0 +1,2081 @@ + + + + + + + + + + + + + + + + + + + + + + Stateless exeutor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use stateless executor

+
using LLama.Common;
+using System;
+using System.Collections.Generic;
+using System.Linq;
+using System.Text;
+using System.Threading.Tasks;
+
+public class StatelessModeExecute
+{
+    public static void Run()
+    {
+        Console.Write("Please input your model path: ");
+        string modelPath = Console.ReadLine();
+
+        StatelessExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));
+
+        Console.ForegroundColor = ConsoleColor.Yellow;
+        Console.WriteLine("The executor has been enabled. In this example, the inference is an one-time job. That says, the previous input and response has " +
+            "no impact on the current response. Now you can ask it questions. Note that in this example, no prompt was set for LLM and the maximum response tokens is 50. " +
+            "It may not perform well because of lack of prompt. This is also an example that could indicate the improtance of prompt in LLM. To improve it, you can add " +
+            "a prompt for it yourself!");
+        Console.ForegroundColor = ConsoleColor.White;
+
+        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "Question:", "#", "Question: ", ".\n" }, MaxTokens = 50 };
+
+        while (true)
+        {
+            Console.Write("\nQuestion: ");
+            Console.ForegroundColor = ConsoleColor.Green;
+            string prompt = Console.ReadLine();
+            Console.ForegroundColor = ConsoleColor.White; 
+            Console.Write("Answer: ");
+            prompt = $"Question: {prompt.Trim()} Answer: ";
+            foreach (var text in ex.Infer(prompt, inferenceParams))
+            {
+                Console.Write(text);
+            }
+        }
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/GetStarted/index.html b/0.5/GetStarted/index.html new file mode 100755 index 00000000..235e1bde --- /dev/null +++ b/0.5/GetStarted/index.html @@ -0,0 +1,2239 @@ + + + + + + + + + + + + + + + + + + + + + + Get Started - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Get Started

+

Install packages

+

Firstly, search LLamaSharp in nuget package manager and install it.

+
PM> Install-Package LLamaSharp
+
+

Then, search and install one of the following backends:

+
LLamaSharp.Backend.Cpu
+LLamaSharp.Backend.Cuda11
+LLamaSharp.Backend.Cuda12
+
+

Here's the mapping of them and corresponding model samples provided by LLamaSharp. If you're not sure which model is available for a version, please try our sample model.

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
LLamaSharp.BackendLLamaSharpVerified Model Resourcesllama.cpp commit id
-v0.2.0This version is not recommended to use.-
-v0.2.1WizardLM, Vicuna (filenames with "old")-
v0.2.2v0.2.2, v0.2.3WizardLM, Vicuna (filenames without "old")63d2046
v0.3.0v0.3.0LLamaSharpSamples v0.3.0, WizardLM7e4ea5b
+

Download a model

+

One of the following models could be okay:

+ +

Note that because llama.cpp is under fast development now and often introduce break changes, some model weights on huggingface which works under a version may be invalid with another version. If it's your first time to configure LLamaSharp, we'd like to suggest for using verified model weights in the table above.

+

Run the program

+

Please create a console program with dotnet runtime >= netstandard 2.0 (>= net6.0 is more recommended). Then, paste the following code to program.cs;

+
using LLama.Common;
+using LLama;
+
+string modelPath = "<Your model path>" // change it to your own model path
+var prompt = "Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.\r\n\r\nUser: Hello, Bob.\r\nBob: Hello. How may I help you today?\r\nUser: Please tell me the largest city in Europe.\r\nBob: Sure. The largest city in Europe is Moscow, the capital of Russia.\r\nUser:"; // use the "chat-with-bob" prompt here.
+
+// Load model
+var parameters = new ModelParams(modelPath)
+{
+    ContextSize = 1024
+};
+using var model = LLamaWeights.LoadFromFile(parameters);
+
+// Initialize a chat session
+using var context = model.CreateContext(parameters);
+var ex = new InteractiveExecutor(context);
+ChatSession session = new ChatSession(ex);
+
+// show the prompt
+Console.WriteLine();
+Console.Write(prompt);
+
+// run the inference in a loop to chat with LLM
+while (true)
+{
+    foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { "User:" } }))
+    {
+        Console.Write(text);
+    }
+
+    Console.ForegroundColor = ConsoleColor.Green;
+    prompt = Console.ReadLine();
+    Console.ForegroundColor = ConsoleColor.White;
+}
+
+

After starting it, you'll see the following outputs.

+
Please input your model path: D:\development\llama\weights\wizard-vicuna-13B.ggmlv3.q4_1.bin
+llama.cpp: loading model from D:\development\llama\weights\wizard-vicuna-13B.ggmlv3.q4_1.bin
+llama_model_load_internal: format     = ggjt v3 (latest)
+llama_model_load_internal: n_vocab    = 32000
+llama_model_load_internal: n_ctx      = 1024
+llama_model_load_internal: n_embd     = 5120
+llama_model_load_internal: n_mult     = 256
+llama_model_load_internal: n_head     = 40
+llama_model_load_internal: n_layer    = 40
+llama_model_load_internal: n_rot      = 128
+llama_model_load_internal: ftype      = 3 (mostly Q4_1)
+llama_model_load_internal: n_ff       = 13824
+llama_model_load_internal: n_parts    = 1
+llama_model_load_internal: model size = 13B
+llama_model_load_internal: ggml ctx size = 7759.48 MB
+llama_model_load_internal: mem required  = 9807.48 MB (+ 1608.00 MB per state)
+....................................................................................................
+llama_init_from_file: kv self size  =  800.00 MB
+
+Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.
+
+User: Hello, Bob.
+Bob: Hello. How may I help you today?
+User: Please tell me the largest city in Europe.
+Bob: Sure. The largest city in Europe is Moscow, the capital of Russia.
+User:
+
+

Now, enjoy chatting with LLM!

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/HighLevelApps/bot-sharp/index.html b/0.5/HighLevelApps/bot-sharp/index.html new file mode 100755 index 00000000..1e00de45 --- /dev/null +++ b/0.5/HighLevelApps/bot-sharp/index.html @@ -0,0 +1,2041 @@ + + + + + + + + + + + + + + + + + + + + + + BotSharp - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

The Usage of BotSharp Integration

+

The document is under work, please have a wait. Thank you for your support! :)

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/HighLevelApps/semantic-kernel/index.html b/0.5/HighLevelApps/semantic-kernel/index.html new file mode 100755 index 00000000..691904e6 --- /dev/null +++ b/0.5/HighLevelApps/semantic-kernel/index.html @@ -0,0 +1,2041 @@ + + + + + + + + + + + + + + + + + + + + + + semantic-kernel - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

The Usage of semantic-kernel Integration

+

Please see this doc

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaExecutors/differences/index.html b/0.5/LLamaExecutors/differences/index.html new file mode 100755 index 00000000..04088f2e --- /dev/null +++ b/0.5/LLamaExecutors/differences/index.html @@ -0,0 +1,2150 @@ + + + + + + + + + + + + + + + + + + + + + + Differences of Executors - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Differences of Executors

+ +

Differences between the executors

+

There're currently three kinds of executors provided, which are InteractiveExecutor, InstructExecutor and StatelessExecutor.

+

In a word, InteractiveExecutor is suitable for getting answer of your questions from LLM continuously. InstructExecutor let LLM execute your instructions, such as "continue writing". StatelessExecutor is best for one-time job because the previous inference has no impact on the current inference.

+

Interactive mode & Instruct mode

+

Both of them are taking "completing the prompt" as the goal to generate the response. For example, if you input Long long ago, there was a fox who wanted to make friend with humen. One day, then the LLM will continue to write the story.

+

Under interactive mode, you serve a role of user and the LLM serves the role of assistant. Then it will help you with your question or request.

+

Under instruct mode, you give LLM some instructions and it follows.

+

Though the behaviors of them sounds similar, it could introduce many differences depending on your prompt. For example, "chat-with-bob" has good performance under interactive mode and alpaca does well with instruct mode.

+
// chat-with-bob
+
+Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.
+
+User: Hello, Bob.
+Bob: Hello. How may I help you today?
+User: Please tell me the largest city in Europe.
+Bob: Sure. The largest city in Europe is Moscow, the capital of Russia.
+User:
+
+
// alpaca
+
+Below is an instruction that describes a task. Write a response that appropriately completes the request.
+
+

Therefore, please modify the prompt correspondingly when switching from one mode to the other.

+

Stateful mode and Stateless mode.

+

Despite the differences between interactive mode and instruct mode, both of them are stateful mode. That is, your previous question/instruction will impact on the current response from LLM. On the contrary, the stateless executor does not have such a "memory". No matter how many times you talk to it, it will only concentrate on what you say in this time.

+

Since the stateless executor has no memory of conversations before, you need to input your question with the whole prompt into it to get the better answer.

+

For example, if you feed Q: Who is Trump? A: to the stateless executor, it may give the following answer with the antiprompt Q:.

+
Donald J. Trump, born June 14, 1946, is an American businessman, television personality, politician and the 45th President of the United States (2017-2021). # Anexo:Torneo de Hamburgo 2022 (individual masculino)
+
+## Presentación previa
+
+* Defensor del título:  Daniil Medvédev
+
+

It seems that things went well at first. However, after answering the question itself, LLM began to talk about some other things until the answer reached the token count limit. The reason of this strange behavior is the anti-prompt cannot be match. With the input, LLM cannot decide whether to append a string "A: " at the end of the response.

+

As an improvement, let's take the following text as the input:

+
Q: What is the capital of the USA? A: Washingtong. Q: What is the sum of 1 and 2? A: 3. Q: Who is Trump? A: 
+
+

Then, I got the following answer with the anti-prompt Q:.

+
45th president of the United States.
+
+

At this time, by repeating the same mode of Q: xxx? A: xxx., LLM outputs the anti-prompt we want to help to decide where to stop the generation.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaExecutors/parameters/index.html b/0.5/LLamaExecutors/parameters/index.html new file mode 100755 index 00000000..22b7bc07 --- /dev/null +++ b/0.5/LLamaExecutors/parameters/index.html @@ -0,0 +1,2172 @@ + + + + + + + + + + + + + + + + + + + + + + Inference Parameters - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Inference Parameters

+

Different from LLamaModel, when using an executor, InferenceParams is passed to the Infer method instead of constructor. This is because executors only define the ways to run the model, therefore in each run, you can change the settings for this time inference.

+

InferenceParams

+

Namespace: LLama.Common

+
public class InferenceParams
+
+

Inheritance ObjectInferenceParams

+

Properties

+

TokensKeep

+

number of tokens to keep from initial prompt

+
public int TokensKeep { get; set; }
+
+

Property Value

+

Int32

+

MaxTokens

+

how many new tokens to predict (n_predict), set to -1 to infinitely generate response + until it complete.

+
public int MaxTokens { get; set; }
+
+

Property Value

+

Int32

+

LogitBias

+

logit bias for specific tokens

+
public Dictionary<int, float> LogitBias { get; set; }
+
+

Property Value

+

Dictionary<Int32, Single>

+

AntiPrompts

+

Sequences where the model will stop generating further tokens.

+
public IEnumerable<string> AntiPrompts { get; set; }
+
+

Property Value

+

IEnumerable<String>

+

PathSession

+

path to file for saving/loading model eval state

+
public string PathSession { get; set; }
+
+

Property Value

+

String

+

InputSuffix

+

string to suffix user inputs with

+
public string InputSuffix { get; set; }
+
+

Property Value

+

String

+

InputPrefix

+

string to prefix user inputs with

+
public string InputPrefix { get; set; }
+
+

Property Value

+

String

+

TopK

+

0 or lower to use vocab size

+
public int TopK { get; set; }
+
+

Property Value

+

Int32

+

TopP

+

1.0 = disabled

+
public float TopP { get; set; }
+
+

Property Value

+

Single

+

TfsZ

+

1.0 = disabled

+
public float TfsZ { get; set; }
+
+

Property Value

+

Single

+

TypicalP

+

1.0 = disabled

+
public float TypicalP { get; set; }
+
+

Property Value

+

Single

+

Temperature

+

1.0 = disabled

+
public float Temperature { get; set; }
+
+

Property Value

+

Single

+

RepeatPenalty

+

1.0 = disabled

+
public float RepeatPenalty { get; set; }
+
+

Property Value

+

Single

+

RepeatLastTokensCount

+

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

+
public int RepeatLastTokensCount { get; set; }
+
+

Property Value

+

Int32

+

FrequencyPenalty

+

frequency penalty coefficient + 0.0 = disabled

+
public float FrequencyPenalty { get; set; }
+
+

Property Value

+

Single

+

PresencePenalty

+

presence penalty coefficient + 0.0 = disabled

+
public float PresencePenalty { get; set; }
+
+

Property Value

+

Single

+

Mirostat

+

Mirostat uses tokens instead of words. + algorithm described in the paper https://arxiv.org/abs/2007.14966. + 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

+
public MiroStateType Mirostat { get; set; }
+
+

Property Value

+

MiroStateType

+

MirostatTau

+

target entropy

+
public float MirostatTau { get; set; }
+
+

Property Value

+

Single

+

MirostatEta

+

learning rate

+
public float MirostatEta { get; set; }
+
+

Property Value

+

Single

+

PenalizeNL

+

consider newlines as a repeatable token (penalize_nl)

+
public bool PenalizeNL { get; set; }
+
+

Property Value

+

Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaExecutors/save-load-state/index.html b/0.5/LLamaExecutors/save-load-state/index.html new file mode 100755 index 00000000..05d97b96 --- /dev/null +++ b/0.5/LLamaExecutors/save-load-state/index.html @@ -0,0 +1,2059 @@ + + + + + + + + + + + + + + + + + + + + + + Save/Load State - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Save/Load State of Executor

+

Similar to LLamaModel, an executor also has its state, which can be saved and loaded. Note that in most of cases, the state of executor and the state of the model should be loaded and saved at the same time.

+

To decouple the model and executor, we provide APIs to save/load state for model and executor respectively. However, during the inference, the processed information will leave footprint in LLamaModel's native context. Therefore, if you just load a state from another executor but keep the model unmodified, some strange things may happen. So will loading model state only.

+

Is there a condition that requires to load one of them only? The answer is YES. For example, after resetting the model state, if you don't want the inference starting from the new position, leaving the executor unmodified is okay. But, anyway, this flexible usage may cause some unexpected behaviors, therefore please ensure you know what you're doing before using it in this way.

+

In the future version, we'll open the access for some variables inside the executor to support more flexible usages.

+

The APIs to load/save state of the executors is similar to that of LLamaModel. However, note that StatelessExecutor doesn't have such APIs because it's stateless itself. Besides, the output of GetStateData is an object of type ExecutorBaseState.

+
LLamaModel model = new LLamaModel(new ModelParams("<modelPath>"));
+InteractiveExecutor executor = new InteractiveExecutor(model);
+// do some things...
+executor.SaveState("executor.st");
+var stateData = model.GetStateData();
+
+InteractiveExecutor executor2 = new InteractiveExecutor(model);
+executor2.LoadState(stateData);
+// do some things...
+
+InteractiveExecutor executor3 = new InteractiveExecutor(model);
+executor3.LoadState("executor.st");
+// do some things...
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaExecutors/text-to-text-apis/index.html b/0.5/LLamaExecutors/text-to-text-apis/index.html new file mode 100755 index 00000000..b74ffeb8 --- /dev/null +++ b/0.5/LLamaExecutors/text-to-text-apis/index.html @@ -0,0 +1,2052 @@ + + + + + + + + + + + + + + + + + + + + + + Text-to-Text APIs - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Text-to-Text APIs of the executors

+

All the executors implements the interface ILLamaExecutor, which provides two APIs to execute text-to-text tasks.

+
public interface ILLamaExecutor
+{
+    public LLamaModel Model { get; }
+
+    IEnumerable<string> Infer(string text, InferenceParams? inferenceParams = null, CancellationToken token = default);
+
+    IAsyncEnumerable<string> InferAsync(string text, InferenceParams? inferenceParams = null, CancellationToken token = default);
+}
+
+

Just pass the text to the executor with the inference parameters. For the inference parameters, please refer to executor inference parameters doc.

+

The output of both two APIs are yield enumerable. Therefore, when receiving the output, you can directly use foreach to take actions on each word you get by order, instead of waiting for the whole process completed.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaModel/embeddings/index.html b/0.5/LLamaModel/embeddings/index.html new file mode 100755 index 00000000..3531380a --- /dev/null +++ b/0.5/LLamaModel/embeddings/index.html @@ -0,0 +1,2031 @@ + + + + + + + + + + + + + + + + + + Get Embeddings - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Get Embeddings

+

Getting the embeddings of a text in LLM is sometimes useful, for example, to train other MLP models.

+

To get the embeddings, please initialize a LLamaEmbedder and then call GetEmbeddings.

+
var embedder = new LLamaEmbedder(new ModelParams("<modelPath>"));
+string text = "hello, LLM.";
+float[] embeddings = embedder.GetEmbeddings(text);
+
+

The output is a float array. Note that the length of the array is related with the model you load. If you just want to get a smaller size embedding, please consider changing a model.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaModel/parameters/index.html b/0.5/LLamaModel/parameters/index.html new file mode 100755 index 00000000..d678f418 --- /dev/null +++ b/0.5/LLamaModel/parameters/index.html @@ -0,0 +1,2135 @@ + + + + + + + + + + + + + + + + + + LLamaModel Parameters - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaModel Parameters

+

When initializing a LLamaModel object, there're three parameters, ModelParams Params, string encoding = "UTF-8", ILLamaLogger? logger = null.

+

The usage of logger will be further introduced in logger doc. The encoding is the encoding you want to use when dealing with text via this model.

+

The most important of all, is the ModelParams, which is defined as below. We'll explain the parameters step by step in this document.

+
public class ModelParams
+{
+    public int ContextSize { get; set; } = 512;
+    public int GpuLayerCount { get; set; } = 20;
+    public int Seed { get; set; } = 1686349486;
+    public bool UseFp16Memory { get; set; } = true;
+    public bool UseMemorymap { get; set; } = true;
+    public bool UseMemoryLock { get; set; } = false;
+    public bool Perplexity { get; set; } = false;
+    public string ModelPath { get; set; }
+    public string LoraAdapter { get; set; } = string.Empty;
+    public string LoraBase { get; set; } = string.Empty;
+    public int Threads { get; set; } = Math.Max(Environment.ProcessorCount / 2, 1);
+    public int BatchSize { get; set; } = 512;
+    public bool ConvertEosToNewLine { get; set; } = false;
+}
+
+

ModelParams

+

Namespace: LLama.Common

+
public class ModelParams
+
+

Inheritance ObjectModelParams

+

Properties

+

ContextSize

+

Model context size (n_ctx)

+
public int ContextSize { get; set; }
+
+

Property Value

+

Int32

+

GpuLayerCount

+

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

+
public int GpuLayerCount { get; set; }
+
+

Property Value

+

Int32

+

Seed

+

Seed for the random number generator (seed)

+
public int Seed { get; set; }
+
+

Property Value

+

Int32

+

UseFp16Memory

+

Use f16 instead of f32 for memory kv (memory_f16)

+
public bool UseFp16Memory { get; set; }
+
+

Property Value

+

Boolean

+

UseMemorymap

+

Use mmap for faster loads (use_mmap)

+
public bool UseMemorymap { get; set; }
+
+

Property Value

+

Boolean

+

UseMemoryLock

+

Use mlock to keep model in memory (use_mlock)

+
public bool UseMemoryLock { get; set; }
+
+

Property Value

+

Boolean

+

Perplexity

+

Compute perplexity over the prompt (perplexity)

+
public bool Perplexity { get; set; }
+
+

Property Value

+

Boolean

+

ModelPath

+

Model path (model)

+
public string ModelPath { get; set; }
+
+

Property Value

+

String

+

LoraAdapter

+

lora adapter path (lora_adapter)

+
public string LoraAdapter { get; set; }
+
+

Property Value

+

String

+

LoraBase

+

base model path for the lora adapter (lora_base)

+
public string LoraBase { get; set; }
+
+

Property Value

+

String

+

Threads

+

Number of threads (-1 = autodetect) (n_threads)

+
public int Threads { get; set; }
+
+

Property Value

+

Int32

+

BatchSize

+

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

+
public int BatchSize { get; set; }
+
+

Property Value

+

Int32

+

ConvertEosToNewLine

+

Whether to convert eos to newline during the inference.

+
public bool ConvertEosToNewLine { get; set; }
+
+

Property Value

+

Boolean

+

EmbeddingMode

+

Whether to use embedding mode. (embedding) Note that if this is set to true, + The LLamaModel won't produce text response anymore.

+
public bool EmbeddingMode { get; set; }
+
+

Property Value

+

Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaModel/quantization/index.html b/0.5/LLamaModel/quantization/index.html new file mode 100755 index 00000000..30b8d6ef --- /dev/null +++ b/0.5/LLamaModel/quantization/index.html @@ -0,0 +1,2041 @@ + + + + + + + + + + + + + + + + + + Quantization - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Quantization

+

Quantization is significant to accelerate the model inference. Since there's little accuracy (performance) reduction when quantizing the model, get it easy to quantize it!

+

To quantize the model, please call Quantize from LLamaQuantizer, which is a static method.

+
string srcPath = "<model.bin>";
+string dstPath = "<model_q4_0.bin>";
+LLamaQuantizer.Quantize(srcPath, dstPath, "q4_0");
+// The following overload is also okay.
+// LLamaQuantizer.Quantize(srcPath, dstPath, LLamaFtype.LLAMA_FTYPE_MOSTLY_Q4_0);
+
+

After calling it, a quantized model file will be saved.

+

There're currently 5 types of quantization supported:

+
    +
  • q4_0
  • +
  • q4_1
  • +
  • q5_0
  • +
  • q5_1
  • +
  • q8_0
  • +
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaModel/save-load-state/index.html b/0.5/LLamaModel/save-load-state/index.html new file mode 100755 index 00000000..b5e93265 --- /dev/null +++ b/0.5/LLamaModel/save-load-state/index.html @@ -0,0 +1,2039 @@ + + + + + + + + + + + + + + + + + + Save/Load State - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Save/Load State

+

There're two ways to load state: loading from path and loading from bite array. Therefore, correspondingly, state data can be extracted as byte array or saved to a file.

+
LLamaModel model = new LLamaModel(new ModelParams("<modelPath>"));
+// do some things...
+model.SaveState("model.st");
+var stateData = model.GetStateData();
+model.Dispose();
+
+LLamaModel model2 = new LLamaModel(new ModelParams("<modelPath>"));
+model2.LoadState(stateData);
+// do some things...
+
+LLamaModel model3 = new LLamaModel(new ModelParams("<modelPath>"));
+model3.LoadState("model.st");
+// do some things...
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/LLamaModel/tokenization/index.html b/0.5/LLamaModel/tokenization/index.html new file mode 100755 index 00000000..de4c6dc0 --- /dev/null +++ b/0.5/LLamaModel/tokenization/index.html @@ -0,0 +1,2060 @@ + + + + + + + + + + + + + + + + + + Tokenization/Detokenization - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Tokenization/Detokenization

+

A pair of APIs to make conversion between text and tokens.

+

Tokenization

+

The basic usage is to call Tokenize after initializing the model.

+
LLamaModel model = new LLamaModel(new ModelParams("<modelPath>"));
+string text = "hello";
+int[] tokens = model.Tokenize(text).ToArray();
+
+

Depending on different model (or vocab), the output will be various.

+

Detokenization

+

Similar to tokenization, just pass an IEnumerable<int> to Detokenize method.

+
LLamaModel model = new LLamaModel(new ModelParams("<modelPath>"));
+int[] tokens = new int[] {125, 2568, 13245};
+string text = model.Detokenize(tokens);
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/More/log/index.html b/0.5/More/log/index.html new file mode 100755 index 00000000..ba58a4f6 --- /dev/null +++ b/0.5/More/log/index.html @@ -0,0 +1,2237 @@ + + + + + + + + + + + + + + + + + + + + + + Logger - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

The Logger in LLamaSharp

+

LLamaSharp supports customized logger because it could be used in many kinds of applications, like Winform/WPF, WebAPI and Blazor, so that the preference of logger varies.

+

Define customized logger

+

What you need to do is to implement the ILogger interface.

+
public interface ILLamaLogger
+{
+    public enum LogLevel
+    {
+        Info,
+        Debug,
+        Warning,
+        Error
+    }
+    void Log(string source, string message, LogLevel level);
+}
+
+

The source specifies where the log message is from, which could be a function, a class, etc..

+

The message is the log message itself.

+

The level is the level of the information in the log. As shown above, there're four levels, which are info, debug, warning and error respectively.

+

The following is a simple example of the logger implementation:

+
public sealed class LLamaDefaultLogger : ILLamaLogger
+{
+    private static readonly Lazy<LLamaDefaultLogger> _instance = new Lazy<LLamaDefaultLogger>(() => new LLamaDefaultLogger());
+
+    private bool _toConsole = true;
+    private bool _toFile = false;
+
+    private FileStream? _fileStream = null;
+    private StreamWriter _fileWriter = null;
+
+    public static LLamaDefaultLogger Default => _instance.Value;
+
+    private LLamaDefaultLogger()
+    {
+
+    }
+
+    public LLamaDefaultLogger EnableConsole()
+    {
+        _toConsole = true;
+        return this;
+    }
+
+    public LLamaDefaultLogger DisableConsole()
+    {
+        _toConsole = false;
+        return this;
+    }
+
+    public LLamaDefaultLogger EnableFile(string filename, FileMode mode = FileMode.Append)
+    {
+        _fileStream = new FileStream(filename, mode, FileAccess.Write);
+        _fileWriter = new StreamWriter(_fileStream);
+        _toFile = true;
+        return this;
+    }
+
+    public LLamaDefaultLogger DisableFile(string filename)
+    {
+        if (_fileWriter is not null)
+        {
+            _fileWriter.Close();
+            _fileWriter = null;
+        }
+        if (_fileStream is not null)
+        {
+            _fileStream.Close();
+            _fileStream = null;
+        }
+        _toFile = false;
+        return this;
+    }
+
+    public void Log(string source, string message, LogLevel level)
+    {
+        if (level == LogLevel.Info)
+        {
+            Info(message);
+        }
+        else if (level == LogLevel.Debug)
+        {
+
+        }
+        else if (level == LogLevel.Warning)
+        {
+            Warn(message);
+        }
+        else if (level == LogLevel.Error)
+        {
+            Error(message);
+        }
+    }
+
+    public void Info(string message)
+    {
+        message = MessageFormat("info", message);
+        if (_toConsole)
+        {
+            Console.ForegroundColor = ConsoleColor.White;
+            Console.WriteLine(message);
+            Console.ResetColor();
+        }
+        if (_toFile)
+        {
+            Debug.Assert(_fileStream is not null);
+            Debug.Assert(_fileWriter is not null);
+            _fileWriter.WriteLine(message);
+        }
+    }
+
+    public void Warn(string message)
+    {
+        message = MessageFormat("warn", message);
+        if (_toConsole)
+        {
+            Console.ForegroundColor = ConsoleColor.Yellow;
+            Console.WriteLine(message);
+            Console.ResetColor();
+        }
+        if (_toFile)
+        {
+            Debug.Assert(_fileStream is not null);
+            Debug.Assert(_fileWriter is not null);
+            _fileWriter.WriteLine(message);
+        }
+    }
+
+    public void Error(string message)
+    {
+        message = MessageFormat("error", message);
+        if (_toConsole)
+        {
+            Console.ForegroundColor = ConsoleColor.Red;
+            Console.WriteLine(message);
+            Console.ResetColor();
+        }
+        if (_toFile)
+        {
+            Debug.Assert(_fileStream is not null);
+            Debug.Assert(_fileWriter is not null);
+            _fileWriter.WriteLine(message);
+        }
+    }
+
+    private string MessageFormat(string level, string message)
+    {
+        DateTime now = DateTime.Now;
+        string formattedDate = now.ToString("yyyy.MM.dd HH:mm:ss");
+        return $"[{formattedDate}][{level}]: {message}";
+    }
+}
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/NonEnglishUsage/Chinese/index.html b/0.5/NonEnglishUsage/Chinese/index.html new file mode 100755 index 00000000..92f2026e --- /dev/null +++ b/0.5/NonEnglishUsage/Chinese/index.html @@ -0,0 +1,2041 @@ + + + + + + + + + + + + + + + + + + + + + + Chinese - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Use LLamaSharp with Chinese

+

It's supported now but the document is under work. Please wait for some time. Thank you for your support! :)

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/Tricks/index.html b/0.5/Tricks/index.html new file mode 100755 index 00000000..4cb990c1 --- /dev/null +++ b/0.5/Tricks/index.html @@ -0,0 +1,2164 @@ + + + + + + + + + + + + + + + + + + + + + + Tricks for FAQ - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ + + + Skip to content + + +
+
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + + + + + +
+
+ + + + +

Tricks for FAQ

+

Sometimes, your application with LLM and LLamaSharp may have strange behaviours. Before opening an issue to report the BUG, the following tricks may worth a try.

+

Carefully set the anti-prompts

+

Anti-prompt can also be called as "Stop-keyword", which decides when to stop the response generation. Under interactive mode, the maximum tokens count is always not set, which makes the LLM generates responses infinitively. Therefore, setting anti-prompt correctly helps a lot to avoid the strange behaviours. For example, the prompt file chat-with-bob.txt has the following content:

+
Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.
+
+User: Hello, Bob.
+Bob: Hello. How may I help you today?
+User: Please tell me the largest city in Europe.
+Bob: Sure. The largest city in Europe is Moscow, the capital of Russia.
+User:
+
+

Therefore, the anti-prompt should be set as "User:". If the last line of the prompt is removed, LLM will automatically generate a question (user) and a response (bob) for one time when running the chat session. Therefore, the antiprompt is suggested to be appended to the prompt when starting a chat session.

+

What if an extra line is appended? The string "User:" in the prompt will be followed with a char "\n". Thus when running the model, the automatic generation of a pair of question and response may appear because the anti-prompt is "User:" but the last token is "User:\n". As for whether it will appear, it's an undefined behaviour, which depends on the implementation inside the LLamaExecutor. Anyway, since it may leads to unexpected behaviors, it's recommended to trim your prompt or carefully keep consistent with your anti-prompt.

+

Pay attention to the length of prompt

+

Sometimes we want to input a long prompt to execute a task. However, the context size may limit the inference of LLama model. Please ensure the inequality below holds.

+

$$ len(prompt) + len(response) < len(context) $$

+

In this inequality, len(response) refers to the expected tokens for LLM to generate.

+

Try different executors with a prompt

+

Some prompt works well under interactive mode, such as chat-with-bob, some others may work well with instruct mode, such as alpaca. Besides, if your input is quite simple and one-time job, such as "Q: what is the satellite of the earth? A: ", stateless mode will be a good choice.

+

If your chat bot has bad performance, trying different executor will possibly make it work well.

+

Choose models weight depending on you task

+

The differences between modes may lead to much different behaviours under the same task. For example, if you're building a chat bot with non-English, a fine-tuned model specially for the language you want to use will have huge effect on the performance.

+

Set the layer count you want to offload to GPU

+

Currently, the GpuLayerCount parameter, which decides the number of layer loaded into GPU, is set to 20 by default. However, if you have some efficient GPUs, setting it as a larger number will attain faster inference.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/assets/images/favicon.png b/0.5/assets/images/favicon.png new file mode 100755 index 00000000..1cf13b9f Binary files /dev/null and b/0.5/assets/images/favicon.png differ diff --git a/0.5/assets/javascripts/bundle.a51614de.min.js b/0.5/assets/javascripts/bundle.a51614de.min.js new file mode 100755 index 00000000..5afb7820 --- /dev/null +++ b/0.5/assets/javascripts/bundle.a51614de.min.js @@ -0,0 +1,29 @@ +"use strict";(()=>{var Ci=Object.create;var gr=Object.defineProperty;var Ri=Object.getOwnPropertyDescriptor;var ki=Object.getOwnPropertyNames,Ht=Object.getOwnPropertySymbols,Hi=Object.getPrototypeOf,yr=Object.prototype.hasOwnProperty,nn=Object.prototype.propertyIsEnumerable;var rn=(e,t,r)=>t in e?gr(e,t,{enumerable:!0,configurable:!0,writable:!0,value:r}):e[t]=r,P=(e,t)=>{for(var r in t||(t={}))yr.call(t,r)&&rn(e,r,t[r]);if(Ht)for(var r of Ht(t))nn.call(t,r)&&rn(e,r,t[r]);return e};var on=(e,t)=>{var r={};for(var n in e)yr.call(e,n)&&t.indexOf(n)<0&&(r[n]=e[n]);if(e!=null&&Ht)for(var n of Ht(e))t.indexOf(n)<0&&nn.call(e,n)&&(r[n]=e[n]);return r};var Pt=(e,t)=>()=>(t||e((t={exports:{}}).exports,t),t.exports);var Pi=(e,t,r,n)=>{if(t&&typeof t=="object"||typeof t=="function")for(let o of ki(t))!yr.call(e,o)&&o!==r&&gr(e,o,{get:()=>t[o],enumerable:!(n=Ri(t,o))||n.enumerable});return e};var yt=(e,t,r)=>(r=e!=null?Ci(Hi(e)):{},Pi(t||!e||!e.__esModule?gr(r,"default",{value:e,enumerable:!0}):r,e));var sn=Pt((xr,an)=>{(function(e,t){typeof xr=="object"&&typeof an!="undefined"?t():typeof define=="function"&&define.amd?define(t):t()})(xr,function(){"use strict";function e(r){var n=!0,o=!1,i=null,s={text:!0,search:!0,url:!0,tel:!0,email:!0,password:!0,number:!0,date:!0,month:!0,week:!0,time:!0,datetime:!0,"datetime-local":!0};function a(O){return!!(O&&O!==document&&O.nodeName!=="HTML"&&O.nodeName!=="BODY"&&"classList"in O&&"contains"in O.classList)}function f(O){var Qe=O.type,De=O.tagName;return!!(De==="INPUT"&&s[Qe]&&!O.readOnly||De==="TEXTAREA"&&!O.readOnly||O.isContentEditable)}function c(O){O.classList.contains("focus-visible")||(O.classList.add("focus-visible"),O.setAttribute("data-focus-visible-added",""))}function u(O){O.hasAttribute("data-focus-visible-added")&&(O.classList.remove("focus-visible"),O.removeAttribute("data-focus-visible-added"))}function p(O){O.metaKey||O.altKey||O.ctrlKey||(a(r.activeElement)&&c(r.activeElement),n=!0)}function m(O){n=!1}function d(O){a(O.target)&&(n||f(O.target))&&c(O.target)}function h(O){a(O.target)&&(O.target.classList.contains("focus-visible")||O.target.hasAttribute("data-focus-visible-added"))&&(o=!0,window.clearTimeout(i),i=window.setTimeout(function(){o=!1},100),u(O.target))}function v(O){document.visibilityState==="hidden"&&(o&&(n=!0),Y())}function Y(){document.addEventListener("mousemove",N),document.addEventListener("mousedown",N),document.addEventListener("mouseup",N),document.addEventListener("pointermove",N),document.addEventListener("pointerdown",N),document.addEventListener("pointerup",N),document.addEventListener("touchmove",N),document.addEventListener("touchstart",N),document.addEventListener("touchend",N)}function B(){document.removeEventListener("mousemove",N),document.removeEventListener("mousedown",N),document.removeEventListener("mouseup",N),document.removeEventListener("pointermove",N),document.removeEventListener("pointerdown",N),document.removeEventListener("pointerup",N),document.removeEventListener("touchmove",N),document.removeEventListener("touchstart",N),document.removeEventListener("touchend",N)}function N(O){O.target.nodeName&&O.target.nodeName.toLowerCase()==="html"||(n=!1,B())}document.addEventListener("keydown",p,!0),document.addEventListener("mousedown",m,!0),document.addEventListener("pointerdown",m,!0),document.addEventListener("touchstart",m,!0),document.addEventListener("visibilitychange",v,!0),Y(),r.addEventListener("focus",d,!0),r.addEventListener("blur",h,!0),r.nodeType===Node.DOCUMENT_FRAGMENT_NODE&&r.host?r.host.setAttribute("data-js-focus-visible",""):r.nodeType===Node.DOCUMENT_NODE&&(document.documentElement.classList.add("js-focus-visible"),document.documentElement.setAttribute("data-js-focus-visible",""))}if(typeof window!="undefined"&&typeof document!="undefined"){window.applyFocusVisiblePolyfill=e;var t;try{t=new CustomEvent("focus-visible-polyfill-ready")}catch(r){t=document.createEvent("CustomEvent"),t.initCustomEvent("focus-visible-polyfill-ready",!1,!1,{})}window.dispatchEvent(t)}typeof document!="undefined"&&e(document)})});var cn=Pt(Er=>{(function(e){var t=function(){try{return!!Symbol.iterator}catch(c){return!1}},r=t(),n=function(c){var u={next:function(){var p=c.shift();return{done:p===void 0,value:p}}};return r&&(u[Symbol.iterator]=function(){return u}),u},o=function(c){return encodeURIComponent(c).replace(/%20/g,"+")},i=function(c){return decodeURIComponent(String(c).replace(/\+/g," "))},s=function(){var c=function(p){Object.defineProperty(this,"_entries",{writable:!0,value:{}});var m=typeof p;if(m!=="undefined")if(m==="string")p!==""&&this._fromString(p);else if(p instanceof c){var d=this;p.forEach(function(B,N){d.append(N,B)})}else if(p!==null&&m==="object")if(Object.prototype.toString.call(p)==="[object Array]")for(var h=0;hd[0]?1:0}),c._entries&&(c._entries={});for(var p=0;p1?i(d[1]):"")}})})(typeof global!="undefined"?global:typeof window!="undefined"?window:typeof self!="undefined"?self:Er);(function(e){var t=function(){try{var o=new e.URL("b","http://a");return o.pathname="c d",o.href==="http://a/c%20d"&&o.searchParams}catch(i){return!1}},r=function(){var o=e.URL,i=function(f,c){typeof f!="string"&&(f=String(f)),c&&typeof c!="string"&&(c=String(c));var u=document,p;if(c&&(e.location===void 0||c!==e.location.href)){c=c.toLowerCase(),u=document.implementation.createHTMLDocument(""),p=u.createElement("base"),p.href=c,u.head.appendChild(p);try{if(p.href.indexOf(c)!==0)throw new Error(p.href)}catch(O){throw new Error("URL unable to set base "+c+" due to "+O)}}var m=u.createElement("a");m.href=f,p&&(u.body.appendChild(m),m.href=m.href);var d=u.createElement("input");if(d.type="url",d.value=f,m.protocol===":"||!/:/.test(m.href)||!d.checkValidity()&&!c)throw new TypeError("Invalid URL");Object.defineProperty(this,"_anchorElement",{value:m});var h=new e.URLSearchParams(this.search),v=!0,Y=!0,B=this;["append","delete","set"].forEach(function(O){var Qe=h[O];h[O]=function(){Qe.apply(h,arguments),v&&(Y=!1,B.search=h.toString(),Y=!0)}}),Object.defineProperty(this,"searchParams",{value:h,enumerable:!0});var N=void 0;Object.defineProperty(this,"_updateSearchParams",{enumerable:!1,configurable:!1,writable:!1,value:function(){this.search!==N&&(N=this.search,Y&&(v=!1,this.searchParams._fromString(this.search),v=!0))}})},s=i.prototype,a=function(f){Object.defineProperty(s,f,{get:function(){return this._anchorElement[f]},set:function(c){this._anchorElement[f]=c},enumerable:!0})};["hash","host","hostname","port","protocol"].forEach(function(f){a(f)}),Object.defineProperty(s,"search",{get:function(){return this._anchorElement.search},set:function(f){this._anchorElement.search=f,this._updateSearchParams()},enumerable:!0}),Object.defineProperties(s,{toString:{get:function(){var f=this;return function(){return f.href}}},href:{get:function(){return this._anchorElement.href.replace(/\?$/,"")},set:function(f){this._anchorElement.href=f,this._updateSearchParams()},enumerable:!0},pathname:{get:function(){return this._anchorElement.pathname.replace(/(^\/?)/,"/")},set:function(f){this._anchorElement.pathname=f},enumerable:!0},origin:{get:function(){var f={"http:":80,"https:":443,"ftp:":21}[this._anchorElement.protocol],c=this._anchorElement.port!=f&&this._anchorElement.port!=="";return this._anchorElement.protocol+"//"+this._anchorElement.hostname+(c?":"+this._anchorElement.port:"")},enumerable:!0},password:{get:function(){return""},set:function(f){},enumerable:!0},username:{get:function(){return""},set:function(f){},enumerable:!0}}),i.createObjectURL=function(f){return o.createObjectURL.apply(o,arguments)},i.revokeObjectURL=function(f){return o.revokeObjectURL.apply(o,arguments)},e.URL=i};if(t()||r(),e.location!==void 0&&!("origin"in e.location)){var n=function(){return e.location.protocol+"//"+e.location.hostname+(e.location.port?":"+e.location.port:"")};try{Object.defineProperty(e.location,"origin",{get:n,enumerable:!0})}catch(o){setInterval(function(){e.location.origin=n()},100)}}})(typeof global!="undefined"?global:typeof window!="undefined"?window:typeof self!="undefined"?self:Er)});var qr=Pt((Mt,Nr)=>{/*! + * clipboard.js v2.0.11 + * https://clipboardjs.com/ + * + * Licensed MIT © Zeno Rocha + */(function(t,r){typeof Mt=="object"&&typeof Nr=="object"?Nr.exports=r():typeof define=="function"&&define.amd?define([],r):typeof Mt=="object"?Mt.ClipboardJS=r():t.ClipboardJS=r()})(Mt,function(){return function(){var e={686:function(n,o,i){"use strict";i.d(o,{default:function(){return Ai}});var s=i(279),a=i.n(s),f=i(370),c=i.n(f),u=i(817),p=i.n(u);function m(j){try{return document.execCommand(j)}catch(T){return!1}}var d=function(T){var E=p()(T);return m("cut"),E},h=d;function v(j){var T=document.documentElement.getAttribute("dir")==="rtl",E=document.createElement("textarea");E.style.fontSize="12pt",E.style.border="0",E.style.padding="0",E.style.margin="0",E.style.position="absolute",E.style[T?"right":"left"]="-9999px";var H=window.pageYOffset||document.documentElement.scrollTop;return E.style.top="".concat(H,"px"),E.setAttribute("readonly",""),E.value=j,E}var Y=function(T,E){var H=v(T);E.container.appendChild(H);var I=p()(H);return m("copy"),H.remove(),I},B=function(T){var E=arguments.length>1&&arguments[1]!==void 0?arguments[1]:{container:document.body},H="";return typeof T=="string"?H=Y(T,E):T instanceof HTMLInputElement&&!["text","search","url","tel","password"].includes(T==null?void 0:T.type)?H=Y(T.value,E):(H=p()(T),m("copy")),H},N=B;function O(j){"@babel/helpers - typeof";return typeof Symbol=="function"&&typeof Symbol.iterator=="symbol"?O=function(E){return typeof E}:O=function(E){return E&&typeof Symbol=="function"&&E.constructor===Symbol&&E!==Symbol.prototype?"symbol":typeof E},O(j)}var Qe=function(){var T=arguments.length>0&&arguments[0]!==void 0?arguments[0]:{},E=T.action,H=E===void 0?"copy":E,I=T.container,q=T.target,Me=T.text;if(H!=="copy"&&H!=="cut")throw new Error('Invalid "action" value, use either "copy" or "cut"');if(q!==void 0)if(q&&O(q)==="object"&&q.nodeType===1){if(H==="copy"&&q.hasAttribute("disabled"))throw new Error('Invalid "target" attribute. Please use "readonly" instead of "disabled" attribute');if(H==="cut"&&(q.hasAttribute("readonly")||q.hasAttribute("disabled")))throw new Error(`Invalid "target" attribute. You can't cut text from elements with "readonly" or "disabled" attributes`)}else throw new Error('Invalid "target" value, use a valid Element');if(Me)return N(Me,{container:I});if(q)return H==="cut"?h(q):N(q,{container:I})},De=Qe;function $e(j){"@babel/helpers - typeof";return typeof Symbol=="function"&&typeof Symbol.iterator=="symbol"?$e=function(E){return typeof E}:$e=function(E){return E&&typeof Symbol=="function"&&E.constructor===Symbol&&E!==Symbol.prototype?"symbol":typeof E},$e(j)}function Ei(j,T){if(!(j instanceof T))throw new TypeError("Cannot call a class as a function")}function tn(j,T){for(var E=0;E0&&arguments[0]!==void 0?arguments[0]:{};this.action=typeof I.action=="function"?I.action:this.defaultAction,this.target=typeof I.target=="function"?I.target:this.defaultTarget,this.text=typeof I.text=="function"?I.text:this.defaultText,this.container=$e(I.container)==="object"?I.container:document.body}},{key:"listenClick",value:function(I){var q=this;this.listener=c()(I,"click",function(Me){return q.onClick(Me)})}},{key:"onClick",value:function(I){var q=I.delegateTarget||I.currentTarget,Me=this.action(q)||"copy",kt=De({action:Me,container:this.container,target:this.target(q),text:this.text(q)});this.emit(kt?"success":"error",{action:Me,text:kt,trigger:q,clearSelection:function(){q&&q.focus(),window.getSelection().removeAllRanges()}})}},{key:"defaultAction",value:function(I){return vr("action",I)}},{key:"defaultTarget",value:function(I){var q=vr("target",I);if(q)return document.querySelector(q)}},{key:"defaultText",value:function(I){return vr("text",I)}},{key:"destroy",value:function(){this.listener.destroy()}}],[{key:"copy",value:function(I){var q=arguments.length>1&&arguments[1]!==void 0?arguments[1]:{container:document.body};return N(I,q)}},{key:"cut",value:function(I){return h(I)}},{key:"isSupported",value:function(){var I=arguments.length>0&&arguments[0]!==void 0?arguments[0]:["copy","cut"],q=typeof I=="string"?[I]:I,Me=!!document.queryCommandSupported;return q.forEach(function(kt){Me=Me&&!!document.queryCommandSupported(kt)}),Me}}]),E}(a()),Ai=Li},828:function(n){var o=9;if(typeof Element!="undefined"&&!Element.prototype.matches){var i=Element.prototype;i.matches=i.matchesSelector||i.mozMatchesSelector||i.msMatchesSelector||i.oMatchesSelector||i.webkitMatchesSelector}function s(a,f){for(;a&&a.nodeType!==o;){if(typeof a.matches=="function"&&a.matches(f))return a;a=a.parentNode}}n.exports=s},438:function(n,o,i){var s=i(828);function a(u,p,m,d,h){var v=c.apply(this,arguments);return u.addEventListener(m,v,h),{destroy:function(){u.removeEventListener(m,v,h)}}}function f(u,p,m,d,h){return typeof u.addEventListener=="function"?a.apply(null,arguments):typeof m=="function"?a.bind(null,document).apply(null,arguments):(typeof u=="string"&&(u=document.querySelectorAll(u)),Array.prototype.map.call(u,function(v){return a(v,p,m,d,h)}))}function c(u,p,m,d){return function(h){h.delegateTarget=s(h.target,p),h.delegateTarget&&d.call(u,h)}}n.exports=f},879:function(n,o){o.node=function(i){return i!==void 0&&i instanceof HTMLElement&&i.nodeType===1},o.nodeList=function(i){var s=Object.prototype.toString.call(i);return i!==void 0&&(s==="[object NodeList]"||s==="[object HTMLCollection]")&&"length"in i&&(i.length===0||o.node(i[0]))},o.string=function(i){return typeof i=="string"||i instanceof String},o.fn=function(i){var s=Object.prototype.toString.call(i);return s==="[object Function]"}},370:function(n,o,i){var s=i(879),a=i(438);function f(m,d,h){if(!m&&!d&&!h)throw new Error("Missing required arguments");if(!s.string(d))throw new TypeError("Second argument must be a String");if(!s.fn(h))throw new TypeError("Third argument must be a Function");if(s.node(m))return c(m,d,h);if(s.nodeList(m))return u(m,d,h);if(s.string(m))return p(m,d,h);throw new TypeError("First argument must be a String, HTMLElement, HTMLCollection, or NodeList")}function c(m,d,h){return m.addEventListener(d,h),{destroy:function(){m.removeEventListener(d,h)}}}function u(m,d,h){return Array.prototype.forEach.call(m,function(v){v.addEventListener(d,h)}),{destroy:function(){Array.prototype.forEach.call(m,function(v){v.removeEventListener(d,h)})}}}function p(m,d,h){return a(document.body,m,d,h)}n.exports=f},817:function(n){function o(i){var s;if(i.nodeName==="SELECT")i.focus(),s=i.value;else if(i.nodeName==="INPUT"||i.nodeName==="TEXTAREA"){var a=i.hasAttribute("readonly");a||i.setAttribute("readonly",""),i.select(),i.setSelectionRange(0,i.value.length),a||i.removeAttribute("readonly"),s=i.value}else{i.hasAttribute("contenteditable")&&i.focus();var f=window.getSelection(),c=document.createRange();c.selectNodeContents(i),f.removeAllRanges(),f.addRange(c),s=f.toString()}return s}n.exports=o},279:function(n){function o(){}o.prototype={on:function(i,s,a){var f=this.e||(this.e={});return(f[i]||(f[i]=[])).push({fn:s,ctx:a}),this},once:function(i,s,a){var f=this;function c(){f.off(i,c),s.apply(a,arguments)}return c._=s,this.on(i,c,a)},emit:function(i){var s=[].slice.call(arguments,1),a=((this.e||(this.e={}))[i]||[]).slice(),f=0,c=a.length;for(f;f{"use strict";/*! + * escape-html + * Copyright(c) 2012-2013 TJ Holowaychuk + * Copyright(c) 2015 Andreas Lubbe + * Copyright(c) 2015 Tiancheng "Timothy" Gu + * MIT Licensed + */var rs=/["'&<>]/;Yo.exports=ns;function ns(e){var t=""+e,r=rs.exec(t);if(!r)return t;var n,o="",i=0,s=0;for(i=r.index;i0&&i[i.length-1])&&(c[0]===6||c[0]===2)){r=0;continue}if(c[0]===3&&(!i||c[1]>i[0]&&c[1]=e.length&&(e=void 0),{value:e&&e[n++],done:!e}}};throw new TypeError(t?"Object is not iterable.":"Symbol.iterator is not defined.")}function W(e,t){var r=typeof Symbol=="function"&&e[Symbol.iterator];if(!r)return e;var n=r.call(e),o,i=[],s;try{for(;(t===void 0||t-- >0)&&!(o=n.next()).done;)i.push(o.value)}catch(a){s={error:a}}finally{try{o&&!o.done&&(r=n.return)&&r.call(n)}finally{if(s)throw s.error}}return i}function D(e,t,r){if(r||arguments.length===2)for(var n=0,o=t.length,i;n1||a(m,d)})})}function a(m,d){try{f(n[m](d))}catch(h){p(i[0][3],h)}}function f(m){m.value instanceof et?Promise.resolve(m.value.v).then(c,u):p(i[0][2],m)}function c(m){a("next",m)}function u(m){a("throw",m)}function p(m,d){m(d),i.shift(),i.length&&a(i[0][0],i[0][1])}}function pn(e){if(!Symbol.asyncIterator)throw new TypeError("Symbol.asyncIterator is not defined.");var t=e[Symbol.asyncIterator],r;return t?t.call(e):(e=typeof Ee=="function"?Ee(e):e[Symbol.iterator](),r={},n("next"),n("throw"),n("return"),r[Symbol.asyncIterator]=function(){return this},r);function n(i){r[i]=e[i]&&function(s){return new Promise(function(a,f){s=e[i](s),o(a,f,s.done,s.value)})}}function o(i,s,a,f){Promise.resolve(f).then(function(c){i({value:c,done:a})},s)}}function C(e){return typeof e=="function"}function at(e){var t=function(n){Error.call(n),n.stack=new Error().stack},r=e(t);return r.prototype=Object.create(Error.prototype),r.prototype.constructor=r,r}var It=at(function(e){return function(r){e(this),this.message=r?r.length+` errors occurred during unsubscription: +`+r.map(function(n,o){return o+1+") "+n.toString()}).join(` + `):"",this.name="UnsubscriptionError",this.errors=r}});function Ve(e,t){if(e){var r=e.indexOf(t);0<=r&&e.splice(r,1)}}var Ie=function(){function e(t){this.initialTeardown=t,this.closed=!1,this._parentage=null,this._finalizers=null}return e.prototype.unsubscribe=function(){var t,r,n,o,i;if(!this.closed){this.closed=!0;var s=this._parentage;if(s)if(this._parentage=null,Array.isArray(s))try{for(var a=Ee(s),f=a.next();!f.done;f=a.next()){var c=f.value;c.remove(this)}}catch(v){t={error:v}}finally{try{f&&!f.done&&(r=a.return)&&r.call(a)}finally{if(t)throw t.error}}else s.remove(this);var u=this.initialTeardown;if(C(u))try{u()}catch(v){i=v instanceof It?v.errors:[v]}var p=this._finalizers;if(p){this._finalizers=null;try{for(var m=Ee(p),d=m.next();!d.done;d=m.next()){var h=d.value;try{ln(h)}catch(v){i=i!=null?i:[],v instanceof It?i=D(D([],W(i)),W(v.errors)):i.push(v)}}}catch(v){n={error:v}}finally{try{d&&!d.done&&(o=m.return)&&o.call(m)}finally{if(n)throw n.error}}}if(i)throw new It(i)}},e.prototype.add=function(t){var r;if(t&&t!==this)if(this.closed)ln(t);else{if(t instanceof e){if(t.closed||t._hasParent(this))return;t._addParent(this)}(this._finalizers=(r=this._finalizers)!==null&&r!==void 0?r:[]).push(t)}},e.prototype._hasParent=function(t){var r=this._parentage;return r===t||Array.isArray(r)&&r.includes(t)},e.prototype._addParent=function(t){var r=this._parentage;this._parentage=Array.isArray(r)?(r.push(t),r):r?[r,t]:t},e.prototype._removeParent=function(t){var r=this._parentage;r===t?this._parentage=null:Array.isArray(r)&&Ve(r,t)},e.prototype.remove=function(t){var r=this._finalizers;r&&Ve(r,t),t instanceof e&&t._removeParent(this)},e.EMPTY=function(){var t=new e;return t.closed=!0,t}(),e}();var Sr=Ie.EMPTY;function jt(e){return e instanceof Ie||e&&"closed"in e&&C(e.remove)&&C(e.add)&&C(e.unsubscribe)}function ln(e){C(e)?e():e.unsubscribe()}var Le={onUnhandledError:null,onStoppedNotification:null,Promise:void 0,useDeprecatedSynchronousErrorHandling:!1,useDeprecatedNextContext:!1};var st={setTimeout:function(e,t){for(var r=[],n=2;n0},enumerable:!1,configurable:!0}),t.prototype._trySubscribe=function(r){return this._throwIfClosed(),e.prototype._trySubscribe.call(this,r)},t.prototype._subscribe=function(r){return this._throwIfClosed(),this._checkFinalizedStatuses(r),this._innerSubscribe(r)},t.prototype._innerSubscribe=function(r){var n=this,o=this,i=o.hasError,s=o.isStopped,a=o.observers;return i||s?Sr:(this.currentObservers=null,a.push(r),new Ie(function(){n.currentObservers=null,Ve(a,r)}))},t.prototype._checkFinalizedStatuses=function(r){var n=this,o=n.hasError,i=n.thrownError,s=n.isStopped;o?r.error(i):s&&r.complete()},t.prototype.asObservable=function(){var r=new F;return r.source=this,r},t.create=function(r,n){return new xn(r,n)},t}(F);var xn=function(e){ie(t,e);function t(r,n){var o=e.call(this)||this;return o.destination=r,o.source=n,o}return t.prototype.next=function(r){var n,o;(o=(n=this.destination)===null||n===void 0?void 0:n.next)===null||o===void 0||o.call(n,r)},t.prototype.error=function(r){var n,o;(o=(n=this.destination)===null||n===void 0?void 0:n.error)===null||o===void 0||o.call(n,r)},t.prototype.complete=function(){var r,n;(n=(r=this.destination)===null||r===void 0?void 0:r.complete)===null||n===void 0||n.call(r)},t.prototype._subscribe=function(r){var n,o;return(o=(n=this.source)===null||n===void 0?void 0:n.subscribe(r))!==null&&o!==void 0?o:Sr},t}(x);var Et={now:function(){return(Et.delegate||Date).now()},delegate:void 0};var wt=function(e){ie(t,e);function t(r,n,o){r===void 0&&(r=1/0),n===void 0&&(n=1/0),o===void 0&&(o=Et);var i=e.call(this)||this;return i._bufferSize=r,i._windowTime=n,i._timestampProvider=o,i._buffer=[],i._infiniteTimeWindow=!0,i._infiniteTimeWindow=n===1/0,i._bufferSize=Math.max(1,r),i._windowTime=Math.max(1,n),i}return t.prototype.next=function(r){var n=this,o=n.isStopped,i=n._buffer,s=n._infiniteTimeWindow,a=n._timestampProvider,f=n._windowTime;o||(i.push(r),!s&&i.push(a.now()+f)),this._trimBuffer(),e.prototype.next.call(this,r)},t.prototype._subscribe=function(r){this._throwIfClosed(),this._trimBuffer();for(var n=this._innerSubscribe(r),o=this,i=o._infiniteTimeWindow,s=o._buffer,a=s.slice(),f=0;f0?e.prototype.requestAsyncId.call(this,r,n,o):(r.actions.push(this),r._scheduled||(r._scheduled=ut.requestAnimationFrame(function(){return r.flush(void 0)})))},t.prototype.recycleAsyncId=function(r,n,o){var i;if(o===void 0&&(o=0),o!=null?o>0:this.delay>0)return e.prototype.recycleAsyncId.call(this,r,n,o);var s=r.actions;n!=null&&((i=s[s.length-1])===null||i===void 0?void 0:i.id)!==n&&(ut.cancelAnimationFrame(n),r._scheduled=void 0)},t}(Wt);var Sn=function(e){ie(t,e);function t(){return e!==null&&e.apply(this,arguments)||this}return t.prototype.flush=function(r){this._active=!0;var n=this._scheduled;this._scheduled=void 0;var o=this.actions,i;r=r||o.shift();do if(i=r.execute(r.state,r.delay))break;while((r=o[0])&&r.id===n&&o.shift());if(this._active=!1,i){for(;(r=o[0])&&r.id===n&&o.shift();)r.unsubscribe();throw i}},t}(Dt);var Oe=new Sn(wn);var M=new F(function(e){return e.complete()});function Vt(e){return e&&C(e.schedule)}function Cr(e){return e[e.length-1]}function Ye(e){return C(Cr(e))?e.pop():void 0}function Te(e){return Vt(Cr(e))?e.pop():void 0}function zt(e,t){return typeof Cr(e)=="number"?e.pop():t}var pt=function(e){return e&&typeof e.length=="number"&&typeof e!="function"};function Nt(e){return C(e==null?void 0:e.then)}function qt(e){return C(e[ft])}function Kt(e){return Symbol.asyncIterator&&C(e==null?void 0:e[Symbol.asyncIterator])}function Qt(e){return new TypeError("You provided "+(e!==null&&typeof e=="object"?"an invalid object":"'"+e+"'")+" where a stream was expected. You can provide an Observable, Promise, ReadableStream, Array, AsyncIterable, or Iterable.")}function zi(){return typeof Symbol!="function"||!Symbol.iterator?"@@iterator":Symbol.iterator}var Yt=zi();function Gt(e){return C(e==null?void 0:e[Yt])}function Bt(e){return un(this,arguments,function(){var r,n,o,i;return $t(this,function(s){switch(s.label){case 0:r=e.getReader(),s.label=1;case 1:s.trys.push([1,,9,10]),s.label=2;case 2:return[4,et(r.read())];case 3:return n=s.sent(),o=n.value,i=n.done,i?[4,et(void 0)]:[3,5];case 4:return[2,s.sent()];case 5:return[4,et(o)];case 6:return[4,s.sent()];case 7:return s.sent(),[3,2];case 8:return[3,10];case 9:return r.releaseLock(),[7];case 10:return[2]}})})}function Jt(e){return C(e==null?void 0:e.getReader)}function U(e){if(e instanceof F)return e;if(e!=null){if(qt(e))return Ni(e);if(pt(e))return qi(e);if(Nt(e))return Ki(e);if(Kt(e))return On(e);if(Gt(e))return Qi(e);if(Jt(e))return Yi(e)}throw Qt(e)}function Ni(e){return new F(function(t){var r=e[ft]();if(C(r.subscribe))return r.subscribe(t);throw new TypeError("Provided object does not correctly implement Symbol.observable")})}function qi(e){return new F(function(t){for(var r=0;r=2;return function(n){return n.pipe(e?A(function(o,i){return e(o,i,n)}):de,ge(1),r?He(t):Dn(function(){return new Zt}))}}function Vn(){for(var e=[],t=0;t=2,!0))}function pe(e){e===void 0&&(e={});var t=e.connector,r=t===void 0?function(){return new x}:t,n=e.resetOnError,o=n===void 0?!0:n,i=e.resetOnComplete,s=i===void 0?!0:i,a=e.resetOnRefCountZero,f=a===void 0?!0:a;return function(c){var u,p,m,d=0,h=!1,v=!1,Y=function(){p==null||p.unsubscribe(),p=void 0},B=function(){Y(),u=m=void 0,h=v=!1},N=function(){var O=u;B(),O==null||O.unsubscribe()};return y(function(O,Qe){d++,!v&&!h&&Y();var De=m=m!=null?m:r();Qe.add(function(){d--,d===0&&!v&&!h&&(p=$r(N,f))}),De.subscribe(Qe),!u&&d>0&&(u=new rt({next:function($e){return De.next($e)},error:function($e){v=!0,Y(),p=$r(B,o,$e),De.error($e)},complete:function(){h=!0,Y(),p=$r(B,s),De.complete()}}),U(O).subscribe(u))})(c)}}function $r(e,t){for(var r=[],n=2;ne.next(document)),e}function K(e,t=document){return Array.from(t.querySelectorAll(e))}function z(e,t=document){let r=ce(e,t);if(typeof r=="undefined")throw new ReferenceError(`Missing element: expected "${e}" to be present`);return r}function ce(e,t=document){return t.querySelector(e)||void 0}function _e(){return document.activeElement instanceof HTMLElement&&document.activeElement||void 0}function tr(e){return L(b(document.body,"focusin"),b(document.body,"focusout")).pipe(ke(1),l(()=>{let t=_e();return typeof t!="undefined"?e.contains(t):!1}),V(e===_e()),J())}function Xe(e){return{x:e.offsetLeft,y:e.offsetTop}}function Kn(e){return L(b(window,"load"),b(window,"resize")).pipe(Ce(0,Oe),l(()=>Xe(e)),V(Xe(e)))}function rr(e){return{x:e.scrollLeft,y:e.scrollTop}}function dt(e){return L(b(e,"scroll"),b(window,"resize")).pipe(Ce(0,Oe),l(()=>rr(e)),V(rr(e)))}var Yn=function(){if(typeof Map!="undefined")return Map;function e(t,r){var n=-1;return t.some(function(o,i){return o[0]===r?(n=i,!0):!1}),n}return function(){function t(){this.__entries__=[]}return Object.defineProperty(t.prototype,"size",{get:function(){return this.__entries__.length},enumerable:!0,configurable:!0}),t.prototype.get=function(r){var n=e(this.__entries__,r),o=this.__entries__[n];return o&&o[1]},t.prototype.set=function(r,n){var o=e(this.__entries__,r);~o?this.__entries__[o][1]=n:this.__entries__.push([r,n])},t.prototype.delete=function(r){var n=this.__entries__,o=e(n,r);~o&&n.splice(o,1)},t.prototype.has=function(r){return!!~e(this.__entries__,r)},t.prototype.clear=function(){this.__entries__.splice(0)},t.prototype.forEach=function(r,n){n===void 0&&(n=null);for(var o=0,i=this.__entries__;o0},e.prototype.connect_=function(){!Wr||this.connected_||(document.addEventListener("transitionend",this.onTransitionEnd_),window.addEventListener("resize",this.refresh),va?(this.mutationsObserver_=new MutationObserver(this.refresh),this.mutationsObserver_.observe(document,{attributes:!0,childList:!0,characterData:!0,subtree:!0})):(document.addEventListener("DOMSubtreeModified",this.refresh),this.mutationEventsAdded_=!0),this.connected_=!0)},e.prototype.disconnect_=function(){!Wr||!this.connected_||(document.removeEventListener("transitionend",this.onTransitionEnd_),window.removeEventListener("resize",this.refresh),this.mutationsObserver_&&this.mutationsObserver_.disconnect(),this.mutationEventsAdded_&&document.removeEventListener("DOMSubtreeModified",this.refresh),this.mutationsObserver_=null,this.mutationEventsAdded_=!1,this.connected_=!1)},e.prototype.onTransitionEnd_=function(t){var r=t.propertyName,n=r===void 0?"":r,o=ba.some(function(i){return!!~n.indexOf(i)});o&&this.refresh()},e.getInstance=function(){return this.instance_||(this.instance_=new e),this.instance_},e.instance_=null,e}(),Gn=function(e,t){for(var r=0,n=Object.keys(t);r0},e}(),Jn=typeof WeakMap!="undefined"?new WeakMap:new Yn,Xn=function(){function e(t){if(!(this instanceof e))throw new TypeError("Cannot call a class as a function.");if(!arguments.length)throw new TypeError("1 argument required, but only 0 present.");var r=ga.getInstance(),n=new La(t,r,this);Jn.set(this,n)}return e}();["observe","unobserve","disconnect"].forEach(function(e){Xn.prototype[e]=function(){var t;return(t=Jn.get(this))[e].apply(t,arguments)}});var Aa=function(){return typeof nr.ResizeObserver!="undefined"?nr.ResizeObserver:Xn}(),Zn=Aa;var eo=new x,Ca=$(()=>k(new Zn(e=>{for(let t of e)eo.next(t)}))).pipe(g(e=>L(ze,k(e)).pipe(R(()=>e.disconnect()))),X(1));function he(e){return{width:e.offsetWidth,height:e.offsetHeight}}function ye(e){return Ca.pipe(S(t=>t.observe(e)),g(t=>eo.pipe(A(({target:r})=>r===e),R(()=>t.unobserve(e)),l(()=>he(e)))),V(he(e)))}function bt(e){return{width:e.scrollWidth,height:e.scrollHeight}}function ar(e){let t=e.parentElement;for(;t&&(e.scrollWidth<=t.scrollWidth&&e.scrollHeight<=t.scrollHeight);)t=(e=t).parentElement;return t?e:void 0}var to=new x,Ra=$(()=>k(new IntersectionObserver(e=>{for(let t of e)to.next(t)},{threshold:0}))).pipe(g(e=>L(ze,k(e)).pipe(R(()=>e.disconnect()))),X(1));function sr(e){return Ra.pipe(S(t=>t.observe(e)),g(t=>to.pipe(A(({target:r})=>r===e),R(()=>t.unobserve(e)),l(({isIntersecting:r})=>r))))}function ro(e,t=16){return dt(e).pipe(l(({y:r})=>{let n=he(e),o=bt(e);return r>=o.height-n.height-t}),J())}var cr={drawer:z("[data-md-toggle=drawer]"),search:z("[data-md-toggle=search]")};function no(e){return cr[e].checked}function Ke(e,t){cr[e].checked!==t&&cr[e].click()}function Ue(e){let t=cr[e];return b(t,"change").pipe(l(()=>t.checked),V(t.checked))}function ka(e,t){switch(e.constructor){case HTMLInputElement:return e.type==="radio"?/^Arrow/.test(t):!0;case HTMLSelectElement:case HTMLTextAreaElement:return!0;default:return e.isContentEditable}}function Ha(){return L(b(window,"compositionstart").pipe(l(()=>!0)),b(window,"compositionend").pipe(l(()=>!1))).pipe(V(!1))}function oo(){let e=b(window,"keydown").pipe(A(t=>!(t.metaKey||t.ctrlKey)),l(t=>({mode:no("search")?"search":"global",type:t.key,claim(){t.preventDefault(),t.stopPropagation()}})),A(({mode:t,type:r})=>{if(t==="global"){let n=_e();if(typeof n!="undefined")return!ka(n,r)}return!0}),pe());return Ha().pipe(g(t=>t?M:e))}function le(){return new URL(location.href)}function ot(e){location.href=e.href}function io(){return new x}function ao(e,t){if(typeof t=="string"||typeof t=="number")e.innerHTML+=t.toString();else if(t instanceof Node)e.appendChild(t);else if(Array.isArray(t))for(let r of t)ao(e,r)}function _(e,t,...r){let n=document.createElement(e);if(t)for(let o of Object.keys(t))typeof t[o]!="undefined"&&(typeof t[o]!="boolean"?n.setAttribute(o,t[o]):n.setAttribute(o,""));for(let o of r)ao(n,o);return n}function fr(e){if(e>999){let t=+((e-950)%1e3>99);return`${((e+1e-6)/1e3).toFixed(t)}k`}else return e.toString()}function so(){return location.hash.substring(1)}function Dr(e){let t=_("a",{href:e});t.addEventListener("click",r=>r.stopPropagation()),t.click()}function Pa(e){return L(b(window,"hashchange"),e).pipe(l(so),V(so()),A(t=>t.length>0),X(1))}function co(e){return Pa(e).pipe(l(t=>ce(`[id="${t}"]`)),A(t=>typeof t!="undefined"))}function Vr(e){let t=matchMedia(e);return er(r=>t.addListener(()=>r(t.matches))).pipe(V(t.matches))}function fo(){let e=matchMedia("print");return L(b(window,"beforeprint").pipe(l(()=>!0)),b(window,"afterprint").pipe(l(()=>!1))).pipe(V(e.matches))}function zr(e,t){return e.pipe(g(r=>r?t():M))}function ur(e,t={credentials:"same-origin"}){return ue(fetch(`${e}`,t)).pipe(fe(()=>M),g(r=>r.status!==200?Ot(()=>new Error(r.statusText)):k(r)))}function We(e,t){return ur(e,t).pipe(g(r=>r.json()),X(1))}function uo(e,t){let r=new DOMParser;return ur(e,t).pipe(g(n=>n.text()),l(n=>r.parseFromString(n,"text/xml")),X(1))}function pr(e){let t=_("script",{src:e});return $(()=>(document.head.appendChild(t),L(b(t,"load"),b(t,"error").pipe(g(()=>Ot(()=>new ReferenceError(`Invalid script: ${e}`))))).pipe(l(()=>{}),R(()=>document.head.removeChild(t)),ge(1))))}function po(){return{x:Math.max(0,scrollX),y:Math.max(0,scrollY)}}function lo(){return L(b(window,"scroll",{passive:!0}),b(window,"resize",{passive:!0})).pipe(l(po),V(po()))}function mo(){return{width:innerWidth,height:innerHeight}}function ho(){return b(window,"resize",{passive:!0}).pipe(l(mo),V(mo()))}function bo(){return G([lo(),ho()]).pipe(l(([e,t])=>({offset:e,size:t})),X(1))}function lr(e,{viewport$:t,header$:r}){let n=t.pipe(ee("size")),o=G([n,r]).pipe(l(()=>Xe(e)));return G([r,t,o]).pipe(l(([{height:i},{offset:s,size:a},{x:f,y:c}])=>({offset:{x:s.x-f,y:s.y-c+i},size:a})))}(()=>{function e(n,o){parent.postMessage(n,o||"*")}function t(...n){return n.reduce((o,i)=>o.then(()=>new Promise(s=>{let a=document.createElement("script");a.src=i,a.onload=s,document.body.appendChild(a)})),Promise.resolve())}var r=class extends EventTarget{constructor(n){super(),this.url=n,this.m=i=>{i.source===this.w&&(this.dispatchEvent(new MessageEvent("message",{data:i.data})),this.onmessage&&this.onmessage(i))},this.e=(i,s,a,f,c)=>{if(s===`${this.url}`){let u=new ErrorEvent("error",{message:i,filename:s,lineno:a,colno:f,error:c});this.dispatchEvent(u),this.onerror&&this.onerror(u)}};let o=document.createElement("iframe");o.hidden=!0,document.body.appendChild(this.iframe=o),this.w.document.open(),this.w.document.write(` + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Overview

+

logo

+

LLamaSharp is the C#/.NET binding of llama.cpp. It provides APIs to inference the LLaMa Models and deploy it on native environment or Web. It could help C# developers to deploy the LLM (Large Language Model) locally and integrate with C# apps.

+

Main features

+
    +
  • Model inference
  • +
  • Model quantization
  • +
  • Generating embeddings
  • +
  • Grammar parse
  • +
  • Interactive/Instruct/Stateless executor mode
  • +
  • Chat session APIs
  • +
  • Save/load the state
  • +
  • Integration with other applications like BotSharp and semantic-kernel
  • +
+

Essential insights for novice learners

+

If you are new to LLM, here're some tips for you to help you to get start with LLamaSharp. If you are experienced in this field, we'd still recommend you to take a few minutes to read it because some things perform differently compared to cpp/python.

+
    +
  1. The main ability of LLamaSharp is to provide an efficient way to run inference of LLM (Large Language Model) locally (and fine-tune model in the future). The model weights, however, need to be downloaded from other resources such as huggingface.
  2. +
  3. Since LLamaSharp supports multiple platforms, The nuget package is split into LLamaSharp and LLama.Backend. After installing LLamaSharp, please install one of LLama.Backend.Cpu, LLama.Backend.Cuda11 or LLama.Backend.Cuda12. If you use the source code, dynamic libraries can be found in LLama/Runtimes. Rename the one you want to use to libllama.dll.
  4. +
  5. LLaMa originally refers to the weights released by Meta (Facebook Research). After that, many models are fine-tuned based on it, such as Vicuna, GPT4All, and Pyglion. Though all of these models are supported by LLamaSharp, some steps are necessary with different file formats. There're mainly three kinds of files, which are .pth, .bin (ggml), .bin (quantized). If you have the .bin (quantized) file, it could be used directly by LLamaSharp. If you have the .bin (ggml) file, you could use it directly but get higher inference speed after the quantization. If you have the .pth file, you need to follow the instructions in llama.cpp to convert it to .bin (ggml) file at first.
  6. +
  7. LLamaSharp supports GPU acceleration, but it requires cuda installation. Please install cuda 11 or cuda 12 on your system before using LLamaSharp to enable GPU. If you have another cuda version, you could compile llama.cpp from source to get the dll. For building from source, please refer to issue #5.
  8. +
+

Welcome to join the development!

+

Community effort is always one of the most important things in open-source projects. Any contribution in any way is welcomed here. For example, the following things mean a lot for LLamaSharp:

+
    +
  1. Open an issue when you find something wrong.
  2. +
  3. Open an PR if you've fixed something. Even if just correcting a typo, it also makes great sense.
  4. +
  5. Help to optimize the documentation.
  6. +
  7. Write an example or blog about how to integrate LLamaSharp with your APPs.
  8. +
  9. Ask for a missed feature and discuss with other developers.
  10. +
+

If you'd like to get deeply involved in development, please touch us in discord channel or send email to AsakusaRinne@gmail.com. :)

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/media/LLamaSharpLogo.png b/0.5/media/LLamaSharpLogo.png new file mode 100755 index 00000000..62df789a Binary files /dev/null and b/0.5/media/LLamaSharpLogo.png differ diff --git a/0.5/media/structure.jpg b/0.5/media/structure.jpg new file mode 100755 index 00000000..74173977 Binary files /dev/null and b/0.5/media/structure.jpg differ diff --git a/0.5/media/structure.vsdx b/0.5/media/structure.vsdx new file mode 100755 index 00000000..c36500eb Binary files /dev/null and b/0.5/media/structure.vsdx differ diff --git a/0.5/sciprts/map_xml_files_to_yml.py b/0.5/sciprts/map_xml_files_to_yml.py new file mode 100755 index 00000000..e51ff490 --- /dev/null +++ b/0.5/sciprts/map_xml_files_to_yml.py @@ -0,0 +1,16 @@ +import os + +def generate_string_list(folder_path, prefix): + file_names = os.listdir(folder_path) + string_list = [] + for file_name in file_names: + new_string = f"- {'.'.join(file_name.split('.')[:-1])}: {prefix}{file_name}" + string_list.append(new_string) + return string_list + +folder_path = "./docs/xmldocs" +prefix = "./xmldocs/" + +string_list = generate_string_list(folder_path, prefix) +result = '\n'.join(string_list) +print(result) diff --git a/0.5/search/search_index.json b/0.5/search/search_index.json new file mode 100755 index 00000000..668250e9 --- /dev/null +++ b/0.5/search/search_index.json @@ -0,0 +1 @@ +{"config":{"lang":["en"],"separator":"[\\s\\-]+","pipeline":["stopWordFilter"]},"docs":[{"location":"","title":"Overview","text":"

LLamaSharp is the C#/.NET binding of llama.cpp. It provides APIs to inference the LLaMa Models and deploy it on native environment or Web. It could help C# developers to deploy the LLM (Large Language Model) locally and integrate with C# apps.

"},{"location":"#main-features","title":"Main features","text":"
  • Model inference
  • Model quantization
  • Generating embeddings
  • Grammar parse
  • Interactive/Instruct/Stateless executor mode
  • Chat session APIs
  • Save/load the state
  • Integration with other applications like BotSharp and semantic-kernel
"},{"location":"#essential-insights-for-novice-learners","title":"Essential insights for novice learners","text":"

If you are new to LLM, here're some tips for you to help you to get start with LLamaSharp. If you are experienced in this field, we'd still recommend you to take a few minutes to read it because some things perform differently compared to cpp/python.

  1. The main ability of LLamaSharp is to provide an efficient way to run inference of LLM (Large Language Model) locally (and fine-tune model in the future). The model weights, however, need to be downloaded from other resources such as huggingface.
  2. Since LLamaSharp supports multiple platforms, The nuget package is split into LLamaSharp and LLama.Backend. After installing LLamaSharp, please install one of LLama.Backend.Cpu, LLama.Backend.Cuda11 or LLama.Backend.Cuda12. If you use the source code, dynamic libraries can be found in LLama/Runtimes. Rename the one you want to use to libllama.dll.
  3. LLaMa originally refers to the weights released by Meta (Facebook Research). After that, many models are fine-tuned based on it, such as Vicuna, GPT4All, and Pyglion. Though all of these models are supported by LLamaSharp, some steps are necessary with different file formats. There're mainly three kinds of files, which are .pth, .bin (ggml), .bin (quantized). If you have the .bin (quantized) file, it could be used directly by LLamaSharp. If you have the .bin (ggml) file, you could use it directly but get higher inference speed after the quantization. If you have the .pth file, you need to follow the instructions in llama.cpp to convert it to .bin (ggml) file at first.
  4. LLamaSharp supports GPU acceleration, but it requires cuda installation. Please install cuda 11 or cuda 12 on your system before using LLamaSharp to enable GPU. If you have another cuda version, you could compile llama.cpp from source to get the dll. For building from source, please refer to issue #5.
"},{"location":"#welcome-to-join-the-development","title":"Welcome to join the development!","text":"

Community effort is always one of the most important things in open-source projects. Any contribution in any way is welcomed here. For example, the following things mean a lot for LLamaSharp:

  1. Open an issue when you find something wrong.
  2. Open an PR if you've fixed something. Even if just correcting a typo, it also makes great sense.
  3. Help to optimize the documentation.
  4. Write an example or blog about how to integrate LLamaSharp with your APPs.
  5. Ask for a missed feature and discuss with other developers.

If you'd like to get deeply involved in development, please touch us in discord channel or send email to AsakusaRinne@gmail.com. :)

"},{"location":"Architecture/","title":"Architecture","text":""},{"location":"Architecture/#architecture-of-main-functions","title":"Architecture of main functions","text":"

The figure below shows the core framework structure, which is separated to four levels.

  • LLamaContext: The holder of a model which directly interact with native library and provide some basic APIs such as tokenization and embedding. Currently it includes three classes: LLamaContext, LLamaEmbedder and LLamaQuantizer.
  • LLamaExecutors: Executors which define the way to run the LLama model. It provides text-to-text APIs to make it easy to use. Currently we provide three kinds of executors: InteractiveExecutor, InstructuExecutor and StatelessExecutor.
  • ChatSession: A wrapping for InteractiveExecutor and LLamaContext, which supports interactive tasks and saving/re-loading sessions. It also provides a flexible way to customize the text process by IHistoryTransform, ITextTransform and ITextStreamTransform.
  • High-level Applications: Some applications that provides higher-level integration. For example, BotSharp provides integration for vector search, Chatbot UI and Web APIs. semantic-kernel provides various APIs for manipulations related with LLM. If you've made an integration, please tell us and add it to the doc!

"},{"location":"Architecture/#recommended-use","title":"Recommended Use","text":"

Since LLamaContext interact with native library, it's not recommended to use the methods of it directly unless you know what you are doing. So does the NativeApi, which is not included in the architecture figure above.

ChatSession is recommended to be used when you want to build an application similar to ChatGPT, or the ChatBot, because it works best with InteractiveExecutor. Though other executors are also allowed to passed as a parameter to initialize a ChatSession, it's not encouraged if you are new to LLamaSharp and LLM.

High-level applications, such as BotSharp, are supposed to be used when you concentrate on the part not related with LLM. For example, if you want to deploy a chat bot to help you remember your schedules, using BotSharp may be a good choice.

Note that the APIs of the high-level applications may not be stable now. Please take it into account when using them.

"},{"location":"ContributingGuide/","title":"LLamaSharp Contributing Guide","text":"

Hi, welcome to develop LLamaSharp with us together! We are always open for every contributor and any format of contributions! If you want to maintain this library actively together, please contact us to get the write access after some PRs. (Email: AsakusaRinne@gmail.com)

In this page, we'd like to introduce how to make contributions here easily. \ud83d\ude0a

"},{"location":"ContributingGuide/#compile-the-native-library-from-source","title":"Compile the native library from source","text":"

Firstly, please clone the llama.cpp repository and following the instructions in llama.cpp readme to configure your local environment.

If you want to support cublas in the compilation, please make sure that you've installed the cuda.

When building from source, please add -DBUILD_SHARED_LIBS=ON to the cmake instruction. For example, when building with cublas but without openblas, use the following instruction:

cmake .. -DLLAMA_CUBLAS=ON -DBUILD_SHARED_LIBS=ON\n

After running cmake --build . --config Release, you could find the llama.dll, llama.so or llama.dylib in your build directory. After pasting it to LLamaSharp/LLama/runtimes and renaming it to libllama.dll, libllama.so or libllama.dylib, you can use it as the native library in LLamaSharp.

"},{"location":"ContributingGuide/#add-a-new-feature-to-llamasharp","title":"Add a new feature to LLamaSharp","text":"

After refactoring the framework in v0.4.0, LLamaSharp will try to maintain the backward compatibility. However, in the following cases a breaking change will be required:

  1. Due to some break changes in llama.cpp, making a breaking change will help to maintain the good abstraction and friendly user APIs.
  2. A very important feature cannot be implemented unless refactoring some parts.
  3. After some discussions, an agreement was reached that making the break change is reasonable.

If a new feature could be added without introducing any break change, please open a PR rather than open an issue first. We will never refuse the PR but help to improve it, unless it's malicious.

When adding the feature, please take care of the namespace and the naming convention. For example, if you are adding an integration for WPF, please put the code under namespace LLama.WPF or LLama.Integration.WPF instead of putting it under the root namespace. The naming convention of LLamaSharp follows the pascal naming convention, but in some parts that are invisible to users, you can do whatever you want.

"},{"location":"ContributingGuide/#find-the-problem-and-fix-the-bug","title":"Find the problem and fix the BUG","text":"

If the issue is related to the LLM internal behaviour, such as endless generating the response, the best way to find the problem is to do comparison test between llama.cpp and LLamaSharp.

You could use exactly the same prompt, the same model and the same parameters to run the inference in llama.cpp and LLamaSharp respectively to see if it's really a problem caused by the implementation in LLamaSharp.

If the experiment showed that it worked well in llama.cpp but didn't in LLamaSharp, a search for the problem could be started. While the reason of the problem could be various, the best way I think is to add log-print in the code of llama.cpp and use it in LLamaSharp after compilation. Thus, when running LLamaSharp, you could see what happened in the native library.

After finding out the reason, a painful but happy process comes. When working on the BUG fix, there's only one rule to follow, that is keeping the examples working well. If the modification fixed the BUG but impact on other functions, it would not be a good fix.

During the BUG fix process, please don't hesitate to discuss together when you stuck on something.

"},{"location":"ContributingGuide/#add-integrations","title":"Add integrations","text":"

All kinds of integration are welcomed here! Currently the following integrations are under work or on our schedule:

  1. BotSharp
  2. semantic-kernel
  3. Unity

Besides, for some other integrations, like ASP.NET core, SQL, Blazor and so on, we'll appreciate it if you could help with that. If the time is limited for you, providing an example for it also means a lot!

"},{"location":"ContributingGuide/#add-examples","title":"Add examples","text":"

There're mainly two ways to add an example:

  1. Add the example to LLama.Examples of the repository.
  2. Put the example in another repository and add the link to the readme or docs of LLamaSharp.
"},{"location":"ContributingGuide/#add-documents","title":"Add documents","text":"

LLamaSharp uses mkdocs to build the documentation, please follow the tutorial of mkdocs to add or modify documents in LLamaSharp.

"},{"location":"GetStarted/","title":"Get Started","text":""},{"location":"GetStarted/#install-packages","title":"Install packages","text":"

Firstly, search LLamaSharp in nuget package manager and install it.

PM> Install-Package LLamaSharp\n

Then, search and install one of the following backends:

LLamaSharp.Backend.Cpu\nLLamaSharp.Backend.Cuda11\nLLamaSharp.Backend.Cuda12\n

Here's the mapping of them and corresponding model samples provided by LLamaSharp. If you're not sure which model is available for a version, please try our sample model.

LLamaSharp.Backend LLamaSharp Verified Model Resources llama.cpp commit id - v0.2.0 This version is not recommended to use. - - v0.2.1 WizardLM, Vicuna (filenames with \"old\") - v0.2.2 v0.2.2, v0.2.3 WizardLM, Vicuna (filenames without \"old\") 63d2046 v0.3.0 v0.3.0 LLamaSharpSamples v0.3.0, WizardLM 7e4ea5b"},{"location":"GetStarted/#download-a-model","title":"Download a model","text":"

One of the following models could be okay:

  • LLaMA \ud83e\udd99
  • Alpaca
  • GPT4All
  • Chinese LLaMA / Alpaca
  • Vigogne (French)
  • Vicuna
  • Koala
  • OpenBuddy \ud83d\udc36 (Multilingual)
  • Pygmalion 7B / Metharme 7B
  • WizardLM

Note that because llama.cpp is under fast development now and often introduce break changes, some model weights on huggingface which works under a version may be invalid with another version. If it's your first time to configure LLamaSharp, we'd like to suggest for using verified model weights in the table above.

"},{"location":"GetStarted/#run-the-program","title":"Run the program","text":"

Please create a console program with dotnet runtime >= netstandard 2.0 (>= net6.0 is more recommended). Then, paste the following code to program.cs;

using LLama.Common;\nusing LLama;\n\nstring modelPath = \"<Your model path>\" // change it to your own model path\nvar prompt = \"Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.\\r\\n\\r\\nUser: Hello, Bob.\\r\\nBob: Hello. How may I help you today?\\r\\nUser: Please tell me the largest city in Europe.\\r\\nBob: Sure. The largest city in Europe is Moscow, the capital of Russia.\\r\\nUser:\"; // use the \"chat-with-bob\" prompt here.\n\n// Load model\nvar parameters = new ModelParams(modelPath)\n{\n    ContextSize = 1024\n};\nusing var model = LLamaWeights.LoadFromFile(parameters);\n\n// Initialize a chat session\nusing var context = model.CreateContext(parameters);\nvar ex = new InteractiveExecutor(context);\nChatSession session = new ChatSession(ex);\n\n// show the prompt\nConsole.WriteLine();\nConsole.Write(prompt);\n\n// run the inference in a loop to chat with LLM\nwhile (true)\n{\n    foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } }))\n    {\n        Console.Write(text);\n    }\n\n    Console.ForegroundColor = ConsoleColor.Green;\n    prompt = Console.ReadLine();\n    Console.ForegroundColor = ConsoleColor.White;\n}\n

After starting it, you'll see the following outputs.

Please input your model path: D:\\development\\llama\\weights\\wizard-vicuna-13B.ggmlv3.q4_1.bin\nllama.cpp: loading model from D:\\development\\llama\\weights\\wizard-vicuna-13B.ggmlv3.q4_1.bin\nllama_model_load_internal: format     = ggjt v3 (latest)\nllama_model_load_internal: n_vocab    = 32000\nllama_model_load_internal: n_ctx      = 1024\nllama_model_load_internal: n_embd     = 5120\nllama_model_load_internal: n_mult     = 256\nllama_model_load_internal: n_head     = 40\nllama_model_load_internal: n_layer    = 40\nllama_model_load_internal: n_rot      = 128\nllama_model_load_internal: ftype      = 3 (mostly Q4_1)\nllama_model_load_internal: n_ff       = 13824\nllama_model_load_internal: n_parts    = 1\nllama_model_load_internal: model size = 13B\nllama_model_load_internal: ggml ctx size = 7759.48 MB\nllama_model_load_internal: mem required  = 9807.48 MB (+ 1608.00 MB per state)\n....................................................................................................\nllama_init_from_file: kv self size  =  800.00 MB\n\nTranscript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.\n\nUser: Hello, Bob.\nBob: Hello. How may I help you today?\nUser: Please tell me the largest city in Europe.\nBob: Sure. The largest city in Europe is Moscow, the capital of Russia.\nUser:\n

Now, enjoy chatting with LLM!

"},{"location":"Tricks/","title":"Tricks for FAQ","text":"

Sometimes, your application with LLM and LLamaSharp may have strange behaviours. Before opening an issue to report the BUG, the following tricks may worth a try.

"},{"location":"Tricks/#carefully-set-the-anti-prompts","title":"Carefully set the anti-prompts","text":"

Anti-prompt can also be called as \"Stop-keyword\", which decides when to stop the response generation. Under interactive mode, the maximum tokens count is always not set, which makes the LLM generates responses infinitively. Therefore, setting anti-prompt correctly helps a lot to avoid the strange behaviours. For example, the prompt file chat-with-bob.txt has the following content:

Transcript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.\n\nUser: Hello, Bob.\nBob: Hello. How may I help you today?\nUser: Please tell me the largest city in Europe.\nBob: Sure. The largest city in Europe is Moscow, the capital of Russia.\nUser:\n

Therefore, the anti-prompt should be set as \"User:\". If the last line of the prompt is removed, LLM will automatically generate a question (user) and a response (bob) for one time when running the chat session. Therefore, the antiprompt is suggested to be appended to the prompt when starting a chat session.

What if an extra line is appended? The string \"User:\" in the prompt will be followed with a char \"\\n\". Thus when running the model, the automatic generation of a pair of question and response may appear because the anti-prompt is \"User:\" but the last token is \"User:\\n\". As for whether it will appear, it's an undefined behaviour, which depends on the implementation inside the LLamaExecutor. Anyway, since it may leads to unexpected behaviors, it's recommended to trim your prompt or carefully keep consistent with your anti-prompt.

"},{"location":"Tricks/#pay-attention-to-the-length-of-prompt","title":"Pay attention to the length of prompt","text":"

Sometimes we want to input a long prompt to execute a task. However, the context size may limit the inference of LLama model. Please ensure the inequality below holds.

$$ len(prompt) + len(response) < len(context) $$

In this inequality, len(response) refers to the expected tokens for LLM to generate.

"},{"location":"Tricks/#try-different-executors-with-a-prompt","title":"Try different executors with a prompt","text":"

Some prompt works well under interactive mode, such as chat-with-bob, some others may work well with instruct mode, such as alpaca. Besides, if your input is quite simple and one-time job, such as \"Q: what is the satellite of the earth? A: \", stateless mode will be a good choice.

If your chat bot has bad performance, trying different executor will possibly make it work well.

"},{"location":"Tricks/#choose-models-weight-depending-on-you-task","title":"Choose models weight depending on you task","text":"

The differences between modes may lead to much different behaviours under the same task. For example, if you're building a chat bot with non-English, a fine-tuned model specially for the language you want to use will have huge effect on the performance.

"},{"location":"Tricks/#set-the-layer-count-you-want-to-offload-to-gpu","title":"Set the layer count you want to offload to GPU","text":"

Currently, the GpuLayerCount parameter, which decides the number of layer loaded into GPU, is set to 20 by default. However, if you have some efficient GPUs, setting it as a larger number will attain faster inference.

"},{"location":"ChatSession/basic-usages/","title":"Basic usages of ChatSession","text":"

ChatSession is a higher-level abstraction than the executors. In the context of a chat application like ChatGPT, a \"chat session\" refers to an interactive conversation or exchange of messages between the user and the chatbot. It represents a continuous flow of communication where the user enters input or asks questions, and the chatbot responds accordingly. A chat session typically starts when the user initiates a conversation with the chatbot and continues until the interaction comes to a natural end or is explicitly terminated by either the user or the system. During a chat session, the chatbot maintains the context of the conversation, remembers previous messages, and generates appropriate responses based on the user's inputs and the ongoing dialogue.

"},{"location":"ChatSession/basic-usages/#initialize-a-session","title":"Initialize a session","text":"

Currently, the only parameter that is accepted is an ILLamaExecutor, because this is the only parameter that we're sure to exist in all the future versions. Since it's the high-level abstraction, we're conservative to the API designs. In the future, there may be more kinds of constructors added.

InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath)));\nChatSession session = new ChatSession(ex);\n
"},{"location":"ChatSession/basic-usages/#chat-with-the-bot","title":"Chat with the bot","text":"

There'll be two kinds of input accepted by the Chat API, which are ChatHistory and String. The API with string is quite similar to that of the executors. Meanwhile, the API with ChatHistory is aimed to provide more flexible usages. For example, you have had a chat with the bot in session A before you open the session B. Now session B has no memory for what you said before. Therefore, you can feed the history of A to B.

string prompt = \"What is C#?\";\n\nforeach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } })) // the inference params should be changed depending on your statement\n{\n    Console.Write(text);\n}\n
"},{"location":"ChatSession/basic-usages/#get-the-history","title":"Get the history","text":"

Currently History is a property of ChatSession.

foreach(var rec in session.History.Messages)\n{\n    Console.WriteLine($\"{rec.AuthorRole}: {rec.Content}\");\n}\n
"},{"location":"ChatSession/save-load-session/","title":"Save/Load Chat Session","text":"

Generally, the chat session could be switched, which requires the ability of loading and saving session.

When building a chat bot app, it's NOT encouraged to initialize many chat sessions and keep them in memory to wait for being switched, because the memory consumption of both CPU and GPU is expensive. It's recommended to save the current session before switching to a new session, and load the file when switching back to the session.

The API is also quite simple, the files will be saved into a directory you specified. If the path does not exist, a new directory will be created.

string savePath = \"<save dir>\";\nsession.SaveSession(savePath);\n\nsession.LoadSession(savePath);\n
"},{"location":"ChatSession/transforms/","title":"Transforms in Chat Session","text":"

There's three important elements in ChatSession, which are input, output and history. Besides, there're some conversions between them. Since the process of them under different conditions varies, LLamaSharp hands over this part of the power to the users.

Currently, there're three kinds of process that could be customized, as introduced below.

"},{"location":"ChatSession/transforms/#input-transform","title":"Input transform","text":"

In general, the input of the chat API is a text (without stream), therefore ChatSession processes it in a pipeline. If you want to use your customized transform, you need to define a transform that implements ITextTransform and add it to the pipeline of ChatSession.

public interface ITextTransform\n{\n    string Transform(string text);\n}\n
public class MyInputTransform1 : ITextTransform\n{\n    public string Transform(string text)\n    {\n        return $\"Question: {text}\\n\";\n    }\n}\n\npublic class MyInputTransform2 : ITextTransform\n{\n    public string Transform(string text)\n    {\n        return text + \"Answer: \";\n    }\n}\n\nsession.AddInputTransform(new MyInputTransform1()).AddInputTransform(new MyInputTransform2());\n
"},{"location":"ChatSession/transforms/#output-transform","title":"Output transform","text":"

Different from the input, the output of chat API is a text stream. Therefore you need to process it word by word, instead of getting the full text at once.

The interface of it has an IEnumerable<string> as input, which is actually a yield sequence.

public interface ITextStreamTransform\n{\n    IEnumerable<string> Transform(IEnumerable<string> tokens);\n    IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens);\n}\n

When implementing it, you could throw a not-implemented exception in one of them if you only need to use the chat API in synchronously or asynchronously.

Different from the input transform pipeline, the output transform only supports one transform.

session.WithOutputTransform(new MyOutputTransform());\n

Here's an example of how to implement the interface. In this example, the transform detects whether there's some keywords in the response and removes them.

/// <summary>\n/// A text output transform that removes the keywords from the response.\n/// </summary>\npublic class KeywordTextOutputStreamTransform : ITextStreamTransform\n{\n    HashSet<string> _keywords;\n    int _maxKeywordLength;\n    bool _removeAllMatchedTokens;\n\n    /// <summary>\n    /// \n    /// </summary>\n    /// <param name=\"keywords\">Keywords that you want to remove from the response.</param>\n    /// <param name=\"redundancyLength\">The extra length when searching for the keyword. For example, if your only keyword is \"highlight\", \n    /// maybe the token you get is \"\\r\\nhighligt\". In this condition, if redundancyLength=0, the token cannot be successfully matched because the length of \"\\r\\nhighligt\" (10)\n    /// has already exceeded the maximum length of the keywords (8). On the contrary, setting redundancyLengyh >= 2 leads to successful match.\n    /// The larger the redundancyLength is, the lower the processing speed. But as an experience, it won't introduce too much performance impact when redundancyLength <= 5 </param>\n    /// <param name=\"removeAllMatchedTokens\">If set to true, when getting a matched keyword, all the related tokens will be removed. Otherwise only the part of keyword will be removed.</param>\n    public KeywordTextOutputStreamTransform(IEnumerable<string> keywords, int redundancyLength = 3, bool removeAllMatchedTokens = false)\n    {\n        _keywords = new(keywords);\n        _maxKeywordLength = keywords.Select(x => x.Length).Max() + redundancyLength;\n        _removeAllMatchedTokens = removeAllMatchedTokens;\n    }\n    /// <inheritdoc />\n    public IEnumerable<string> Transform(IEnumerable<string> tokens)\n    {\n        var window = new Queue<string>();\n\n        foreach (var s in tokens)\n        {\n            window.Enqueue(s);\n            var current = string.Join(\"\", window);\n            if (_keywords.Any(x => current.Contains(x)))\n            {\n                var matchedKeyword = _keywords.First(x => current.Contains(x));\n                int total = window.Count;\n                for (int i = 0; i < total; i++)\n                {\n                    window.Dequeue();\n                }\n                if (!_removeAllMatchedTokens)\n                {\n                    yield return current.Replace(matchedKeyword, \"\");\n                }\n            }\n            if (current.Length >= _maxKeywordLength)\n            {\n                if (_keywords.Any(x => current.Contains(x)))\n                {\n                    var matchedKeyword = _keywords.First(x => current.Contains(x));\n                    int total = window.Count;\n                    for (int i = 0; i < total; i++)\n                    {\n                        window.Dequeue();\n                    }\n                    if (!_removeAllMatchedTokens)\n                    {\n                        yield return current.Replace(matchedKeyword, \"\");\n                    }\n                }\n                else\n                {\n                    int total = window.Count;\n                    for (int i = 0; i < total; i++)\n                    {\n                        yield return window.Dequeue();\n                    }\n                }\n            }\n        }\n        int totalCount = window.Count;\n        for (int i = 0; i < totalCount; i++)\n        {\n            yield return window.Dequeue();\n        }\n    }\n    /// <inheritdoc />\n    public async IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens)\n    {\n        throw new NotImplementedException(); // This is implemented in `LLamaTransforms` but we ignore it here.\n    }\n}\n
"},{"location":"ChatSession/transforms/#history-transform","title":"History transform","text":"

The chat history could be converted to or from a text, which is exactly what the interface of it.

public interface IHistoryTransform\n{\n    string HistoryToText(ChatHistory history);\n    ChatHistory TextToHistory(AuthorRole role, string text);\n}\n

Similar to the output transform, the history transform is added in the following way:

session.WithHistoryTransform(new MyHistoryTransform());\n

The implementation is quite flexible, depending on what you want the history message to be like. Here's an example, which is the default history transform in LLamaSharp.

/// <summary>\n/// The default history transform.\n/// Uses plain text with the following format:\n/// [Author]: [Message]\n/// </summary>\npublic class DefaultHistoryTransform : IHistoryTransform\n{\n    private readonly string defaultUserName = \"User\";\n    private readonly string defaultAssistantName = \"Assistant\";\n    private readonly string defaultSystemName = \"System\";\n    private readonly string defaultUnknownName = \"??\";\n\n    string _userName;\n    string _assistantName;\n    string _systemName;\n    string _unknownName;\n    bool _isInstructMode;\n    public DefaultHistoryTransform(string? userName = null, string? assistantName = null, \n        string? systemName = null, string? unknownName = null, bool isInstructMode = false)\n    {\n        _userName = userName ?? defaultUserName;\n        _assistantName = assistantName ?? defaultAssistantName;\n        _systemName = systemName ?? defaultSystemName;\n        _unknownName = unknownName ?? defaultUnknownName;\n        _isInstructMode = isInstructMode;\n    }\n\n    public virtual string HistoryToText(ChatHistory history)\n    {\n        StringBuilder sb = new();\n        foreach (var message in history.Messages)\n        {\n            if (message.AuthorRole == AuthorRole.User)\n            {\n                sb.AppendLine($\"{_userName}: {message.Content}\");\n            }\n            else if (message.AuthorRole == AuthorRole.System)\n            {\n                sb.AppendLine($\"{_systemName}: {message.Content}\");\n            }\n            else if (message.AuthorRole == AuthorRole.Unknown)\n            {\n                sb.AppendLine($\"{_unknownName}: {message.Content}\");\n            }\n            else if (message.AuthorRole == AuthorRole.Assistant)\n            {\n                sb.AppendLine($\"{_assistantName}: {message.Content}\");\n            }\n        }\n        return sb.ToString();\n    }\n\n    public virtual ChatHistory TextToHistory(AuthorRole role, string text)\n    {\n        ChatHistory history = new ChatHistory();\n        history.AddMessage(role, TrimNamesFromText(text, role));\n        return history;\n    }\n\n    public virtual string TrimNamesFromText(string text, AuthorRole role)\n    {\n        if (role == AuthorRole.User && text.StartsWith($\"{_userName}:\"))\n        {\n            text = text.Substring($\"{_userName}:\".Length).TrimStart();\n        }\n        else if (role == AuthorRole.Assistant && text.EndsWith($\"{_assistantName}:\"))\n        {\n            text = text.Substring(0, text.Length - $\"{_assistantName}:\".Length).TrimEnd();\n        }\n        if (_isInstructMode && role == AuthorRole.Assistant && text.EndsWith(\"\\n> \"))\n        {\n            text = text.Substring(0, text.Length - \"\\n> \".Length).TrimEnd();\n        }\n        return text;\n    }\n}\n
"},{"location":"Examples/ChatSessionStripRoleName/","title":"Use chat session and strip role names","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class ChatSessionStripRoleName\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/chat-with-bob.txt\").Trim();\n        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));\n        ChatSession session = new ChatSession(ex).WithOutputTransform(new LLamaTransforms.KeywordTextOutputStreamTransform(new string[] { \"User:\", \"Bob:\" }, redundancyLength: 8));\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The chat session has started. The role names won't be printed.\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        while (true)\n        {\n            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } }))\n            {\n                Console.Write(text);\n            }\n\n            Console.ForegroundColor = ConsoleColor.Green;\n            prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n        }\n    }\n}\n
"},{"location":"Examples/ChatSessionWithRoleName/","title":"Use chat session without removing role names","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class ChatSessionWithRoleName\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/chat-with-bob.txt\").Trim();\n        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));\n        ChatSession session = new ChatSession(ex); // The only change is to remove the transform for the output text stream.\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The chat session has started. In this example, the prompt is printed for better visual result.\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        // show the prompt\n        Console.Write(prompt);\n        while (true)\n        {\n            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } }))\n            {\n                Console.Write(text);\n            }\n\n            Console.ForegroundColor = ConsoleColor.Green;\n            prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n        }\n    }\n}\n
"},{"location":"Examples/GetEmbeddings/","title":"Get embeddings","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class GetEmbeddings\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var embedder = new LLamaEmbedder(new ModelParams(modelPath));\n\n        while (true)\n        {\n            Console.Write(\"Please input your text: \");\n            Console.ForegroundColor = ConsoleColor.Green;\n            var text = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n\n            Console.WriteLine(string.Join(\", \", embedder.GetEmbeddings(text)));\n            Console.WriteLine();\n        }\n    }\n}\n
"},{"location":"Examples/InstructModeExecute/","title":"Use instruct executor","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class InstructModeExecute\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/dan.txt\").Trim();\n\n        InstructExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024)));\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The executor has been enabled. In this example, the LLM will follow your instructions. For example, you can input \\\"Write a story about a fox who want to \" +\n            \"make friend with human, no less than 200 words.\\\"\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        var inferenceParams = new InferenceParams() { Temperature = 0.8f, MaxTokens = 300 };\n\n        while (true)\n        {\n            foreach (var text in ex.Infer(prompt, inferenceParams))\n            {\n                Console.Write(text);\n            }\n            Console.ForegroundColor = ConsoleColor.Green;\n            prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n        }\n    }\n}\n
"},{"location":"Examples/InteractiveModeExecute/","title":"Use interactive executor","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class InteractiveModeExecute\n{\n    public async static Task Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/chat-with-bob.txt\").Trim();\n\n        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to 64 and the context size is 256. (an example for small scale usage)\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        Console.Write(prompt);\n\n        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" }, MaxTokens = 64 };\n\n        while (true)\n        {\n            await foreach (var text in ex.InferAsync(prompt, inferenceParams))\n            {\n                Console.Write(text);\n            }\n            Console.ForegroundColor = ConsoleColor.Green;\n            prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n        }\n    }\n}\n
"},{"location":"Examples/LoadAndSaveSession/","title":"Load and save chat session","text":"
using LLama.Common;\nusing LLama.OldVersion;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class SaveAndLoadSession\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/chat-with-bob.txt\").Trim();\n        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));\n        ChatSession session = new ChatSession(ex); // The only change is to remove the transform for the output text stream.\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The chat session has started. In this example, the prompt is printed for better visual result. Input \\\"save\\\" to save and reload the session.\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        // show the prompt\n        Console.Write(prompt);\n        while (true)\n        {\n            foreach (var text in session.Chat(prompt, new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } }))\n            {\n                Console.Write(text);\n            }\n\n            Console.ForegroundColor = ConsoleColor.Green;\n            prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White;\n            if (prompt == \"save\")\n            {\n                Console.Write(\"Preparing to save the state, please input the path you want to save it: \");\n                Console.ForegroundColor = ConsoleColor.Green;\n                var statePath = Console.ReadLine();\n                session.SaveSession(statePath);\n                Console.ForegroundColor = ConsoleColor.White;\n                Console.ForegroundColor = ConsoleColor.Yellow;\n                Console.WriteLine(\"Saved session!\");\n                Console.ForegroundColor = ConsoleColor.White;\n\n                ex.Model.Dispose();\n                ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 1024, seed: 1337, gpuLayerCount: 5)));\n                session = new ChatSession(ex).WithOutputTransform(new LLamaTransforms.KeywordTextOutputStreamTransform(new string[] { \"User:\", \"Bob:\" }, redundancyLength: 8));\n                session.LoadSession(statePath);\n\n                Console.ForegroundColor = ConsoleColor.Yellow;\n                Console.WriteLine(\"Loaded session!\");\n                Console.ForegroundColor = ConsoleColor.White;\n\n                Console.Write(\"Now you can continue your session: \");\n                Console.ForegroundColor = ConsoleColor.Green;\n                prompt = Console.ReadLine();\n                Console.ForegroundColor = ConsoleColor.White;\n            }\n        }\n    }\n}\n
"},{"location":"Examples/LoadAndSaveState/","title":"Load and save model/executor state","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class LoadAndSaveState\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n        var prompt = File.ReadAllText(\"Assets/chat-with-bob.txt\").Trim();\n\n        InteractiveExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to 64 and the context size is 256. (an example for small scale usage)\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        Console.Write(prompt);\n\n        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"User:\" } };\n\n        while (true)\n        {\n            foreach (var text in ex.Infer(prompt, inferenceParams))\n            {\n                Console.Write(text);\n            }\n\n            prompt = Console.ReadLine();\n            if (prompt == \"save\")\n            {\n                Console.Write(\"Your path to save model state: \");\n                string modelStatePath = Console.ReadLine();\n                ex.Model.SaveState(modelStatePath);\n\n                Console.Write(\"Your path to save executor state: \");\n                string executorStatePath = Console.ReadLine();\n                ex.SaveState(executorStatePath);\n\n                Console.ForegroundColor = ConsoleColor.Yellow;\n                Console.WriteLine(\"All states saved!\");\n                Console.ForegroundColor = ConsoleColor.White;\n\n                var model = ex.Model;\n                model.LoadState(modelStatePath);\n                ex = new InteractiveExecutor(model);\n                ex.LoadState(executorStatePath);\n                Console.ForegroundColor = ConsoleColor.Yellow;\n                Console.WriteLine(\"Loaded state!\");\n                Console.ForegroundColor = ConsoleColor.White;\n\n                Console.Write(\"Now you can continue your session: \");\n                Console.ForegroundColor = ConsoleColor.Green;\n                prompt = Console.ReadLine();\n                Console.ForegroundColor = ConsoleColor.White;\n            }\n        }\n    }\n}\n
"},{"location":"Examples/QuantizeModel/","title":"Quantize model","text":"
using System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading;\nusing System.Threading.Tasks;\n\npublic class QuantizeModel\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your original model path: \");\n        var inputPath = Console.ReadLine();\n        Console.Write(\"Please input your output model path: \");\n        var outputPath = Console.ReadLine();\n        Console.Write(\"Please input the quantize type (one of q4_0, q4_1, q5_0, q5_1, q8_0): \");\n        var quantizeType = Console.ReadLine();\n        if (LLamaQuantizer.Quantize(inputPath, outputPath, quantizeType))\n        {\n            Console.WriteLine(\"Quantization succeed!\");\n        }\n        else\n        {\n            Console.WriteLine(\"Quantization failed!\");\n        }\n    }\n}\n
"},{"location":"Examples/StatelessModeExecute/","title":"Use stateless executor","text":"
using LLama.Common;\nusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\npublic class StatelessModeExecute\n{\n    public static void Run()\n    {\n        Console.Write(\"Please input your model path: \");\n        string modelPath = Console.ReadLine();\n\n        StatelessExecutor ex = new(new LLamaModel(new ModelParams(modelPath, contextSize: 256)));\n\n        Console.ForegroundColor = ConsoleColor.Yellow;\n        Console.WriteLine(\"The executor has been enabled. In this example, the inference is an one-time job. That says, the previous input and response has \" +\n            \"no impact on the current response. Now you can ask it questions. Note that in this example, no prompt was set for LLM and the maximum response tokens is 50. \" +\n            \"It may not perform well because of lack of prompt. This is also an example that could indicate the improtance of prompt in LLM. To improve it, you can add \" +\n            \"a prompt for it yourself!\");\n        Console.ForegroundColor = ConsoleColor.White;\n\n        var inferenceParams = new InferenceParams() { Temperature = 0.6f, AntiPrompts = new List<string> { \"Question:\", \"#\", \"Question: \", \".\\n\" }, MaxTokens = 50 };\n\n        while (true)\n        {\n            Console.Write(\"\\nQuestion: \");\n            Console.ForegroundColor = ConsoleColor.Green;\n            string prompt = Console.ReadLine();\n            Console.ForegroundColor = ConsoleColor.White; \n            Console.Write(\"Answer: \");\n            prompt = $\"Question: {prompt.Trim()} Answer: \";\n            foreach (var text in ex.Infer(prompt, inferenceParams))\n            {\n                Console.Write(text);\n            }\n        }\n    }\n}\n
"},{"location":"HighLevelApps/bot-sharp/","title":"The Usage of BotSharp Integration","text":"

The document is under work, please have a wait. Thank you for your support! :)

"},{"location":"HighLevelApps/semantic-kernel/","title":"The Usage of semantic-kernel Integration","text":"

Please see this doc

"},{"location":"LLamaExecutors/differences/","title":"Differences of Executors","text":""},{"location":"LLamaExecutors/differences/#differences-between-the-executors","title":"Differences between the executors","text":"

There're currently three kinds of executors provided, which are InteractiveExecutor, InstructExecutor and StatelessExecutor.

In a word, InteractiveExecutor is suitable for getting answer of your questions from LLM continuously. InstructExecutor let LLM execute your instructions, such as \"continue writing\". StatelessExecutor is best for one-time job because the previous inference has no impact on the current inference.

"},{"location":"LLamaExecutors/differences/#interactive-mode-instruct-mode","title":"Interactive mode & Instruct mode","text":"

Both of them are taking \"completing the prompt\" as the goal to generate the response. For example, if you input Long long ago, there was a fox who wanted to make friend with humen. One day, then the LLM will continue to write the story.

Under interactive mode, you serve a role of user and the LLM serves the role of assistant. Then it will help you with your question or request.

Under instruct mode, you give LLM some instructions and it follows.

Though the behaviors of them sounds similar, it could introduce many differences depending on your prompt. For example, \"chat-with-bob\" has good performance under interactive mode and alpaca does well with instruct mode.

// chat-with-bob\n\nTranscript of a dialog, where the User interacts with an Assistant named Bob. Bob is helpful, kind, honest, good at writing, and never fails to answer the User's requests immediately and with precision.\n\nUser: Hello, Bob.\nBob: Hello. How may I help you today?\nUser: Please tell me the largest city in Europe.\nBob: Sure. The largest city in Europe is Moscow, the capital of Russia.\nUser:\n
// alpaca\n\nBelow is an instruction that describes a task. Write a response that appropriately completes the request.\n

Therefore, please modify the prompt correspondingly when switching from one mode to the other.

"},{"location":"LLamaExecutors/differences/#stateful-mode-and-stateless-mode","title":"Stateful mode and Stateless mode.","text":"

Despite the differences between interactive mode and instruct mode, both of them are stateful mode. That is, your previous question/instruction will impact on the current response from LLM. On the contrary, the stateless executor does not have such a \"memory\". No matter how many times you talk to it, it will only concentrate on what you say in this time.

Since the stateless executor has no memory of conversations before, you need to input your question with the whole prompt into it to get the better answer.

For example, if you feed Q: Who is Trump? A: to the stateless executor, it may give the following answer with the antiprompt Q:.

Donald J. Trump, born June 14, 1946, is an American businessman, television personality, politician and the 45th President of the United States (2017-2021). # Anexo:Torneo de Hamburgo 2022 (individual masculino)\n\n## Presentaci\u00f3n previa\n\n* Defensor del t\u00edtulo:  Daniil Medv\u00e9dev\n

It seems that things went well at first. However, after answering the question itself, LLM began to talk about some other things until the answer reached the token count limit. The reason of this strange behavior is the anti-prompt cannot be match. With the input, LLM cannot decide whether to append a string \"A: \" at the end of the response.

As an improvement, let's take the following text as the input:

Q: What is the capital of the USA? A: Washingtong. Q: What is the sum of 1 and 2? A: 3. Q: Who is Trump? A: \n

Then, I got the following answer with the anti-prompt Q:.

45th president of the United States.\n

At this time, by repeating the same mode of Q: xxx? A: xxx., LLM outputs the anti-prompt we want to help to decide where to stop the generation.

"},{"location":"LLamaExecutors/parameters/","title":"Inference Parameters","text":"

Different from LLamaModel, when using an executor, InferenceParams is passed to the Infer method instead of constructor. This is because executors only define the ways to run the model, therefore in each run, you can change the settings for this time inference.

"},{"location":"LLamaExecutors/parameters/#inferenceparams","title":"InferenceParams","text":"

Namespace: LLama.Common

public class InferenceParams\n

Inheritance Object \u2192 InferenceParams

"},{"location":"LLamaExecutors/parameters/#properties","title":"Properties","text":""},{"location":"LLamaExecutors/parameters/#tokenskeep","title":"TokensKeep","text":"

number of tokens to keep from initial prompt

public int TokensKeep { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value","title":"Property Value","text":"

Int32

"},{"location":"LLamaExecutors/parameters/#maxtokens","title":"MaxTokens","text":"

how many new tokens to predict (n_predict), set to -1 to infinitely generate response until it complete.

public int MaxTokens { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"LLamaExecutors/parameters/#logitbias","title":"LogitBias","text":"

logit bias for specific tokens

public Dictionary<int, float> LogitBias { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_2","title":"Property Value","text":"

Dictionary<Int32, Single>

"},{"location":"LLamaExecutors/parameters/#antiprompts","title":"AntiPrompts","text":"

Sequences where the model will stop generating further tokens.

public IEnumerable<string> AntiPrompts { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_3","title":"Property Value","text":"

IEnumerable<String>

"},{"location":"LLamaExecutors/parameters/#pathsession","title":"PathSession","text":"

path to file for saving/loading model eval state

public string PathSession { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_4","title":"Property Value","text":"

String

"},{"location":"LLamaExecutors/parameters/#inputsuffix","title":"InputSuffix","text":"

string to suffix user inputs with

public string InputSuffix { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_5","title":"Property Value","text":"

String

"},{"location":"LLamaExecutors/parameters/#inputprefix","title":"InputPrefix","text":"

string to prefix user inputs with

public string InputPrefix { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_6","title":"Property Value","text":"

String

"},{"location":"LLamaExecutors/parameters/#topk","title":"TopK","text":"

0 or lower to use vocab size

public int TopK { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_7","title":"Property Value","text":"

Int32

"},{"location":"LLamaExecutors/parameters/#topp","title":"TopP","text":"

1.0 = disabled

public float TopP { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_8","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#tfsz","title":"TfsZ","text":"

1.0 = disabled

public float TfsZ { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_9","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#typicalp","title":"TypicalP","text":"

1.0 = disabled

public float TypicalP { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_10","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#temperature","title":"Temperature","text":"

1.0 = disabled

public float Temperature { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_11","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#repeatpenalty","title":"RepeatPenalty","text":"

1.0 = disabled

public float RepeatPenalty { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_12","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#repeatlasttokenscount","title":"RepeatLastTokensCount","text":"

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

public int RepeatLastTokensCount { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_13","title":"Property Value","text":"

Int32

"},{"location":"LLamaExecutors/parameters/#frequencypenalty","title":"FrequencyPenalty","text":"

frequency penalty coefficient 0.0 = disabled

public float FrequencyPenalty { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_14","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#presencepenalty","title":"PresencePenalty","text":"

presence penalty coefficient 0.0 = disabled

public float PresencePenalty { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_15","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#mirostat","title":"Mirostat","text":"

Mirostat uses tokens instead of words. algorithm described in the paper https://arxiv.org/abs/2007.14966. 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

public MiroStateType Mirostat { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_16","title":"Property Value","text":"

MiroStateType

"},{"location":"LLamaExecutors/parameters/#mirostattau","title":"MirostatTau","text":"

target entropy

public float MirostatTau { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_17","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#mirostateta","title":"MirostatEta","text":"

learning rate

public float MirostatEta { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_18","title":"Property Value","text":"

Single

"},{"location":"LLamaExecutors/parameters/#penalizenl","title":"PenalizeNL","text":"

consider newlines as a repeatable token (penalize_nl)

public bool PenalizeNL { get; set; }\n
"},{"location":"LLamaExecutors/parameters/#property-value_19","title":"Property Value","text":"

Boolean

"},{"location":"LLamaExecutors/save-load-state/","title":"Save/Load State of Executor","text":"

Similar to LLamaModel, an executor also has its state, which can be saved and loaded. Note that in most of cases, the state of executor and the state of the model should be loaded and saved at the same time.

To decouple the model and executor, we provide APIs to save/load state for model and executor respectively. However, during the inference, the processed information will leave footprint in LLamaModel's native context. Therefore, if you just load a state from another executor but keep the model unmodified, some strange things may happen. So will loading model state only.

Is there a condition that requires to load one of them only? The answer is YES. For example, after resetting the model state, if you don't want the inference starting from the new position, leaving the executor unmodified is okay. But, anyway, this flexible usage may cause some unexpected behaviors, therefore please ensure you know what you're doing before using it in this way.

In the future version, we'll open the access for some variables inside the executor to support more flexible usages.

The APIs to load/save state of the executors is similar to that of LLamaModel. However, note that StatelessExecutor doesn't have such APIs because it's stateless itself. Besides, the output of GetStateData is an object of type ExecutorBaseState.

LLamaModel model = new LLamaModel(new ModelParams(\"<modelPath>\"));\nInteractiveExecutor executor = new InteractiveExecutor(model);\n// do some things...\nexecutor.SaveState(\"executor.st\");\nvar stateData = model.GetStateData();\n\nInteractiveExecutor executor2 = new InteractiveExecutor(model);\nexecutor2.LoadState(stateData);\n// do some things...\n\nInteractiveExecutor executor3 = new InteractiveExecutor(model);\nexecutor3.LoadState(\"executor.st\");\n// do some things...\n
"},{"location":"LLamaExecutors/text-to-text-apis/","title":"Text-to-Text APIs of the executors","text":"

All the executors implements the interface ILLamaExecutor, which provides two APIs to execute text-to-text tasks.

public interface ILLamaExecutor\n{\n    public LLamaModel Model { get; }\n\n    IEnumerable<string> Infer(string text, InferenceParams? inferenceParams = null, CancellationToken token = default);\n\n    IAsyncEnumerable<string> InferAsync(string text, InferenceParams? inferenceParams = null, CancellationToken token = default);\n}\n

Just pass the text to the executor with the inference parameters. For the inference parameters, please refer to executor inference parameters doc.

The output of both two APIs are yield enumerable. Therefore, when receiving the output, you can directly use foreach to take actions on each word you get by order, instead of waiting for the whole process completed.

"},{"location":"LLamaModel/embeddings/","title":"Get Embeddings","text":"

Getting the embeddings of a text in LLM is sometimes useful, for example, to train other MLP models.

To get the embeddings, please initialize a LLamaEmbedder and then call GetEmbeddings.

var embedder = new LLamaEmbedder(new ModelParams(\"<modelPath>\"));\nstring text = \"hello, LLM.\";\nfloat[] embeddings = embedder.GetEmbeddings(text);\n

The output is a float array. Note that the length of the array is related with the model you load. If you just want to get a smaller size embedding, please consider changing a model.

"},{"location":"LLamaModel/parameters/","title":"LLamaModel Parameters","text":"

When initializing a LLamaModel object, there're three parameters, ModelParams Params, string encoding = \"UTF-8\", ILLamaLogger? logger = null.

The usage of logger will be further introduced in logger doc. The encoding is the encoding you want to use when dealing with text via this model.

The most important of all, is the ModelParams, which is defined as below. We'll explain the parameters step by step in this document.

public class ModelParams\n{\n    public int ContextSize { get; set; } = 512;\n    public int GpuLayerCount { get; set; } = 20;\n    public int Seed { get; set; } = 1686349486;\n    public bool UseFp16Memory { get; set; } = true;\n    public bool UseMemorymap { get; set; } = true;\n    public bool UseMemoryLock { get; set; } = false;\n    public bool Perplexity { get; set; } = false;\n    public string ModelPath { get; set; }\n    public string LoraAdapter { get; set; } = string.Empty;\n    public string LoraBase { get; set; } = string.Empty;\n    public int Threads { get; set; } = Math.Max(Environment.ProcessorCount / 2, 1);\n    public int BatchSize { get; set; } = 512;\n    public bool ConvertEosToNewLine { get; set; } = false;\n}\n
"},{"location":"LLamaModel/parameters/#modelparams","title":"ModelParams","text":"

Namespace: LLama.Common

public class ModelParams\n

Inheritance Object \u2192 ModelParams

"},{"location":"LLamaModel/parameters/#properties","title":"Properties","text":""},{"location":"LLamaModel/parameters/#contextsize","title":"ContextSize","text":"

Model context size (n_ctx)

public int ContextSize { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value","title":"Property Value","text":"

Int32

"},{"location":"LLamaModel/parameters/#gpulayercount","title":"GpuLayerCount","text":"

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

public int GpuLayerCount { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"LLamaModel/parameters/#seed","title":"Seed","text":"

Seed for the random number generator (seed)

public int Seed { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"LLamaModel/parameters/#usefp16memory","title":"UseFp16Memory","text":"

Use f16 instead of f32 for memory kv (memory_f16)

public bool UseFp16Memory { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_3","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/parameters/#usememorymap","title":"UseMemorymap","text":"

Use mmap for faster loads (use_mmap)

public bool UseMemorymap { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_4","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/parameters/#usememorylock","title":"UseMemoryLock","text":"

Use mlock to keep model in memory (use_mlock)

public bool UseMemoryLock { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_5","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/parameters/#perplexity","title":"Perplexity","text":"

Compute perplexity over the prompt (perplexity)

public bool Perplexity { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_6","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/parameters/#modelpath","title":"ModelPath","text":"

Model path (model)

public string ModelPath { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_7","title":"Property Value","text":"

String

"},{"location":"LLamaModel/parameters/#loraadapter","title":"LoraAdapter","text":"

lora adapter path (lora_adapter)

public string LoraAdapter { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_8","title":"Property Value","text":"

String

"},{"location":"LLamaModel/parameters/#lorabase","title":"LoraBase","text":"

base model path for the lora adapter (lora_base)

public string LoraBase { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_9","title":"Property Value","text":"

String

"},{"location":"LLamaModel/parameters/#threads","title":"Threads","text":"

Number of threads (-1 = autodetect) (n_threads)

public int Threads { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_10","title":"Property Value","text":"

Int32

"},{"location":"LLamaModel/parameters/#batchsize","title":"BatchSize","text":"

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

public int BatchSize { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_11","title":"Property Value","text":"

Int32

"},{"location":"LLamaModel/parameters/#converteostonewline","title":"ConvertEosToNewLine","text":"

Whether to convert eos to newline during the inference.

public bool ConvertEosToNewLine { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_12","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/parameters/#embeddingmode","title":"EmbeddingMode","text":"

Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore.

public bool EmbeddingMode { get; set; }\n
"},{"location":"LLamaModel/parameters/#property-value_13","title":"Property Value","text":"

Boolean

"},{"location":"LLamaModel/quantization/","title":"Quantization","text":"

Quantization is significant to accelerate the model inference. Since there's little accuracy (performance) reduction when quantizing the model, get it easy to quantize it!

To quantize the model, please call Quantize from LLamaQuantizer, which is a static method.

string srcPath = \"<model.bin>\";\nstring dstPath = \"<model_q4_0.bin>\";\nLLamaQuantizer.Quantize(srcPath, dstPath, \"q4_0\");\n// The following overload is also okay.\n// LLamaQuantizer.Quantize(srcPath, dstPath, LLamaFtype.LLAMA_FTYPE_MOSTLY_Q4_0);\n

After calling it, a quantized model file will be saved.

There're currently 5 types of quantization supported:

  • q4_0
  • q4_1
  • q5_0
  • q5_1
  • q8_0
"},{"location":"LLamaModel/save-load-state/","title":"Save/Load State","text":"

There're two ways to load state: loading from path and loading from bite array. Therefore, correspondingly, state data can be extracted as byte array or saved to a file.

LLamaModel model = new LLamaModel(new ModelParams(\"<modelPath>\"));\n// do some things...\nmodel.SaveState(\"model.st\");\nvar stateData = model.GetStateData();\nmodel.Dispose();\n\nLLamaModel model2 = new LLamaModel(new ModelParams(\"<modelPath>\"));\nmodel2.LoadState(stateData);\n// do some things...\n\nLLamaModel model3 = new LLamaModel(new ModelParams(\"<modelPath>\"));\nmodel3.LoadState(\"model.st\");\n// do some things...\n
"},{"location":"LLamaModel/tokenization/","title":"Tokenization/Detokenization","text":"

A pair of APIs to make conversion between text and tokens.

"},{"location":"LLamaModel/tokenization/#tokenization","title":"Tokenization","text":"

The basic usage is to call Tokenize after initializing the model.

LLamaModel model = new LLamaModel(new ModelParams(\"<modelPath>\"));\nstring text = \"hello\";\nint[] tokens = model.Tokenize(text).ToArray();\n

Depending on different model (or vocab), the output will be various.

"},{"location":"LLamaModel/tokenization/#detokenization","title":"Detokenization","text":"

Similar to tokenization, just pass an IEnumerable<int> to Detokenize method.

LLamaModel model = new LLamaModel(new ModelParams(\"<modelPath>\"));\nint[] tokens = new int[] {125, 2568, 13245};\nstring text = model.Detokenize(tokens);\n
"},{"location":"More/log/","title":"The Logger in LLamaSharp","text":"

LLamaSharp supports customized logger because it could be used in many kinds of applications, like Winform/WPF, WebAPI and Blazor, so that the preference of logger varies.

"},{"location":"More/log/#define-customized-logger","title":"Define customized logger","text":"

What you need to do is to implement the ILogger interface.

public interface ILLamaLogger\n{\n    public enum LogLevel\n    {\n        Info,\n        Debug,\n        Warning,\n        Error\n    }\n    void Log(string source, string message, LogLevel level);\n}\n

The source specifies where the log message is from, which could be a function, a class, etc..

The message is the log message itself.

The level is the level of the information in the log. As shown above, there're four levels, which are info, debug, warning and error respectively.

The following is a simple example of the logger implementation:

public sealed class LLamaDefaultLogger : ILLamaLogger\n{\n    private static readonly Lazy<LLamaDefaultLogger> _instance = new Lazy<LLamaDefaultLogger>(() => new LLamaDefaultLogger());\n\n    private bool _toConsole = true;\n    private bool _toFile = false;\n\n    private FileStream? _fileStream = null;\n    private StreamWriter _fileWriter = null;\n\n    public static LLamaDefaultLogger Default => _instance.Value;\n\n    private LLamaDefaultLogger()\n    {\n\n    }\n\n    public LLamaDefaultLogger EnableConsole()\n    {\n        _toConsole = true;\n        return this;\n    }\n\n    public LLamaDefaultLogger DisableConsole()\n    {\n        _toConsole = false;\n        return this;\n    }\n\n    public LLamaDefaultLogger EnableFile(string filename, FileMode mode = FileMode.Append)\n    {\n        _fileStream = new FileStream(filename, mode, FileAccess.Write);\n        _fileWriter = new StreamWriter(_fileStream);\n        _toFile = true;\n        return this;\n    }\n\n    public LLamaDefaultLogger DisableFile(string filename)\n    {\n        if (_fileWriter is not null)\n        {\n            _fileWriter.Close();\n            _fileWriter = null;\n        }\n        if (_fileStream is not null)\n        {\n            _fileStream.Close();\n            _fileStream = null;\n        }\n        _toFile = false;\n        return this;\n    }\n\n    public void Log(string source, string message, LogLevel level)\n    {\n        if (level == LogLevel.Info)\n        {\n            Info(message);\n        }\n        else if (level == LogLevel.Debug)\n        {\n\n        }\n        else if (level == LogLevel.Warning)\n        {\n            Warn(message);\n        }\n        else if (level == LogLevel.Error)\n        {\n            Error(message);\n        }\n    }\n\n    public void Info(string message)\n    {\n        message = MessageFormat(\"info\", message);\n        if (_toConsole)\n        {\n            Console.ForegroundColor = ConsoleColor.White;\n            Console.WriteLine(message);\n            Console.ResetColor();\n        }\n        if (_toFile)\n        {\n            Debug.Assert(_fileStream is not null);\n            Debug.Assert(_fileWriter is not null);\n            _fileWriter.WriteLine(message);\n        }\n    }\n\n    public void Warn(string message)\n    {\n        message = MessageFormat(\"warn\", message);\n        if (_toConsole)\n        {\n            Console.ForegroundColor = ConsoleColor.Yellow;\n            Console.WriteLine(message);\n            Console.ResetColor();\n        }\n        if (_toFile)\n        {\n            Debug.Assert(_fileStream is not null);\n            Debug.Assert(_fileWriter is not null);\n            _fileWriter.WriteLine(message);\n        }\n    }\n\n    public void Error(string message)\n    {\n        message = MessageFormat(\"error\", message);\n        if (_toConsole)\n        {\n            Console.ForegroundColor = ConsoleColor.Red;\n            Console.WriteLine(message);\n            Console.ResetColor();\n        }\n        if (_toFile)\n        {\n            Debug.Assert(_fileStream is not null);\n            Debug.Assert(_fileWriter is not null);\n            _fileWriter.WriteLine(message);\n        }\n    }\n\n    private string MessageFormat(string level, string message)\n    {\n        DateTime now = DateTime.Now;\n        string formattedDate = now.ToString(\"yyyy.MM.dd HH:mm:ss\");\n        return $\"[{formattedDate}][{level}]: {message}\";\n    }\n}\n
"},{"location":"NonEnglishUsage/Chinese/","title":"Use LLamaSharp with Chinese","text":"

It's supported now but the document is under work. Please wait for some time. Thank you for your support! :)

"},{"location":"xmldocs/","title":"LLamaSharp","text":""},{"location":"xmldocs/#llama","title":"LLama","text":"

ChatSession

InstructExecutor

InteractiveExecutor

LLamaContext

LLamaEmbedder

LLamaQuantizer

LLamaTransforms

LLamaWeights

StatefulExecutorBase

StatelessExecutor

Utils

"},{"location":"xmldocs/#llamaabstractions","title":"LLama.Abstractions","text":"

IHistoryTransform

IInferenceParams

ILLamaExecutor

IModelParams

ITextStreamTransform

ITextTransform

"},{"location":"xmldocs/#llamacommon","title":"LLama.Common","text":"

AuthorRole

ChatHistory

FixedSizeQueue<T>

ILLamaLogger

InferenceParams

LLamaDefaultLogger

MirostatType

ModelParams

"},{"location":"xmldocs/#llamaexceptions","title":"LLama.Exceptions","text":"

GrammarExpectedName

GrammarExpectedNext

GrammarExpectedPrevious

GrammarFormatException

GrammarUnexpectedCharAltElement

GrammarUnexpectedCharRngElement

GrammarUnexpectedEndElement

GrammarUnexpectedEndOfInput

GrammarUnexpectedHexCharsCount

GrammarUnknownEscapeCharacter

RuntimeError

"},{"location":"xmldocs/#llamaextensions","title":"LLama.Extensions","text":"

IModelParamsExtensions

KeyValuePairExtensions

"},{"location":"xmldocs/#llamagrammars","title":"LLama.Grammars","text":"

Grammar

GrammarRule

"},{"location":"xmldocs/#llamanative","title":"LLama.Native","text":"

LLamaContextParams

LLamaFtype

LLamaGrammarElement

LLamaGrammarElementType

LLamaModelQuantizeParams

LLamaTokenData

LLamaTokenDataArray

LLamaTokenDataArrayNative

NativeApi

SafeLLamaContextHandle

SafeLLamaGrammarHandle

SafeLLamaHandleBase

SafeLlamaModelHandle

SamplingApi

"},{"location":"xmldocs/#llamaoldversion","title":"LLama.OldVersion","text":"

ChatCompletion

ChatCompletionChoice

ChatCompletionChunk

ChatCompletionChunkChoice

ChatCompletionChunkDelta

ChatCompletionMessage

ChatMessageRecord

ChatRole

ChatSession<T>

Completion

CompletionChoice

CompletionChunk

CompletionLogprobs

CompletionUsage

Embedding

EmbeddingData

EmbeddingUsage

IChatModel

LLamaEmbedder

LLamaModel

LLamaParams

"},{"location":"xmldocs/llama.abstractions.ihistorytransform/","title":"IHistoryTransform","text":"

Namespace: LLama.Abstractions

Transform history to plain text and vice versa.

public interface IHistoryTransform\n
"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.abstractions.ihistorytransform/#historytotextchathistory","title":"HistoryToText(ChatHistory)","text":"

Convert a ChatHistory instance to plain text.

string HistoryToText(ChatHistory history)\n
"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#parameters","title":"Parameters","text":"

history ChatHistory The ChatHistory instance

"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#texttohistoryauthorrole-string","title":"TextToHistory(AuthorRole, String)","text":"

Converts plain text to a ChatHistory instance.

ChatHistory TextToHistory(AuthorRole role, string text)\n
"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#parameters_1","title":"Parameters","text":"

role AuthorRole The role for the author.

text String The chat history as plain text.

"},{"location":"xmldocs/llama.abstractions.ihistorytransform/#returns_1","title":"Returns","text":"

ChatHistory The updated history.

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/","title":"IInferenceParams","text":"

Namespace: LLama.Abstractions

The paramters used for inference.

public interface IInferenceParams\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.abstractions.iinferenceparams/#tokenskeep","title":"TokensKeep","text":"

number of tokens to keep from initial prompt

public abstract int TokensKeep { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#maxtokens","title":"MaxTokens","text":"

how many new tokens to predict (n_predict), set to -1 to inifinitely generate response until it complete.

public abstract int MaxTokens { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#logitbias","title":"LogitBias","text":"

logit bias for specific tokens

public abstract Dictionary<int, float> LogitBias { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_2","title":"Property Value","text":"

Dictionary<Int32, Single>

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#antiprompts","title":"AntiPrompts","text":"

Sequences where the model will stop generating further tokens.

public abstract IEnumerable<string> AntiPrompts { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_3","title":"Property Value","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#pathsession","title":"PathSession","text":"

path to file for saving/loading model eval state

public abstract string PathSession { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#inputsuffix","title":"InputSuffix","text":"

string to suffix user inputs with

public abstract string InputSuffix { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#inputprefix","title":"InputPrefix","text":"

string to prefix user inputs with

public abstract string InputPrefix { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_6","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#topk","title":"TopK","text":"

0 or lower to use vocab size

public abstract int TopK { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_7","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#topp","title":"TopP","text":"

1.0 = disabled

public abstract float TopP { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_8","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#tfsz","title":"TfsZ","text":"

1.0 = disabled

public abstract float TfsZ { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_9","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#typicalp","title":"TypicalP","text":"

1.0 = disabled

public abstract float TypicalP { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_10","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#temperature","title":"Temperature","text":"

1.0 = disabled

public abstract float Temperature { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_11","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#repeatpenalty","title":"RepeatPenalty","text":"

1.0 = disabled

public abstract float RepeatPenalty { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_12","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#repeatlasttokenscount","title":"RepeatLastTokensCount","text":"

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

public abstract int RepeatLastTokensCount { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_13","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#frequencypenalty","title":"FrequencyPenalty","text":"

frequency penalty coefficient 0.0 = disabled

public abstract float FrequencyPenalty { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_14","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#presencepenalty","title":"PresencePenalty","text":"

presence penalty coefficient 0.0 = disabled

public abstract float PresencePenalty { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_15","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#mirostat","title":"Mirostat","text":"

Mirostat uses tokens instead of words. algorithm described in the paper https://arxiv.org/abs/2007.14966. 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

public abstract MirostatType Mirostat { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_16","title":"Property Value","text":"

MirostatType

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#mirostattau","title":"MirostatTau","text":"

target entropy

public abstract float MirostatTau { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_17","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#mirostateta","title":"MirostatEta","text":"

learning rate

public abstract float MirostatEta { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_18","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#penalizenl","title":"PenalizeNL","text":"

consider newlines as a repeatable token (penalize_nl)

public abstract bool PenalizeNL { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_19","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#grammar","title":"Grammar","text":"

Grammar to constrain possible tokens

public abstract SafeLLamaGrammarHandle Grammar { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.iinferenceparams/#property-value_20","title":"Property Value","text":"

SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.abstractions.illamaexecutor/","title":"ILLamaExecutor","text":"

Namespace: LLama.Abstractions

A high level interface for LLama models.

public interface ILLamaExecutor\n
"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.abstractions.illamaexecutor/#context","title":"Context","text":"

The loaded context for this executor.

public abstract LLamaContext Context { get; }\n
"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#property-value","title":"Property Value","text":"

LLamaContext

"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.abstractions.illamaexecutor/#inferstring-iinferenceparams-cancellationtoken","title":"Infer(String, IInferenceParams, CancellationToken)","text":"

Infers a response from the model.

IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken token)\n
"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#parameters","title":"Parameters","text":"

text String Your prompt

inferenceParams IInferenceParams Any additional parameters

token CancellationToken A cancellation token.

"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#returns","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#inferasyncstring-iinferenceparams-cancellationtoken","title":"InferAsync(String, IInferenceParams, CancellationToken)","text":"

Asynchronously infers a response from the model.

IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken token)\n
"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#parameters_1","title":"Parameters","text":"

text String Your prompt

inferenceParams IInferenceParams Any additional parameters

token CancellationToken A cancellation token.

"},{"location":"xmldocs/llama.abstractions.illamaexecutor/#returns_1","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.imodelparams/","title":"IModelParams","text":"

Namespace: LLama.Abstractions

The parameters for initializing a LLama model.

public interface IModelParams\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.abstractions.imodelparams/#contextsize","title":"ContextSize","text":"

Model context size (n_ctx)

public abstract int ContextSize { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#maingpu","title":"MainGpu","text":"

the GPU that is used for scratch and small tensors

public abstract int MainGpu { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#lowvram","title":"LowVram","text":"

if true, reduce VRAM usage at the cost of performance

public abstract bool LowVram { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_2","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#gpulayercount","title":"GpuLayerCount","text":"

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

public abstract int GpuLayerCount { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_3","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#seed","title":"Seed","text":"

Seed for the random number generator (seed)

public abstract int Seed { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_4","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#usefp16memory","title":"UseFp16Memory","text":"

Use f16 instead of f32 for memory kv (memory_f16)

public abstract bool UseFp16Memory { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_5","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#usememorymap","title":"UseMemorymap","text":"

Use mmap for faster loads (use_mmap)

public abstract bool UseMemorymap { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_6","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#usememorylock","title":"UseMemoryLock","text":"

Use mlock to keep model in memory (use_mlock)

public abstract bool UseMemoryLock { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_7","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#perplexity","title":"Perplexity","text":"

Compute perplexity over the prompt (perplexity)

public abstract bool Perplexity { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_8","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#modelpath","title":"ModelPath","text":"

Model path (model)

public abstract string ModelPath { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_9","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.imodelparams/#modelalias","title":"ModelAlias","text":"

model alias

public abstract string ModelAlias { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_10","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.imodelparams/#loraadapter","title":"LoraAdapter","text":"

lora adapter path (lora_adapter)

public abstract string LoraAdapter { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_11","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.imodelparams/#lorabase","title":"LoraBase","text":"

base model path for the lora adapter (lora_base)

public abstract string LoraBase { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_12","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.abstractions.imodelparams/#threads","title":"Threads","text":"

Number of threads (-1 = autodetect) (n_threads)

public abstract int Threads { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_13","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#batchsize","title":"BatchSize","text":"

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

public abstract int BatchSize { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_14","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.abstractions.imodelparams/#converteostonewline","title":"ConvertEosToNewLine","text":"

Whether to convert eos to newline during the inference.

public abstract bool ConvertEosToNewLine { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_15","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#embeddingmode","title":"EmbeddingMode","text":"

Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore.

public abstract bool EmbeddingMode { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_16","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#tensorsplits","title":"TensorSplits","text":"

how split tensors should be distributed across GPUs

public abstract Single[] TensorSplits { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_17","title":"Property Value","text":"

Single[]

"},{"location":"xmldocs/llama.abstractions.imodelparams/#ropefrequencybase","title":"RopeFrequencyBase","text":"

RoPE base frequency

public abstract float RopeFrequencyBase { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_18","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.imodelparams/#ropefrequencyscale","title":"RopeFrequencyScale","text":"

RoPE frequency scaling factor

public abstract float RopeFrequencyScale { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_19","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.abstractions.imodelparams/#mulmatq","title":"MulMatQ","text":"

Use experimental mul_mat_q kernels

public abstract bool MulMatQ { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_20","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.abstractions.imodelparams/#encoding","title":"Encoding","text":"

The encoding to use for models

public abstract Encoding Encoding { get; set; }\n
"},{"location":"xmldocs/llama.abstractions.imodelparams/#property-value_21","title":"Property Value","text":"

Encoding

"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/","title":"ITextStreamTransform","text":"

Namespace: LLama.Abstractions

Takes a stream of tokens and transforms them.

public interface ITextStreamTransform\n
"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#transformienumerablestring","title":"Transform(IEnumerable<String>)","text":"

Takes a stream of tokens and transforms them, returning a new stream of tokens.

IEnumerable<string> Transform(IEnumerable<string> tokens)\n
"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#parameters","title":"Parameters","text":"

tokens IEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#returns","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#transformasynciasyncenumerablestring","title":"TransformAsync(IAsyncEnumerable<String>)","text":"

Takes a stream of tokens and transforms them, returning a new stream of tokens asynchronously.

IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens)\n
"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#parameters_1","title":"Parameters","text":"

tokens IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.itextstreamtransform/#returns_1","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.abstractions.itexttransform/","title":"ITextTransform","text":"

Namespace: LLama.Abstractions

An interface for text transformations. These can be used to compose a pipeline of text transformations, such as: - Tokenization - Lowercasing - Punctuation removal - Trimming - etc.

public interface ITextTransform\n
"},{"location":"xmldocs/llama.abstractions.itexttransform/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.abstractions.itexttransform/#transformstring","title":"Transform(String)","text":"

Takes a string and transforms it.

string Transform(string text)\n
"},{"location":"xmldocs/llama.abstractions.itexttransform/#parameters","title":"Parameters","text":"

text String

"},{"location":"xmldocs/llama.abstractions.itexttransform/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.chatsession/","title":"ChatSession","text":"

Namespace: LLama

The main chat session class.

public class ChatSession\n

Inheritance Object \u2192 ChatSession

"},{"location":"xmldocs/llama.chatsession/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.chatsession/#outputtransform","title":"OutputTransform","text":"

The output transform used in this session.

public ITextStreamTransform OutputTransform;\n
"},{"location":"xmldocs/llama.chatsession/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.chatsession/#executor","title":"Executor","text":"

The executor for this session.

public ILLamaExecutor Executor { get; }\n
"},{"location":"xmldocs/llama.chatsession/#property-value","title":"Property Value","text":"

ILLamaExecutor

"},{"location":"xmldocs/llama.chatsession/#history","title":"History","text":"

The chat history for this session.

public ChatHistory History { get; }\n
"},{"location":"xmldocs/llama.chatsession/#property-value_1","title":"Property Value","text":"

ChatHistory

"},{"location":"xmldocs/llama.chatsession/#historytransform","title":"HistoryTransform","text":"

The history transform used in this session.

public IHistoryTransform HistoryTransform { get; set; }\n
"},{"location":"xmldocs/llama.chatsession/#property-value_2","title":"Property Value","text":"

IHistoryTransform

"},{"location":"xmldocs/llama.chatsession/#inputtransformpipeline","title":"InputTransformPipeline","text":"

The input transform pipeline used in this session.

public List<ITextTransform> InputTransformPipeline { get; set; }\n
"},{"location":"xmldocs/llama.chatsession/#property-value_3","title":"Property Value","text":"

List<ITextTransform>

"},{"location":"xmldocs/llama.chatsession/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.chatsession/#chatsessionillamaexecutor","title":"ChatSession(ILLamaExecutor)","text":"
public ChatSession(ILLamaExecutor executor)\n
"},{"location":"xmldocs/llama.chatsession/#parameters","title":"Parameters","text":"

executor ILLamaExecutor The executor for this session

"},{"location":"xmldocs/llama.chatsession/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.chatsession/#withhistorytransformihistorytransform","title":"WithHistoryTransform(IHistoryTransform)","text":"

Use a custom history transform.

public ChatSession WithHistoryTransform(IHistoryTransform transform)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_1","title":"Parameters","text":"

transform IHistoryTransform

"},{"location":"xmldocs/llama.chatsession/#returns","title":"Returns","text":"

ChatSession

"},{"location":"xmldocs/llama.chatsession/#addinputtransformitexttransform","title":"AddInputTransform(ITextTransform)","text":"

Add a text transform to the input transform pipeline.

public ChatSession AddInputTransform(ITextTransform transform)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_2","title":"Parameters","text":"

transform ITextTransform

"},{"location":"xmldocs/llama.chatsession/#returns_1","title":"Returns","text":"

ChatSession

"},{"location":"xmldocs/llama.chatsession/#withoutputtransformitextstreamtransform","title":"WithOutputTransform(ITextStreamTransform)","text":"

Use a custom output transform.

public ChatSession WithOutputTransform(ITextStreamTransform transform)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_3","title":"Parameters","text":"

transform ITextStreamTransform

"},{"location":"xmldocs/llama.chatsession/#returns_2","title":"Returns","text":"

ChatSession

"},{"location":"xmldocs/llama.chatsession/#savesessionstring","title":"SaveSession(String)","text":"
public void SaveSession(string path)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_4","title":"Parameters","text":"

path String The directory name to save the session. If the directory does not exist, a new directory will be created.

"},{"location":"xmldocs/llama.chatsession/#loadsessionstring","title":"LoadSession(String)","text":"
public void LoadSession(string path)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_5","title":"Parameters","text":"

path String The directory name to load the session.

"},{"location":"xmldocs/llama.chatsession/#chatchathistory-iinferenceparams-cancellationtoken","title":"Chat(ChatHistory, IInferenceParams, CancellationToken)","text":"

Get the response from the LLama model with chat histories.

public IEnumerable<string> Chat(ChatHistory history, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_6","title":"Parameters","text":"

history ChatHistory

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.chatsession/#returns_3","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.chatsession/#chatstring-iinferenceparams-cancellationtoken","title":"Chat(String, IInferenceParams, CancellationToken)","text":"

Get the response from the LLama model. Note that prompt could not only be the preset words, but also the question you want to ask.

public IEnumerable<string> Chat(string prompt, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_7","title":"Parameters","text":"

prompt String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.chatsession/#returns_4","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.chatsession/#chatasyncchathistory-iinferenceparams-cancellationtoken","title":"ChatAsync(ChatHistory, IInferenceParams, CancellationToken)","text":"

Get the response from the LLama model with chat histories.

public IAsyncEnumerable<string> ChatAsync(ChatHistory history, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_8","title":"Parameters","text":"

history ChatHistory

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.chatsession/#returns_5","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.chatsession/#chatasyncstring-iinferenceparams-cancellationtoken","title":"ChatAsync(String, IInferenceParams, CancellationToken)","text":"

Get the response from the LLama model with chat histories asynchronously.

public IAsyncEnumerable<string> ChatAsync(string prompt, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.chatsession/#parameters_9","title":"Parameters","text":"

prompt String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.chatsession/#returns_6","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.common.authorrole/","title":"AuthorRole","text":"

Namespace: LLama.Common

Role of the message author, e.g. user/assistant/system

public enum AuthorRole\n

Inheritance Object \u2192 ValueType \u2192 Enum \u2192 AuthorRole Implements IComparable, IFormattable, IConvertible

"},{"location":"xmldocs/llama.common.authorrole/#fields","title":"Fields","text":"Name Value Description Unknown -1 Role is unknown System 0 Message comes from a \"system\" prompt, not written by a user or language model User 1 Message comes from the user Assistant 2 Messages was generated by the language model"},{"location":"xmldocs/llama.common.chathistory/","title":"ChatHistory","text":"

Namespace: LLama.Common

The chat history class

public class ChatHistory\n

Inheritance Object \u2192 ChatHistory

"},{"location":"xmldocs/llama.common.chathistory/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.common.chathistory/#messages","title":"Messages","text":"

List of messages in the chat

public List<Message> Messages { get; }\n
"},{"location":"xmldocs/llama.common.chathistory/#property-value","title":"Property Value","text":"

List<Message>

"},{"location":"xmldocs/llama.common.chathistory/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.common.chathistory/#chathistory_1","title":"ChatHistory()","text":"

Create a new instance of the chat content class

public ChatHistory()\n
"},{"location":"xmldocs/llama.common.chathistory/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.common.chathistory/#addmessageauthorrole-string","title":"AddMessage(AuthorRole, String)","text":"

Add a message to the chat history

public void AddMessage(AuthorRole authorRole, string content)\n
"},{"location":"xmldocs/llama.common.chathistory/#parameters","title":"Parameters","text":"

authorRole AuthorRole Role of the message author

content String Message content

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/","title":"FixedSizeQueue<T>","text":"

Namespace: LLama.Common

A queue with fixed storage size. Currently it's only a naive implementation and needs to be further optimized in the future.

public class FixedSizeQueue<T> : , System.Collections.IEnumerable\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#type-parameters","title":"Type Parameters","text":"

T

Inheritance Object \u2192 FixedSizeQueue<T> Implements IEnumerable<T>, IEnumerable

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.common.fixedsizequeue-1/#count","title":"Count","text":"

Number of items in this queue

public int Count { get; }\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#capacity","title":"Capacity","text":"

Maximum number of items allowed in this queue

public int Capacity { get; }\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.common.fixedsizequeue-1/#fixedsizequeueint32","title":"FixedSizeQueue(Int32)","text":"

Create a new queue

public FixedSizeQueue(int size)\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#parameters","title":"Parameters","text":"

size Int32 the maximum number of items to store in this queue

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#fixedsizequeueint32-ienumerablet","title":"FixedSizeQueue(Int32, IEnumerable<T>)","text":"

Fill the quene with the data. Please ensure that data.Count <= size

public FixedSizeQueue(int size, IEnumerable<T> data)\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#parameters_1","title":"Parameters","text":"

size Int32

data IEnumerable<T>

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.common.fixedsizequeue-1/#fillwitht","title":"FillWith(T)","text":"

Replace every item in the queue with the given value

public FixedSizeQueue<T> FillWith(T value)\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#parameters_2","title":"Parameters","text":"

value T The value to replace all items with

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#returns","title":"Returns","text":"

FixedSizeQueue<T> returns this

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#enqueuet","title":"Enqueue(T)","text":"

Enquene an element.

public void Enqueue(T item)\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#parameters_3","title":"Parameters","text":"

item T

"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#getenumerator","title":"GetEnumerator()","text":"
public IEnumerator<T> GetEnumerator()\n
"},{"location":"xmldocs/llama.common.fixedsizequeue-1/#returns_1","title":"Returns","text":"

IEnumerator<T>

"},{"location":"xmldocs/llama.common.illamalogger/","title":"ILLamaLogger","text":"

Namespace: LLama.Common

receives log messages from LLamaSharp

public interface ILLamaLogger\n
"},{"location":"xmldocs/llama.common.illamalogger/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.common.illamalogger/#logstring-string-loglevel","title":"Log(String, String, LogLevel)","text":"

Write the log in customized way

void Log(string source, string message, LogLevel level)\n
"},{"location":"xmldocs/llama.common.illamalogger/#parameters","title":"Parameters","text":"

source String The source of the log. It may be a method name or class name.

message String The message.

level LogLevel The log level.

"},{"location":"xmldocs/llama.common.inferenceparams/","title":"InferenceParams","text":"

Namespace: LLama.Common

The paramters used for inference.

public class InferenceParams : LLama.Abstractions.IInferenceParams\n

Inheritance Object \u2192 InferenceParams Implements IInferenceParams

"},{"location":"xmldocs/llama.common.inferenceparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.common.inferenceparams/#tokenskeep","title":"TokensKeep","text":"

number of tokens to keep from initial prompt

public int TokensKeep { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.inferenceparams/#maxtokens","title":"MaxTokens","text":"

how many new tokens to predict (n_predict), set to -1 to inifinitely generate response until it complete.

public int MaxTokens { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.inferenceparams/#logitbias","title":"LogitBias","text":"

logit bias for specific tokens

public Dictionary<int, float> LogitBias { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_2","title":"Property Value","text":"

Dictionary<Int32, Single>

"},{"location":"xmldocs/llama.common.inferenceparams/#antiprompts","title":"AntiPrompts","text":"

Sequences where the model will stop generating further tokens.

public IEnumerable<string> AntiPrompts { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_3","title":"Property Value","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.common.inferenceparams/#pathsession","title":"PathSession","text":"

path to file for saving/loading model eval state

public string PathSession { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.inferenceparams/#inputsuffix","title":"InputSuffix","text":"

string to suffix user inputs with

public string InputSuffix { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.inferenceparams/#inputprefix","title":"InputPrefix","text":"

string to prefix user inputs with

public string InputPrefix { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_6","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.inferenceparams/#topk","title":"TopK","text":"

0 or lower to use vocab size

public int TopK { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_7","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.inferenceparams/#topp","title":"TopP","text":"

1.0 = disabled

public float TopP { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_8","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#tfsz","title":"TfsZ","text":"

1.0 = disabled

public float TfsZ { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_9","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#typicalp","title":"TypicalP","text":"

1.0 = disabled

public float TypicalP { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_10","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#temperature","title":"Temperature","text":"

1.0 = disabled

public float Temperature { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_11","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#repeatpenalty","title":"RepeatPenalty","text":"

1.0 = disabled

public float RepeatPenalty { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_12","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#repeatlasttokenscount","title":"RepeatLastTokensCount","text":"

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

public int RepeatLastTokensCount { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_13","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.inferenceparams/#frequencypenalty","title":"FrequencyPenalty","text":"

frequency penalty coefficient 0.0 = disabled

public float FrequencyPenalty { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_14","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#presencepenalty","title":"PresencePenalty","text":"

presence penalty coefficient 0.0 = disabled

public float PresencePenalty { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_15","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#mirostat","title":"Mirostat","text":"

Mirostat uses tokens instead of words. algorithm described in the paper https://arxiv.org/abs/2007.14966. 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

public MirostatType Mirostat { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_16","title":"Property Value","text":"

MirostatType

"},{"location":"xmldocs/llama.common.inferenceparams/#mirostattau","title":"MirostatTau","text":"

target entropy

public float MirostatTau { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_17","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#mirostateta","title":"MirostatEta","text":"

learning rate

public float MirostatEta { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_18","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.inferenceparams/#penalizenl","title":"PenalizeNL","text":"

consider newlines as a repeatable token (penalize_nl)

public bool PenalizeNL { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_19","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.inferenceparams/#grammar","title":"Grammar","text":"

A grammar to constrain the possible tokens

public SafeLLamaGrammarHandle Grammar { get; set; }\n
"},{"location":"xmldocs/llama.common.inferenceparams/#property-value_20","title":"Property Value","text":"

SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.common.inferenceparams/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.common.inferenceparams/#inferenceparams_1","title":"InferenceParams()","text":"
public InferenceParams()\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/","title":"LLamaDefaultLogger","text":"

Namespace: LLama.Common

The default logger of LLamaSharp. On default it write to console. Use methods of LLamaLogger.Default to change the behavior. It's recommended to inherit ILLamaLogger to customize the behavior.

public sealed class LLamaDefaultLogger : ILLamaLogger\n

Inheritance Object \u2192 LLamaDefaultLogger Implements ILLamaLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.common.llamadefaultlogger/#default","title":"Default","text":"

Get the default logger instance

public static LLamaDefaultLogger Default { get; }\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#property-value","title":"Property Value","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.common.llamadefaultlogger/#enablenative","title":"EnableNative()","text":"

Enable logging output from llama.cpp

public LLamaDefaultLogger EnableNative()\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#enableconsole","title":"EnableConsole()","text":"

Enable writing log messages to console

public LLamaDefaultLogger EnableConsole()\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns_1","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#disableconsole","title":"DisableConsole()","text":"

Disable writing messages to console

public LLamaDefaultLogger DisableConsole()\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns_2","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#enablefilestring-filemode","title":"EnableFile(String, FileMode)","text":"

Enable writing log messages to file

public LLamaDefaultLogger EnableFile(string filename, FileMode mode)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters","title":"Parameters","text":"

filename String

mode FileMode

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns_3","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#disablefilestring","title":"DisableFile(String)","text":""},{"location":"xmldocs/llama.common.llamadefaultlogger/#caution","title":"Caution","text":"

Use DisableFile method without 'filename' parameter

Disable writing log messages to file

public LLamaDefaultLogger DisableFile(string filename)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters_1","title":"Parameters","text":"

filename String unused!

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns_4","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#disablefile","title":"DisableFile()","text":"

Disable writing log messages to file

public LLamaDefaultLogger DisableFile()\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#returns_5","title":"Returns","text":"

LLamaDefaultLogger

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#logstring-string-loglevel","title":"Log(String, String, LogLevel)","text":"

Log a message

public void Log(string source, string message, LogLevel level)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters_2","title":"Parameters","text":"

source String The source of this message (e.g. class name)

message String The message to log

level LogLevel Severity level of this message

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#infostring","title":"Info(String)","text":"

Write a log message with \"Info\" severity

public void Info(string message)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters_3","title":"Parameters","text":"

message String

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#warnstring","title":"Warn(String)","text":"

Write a log message with \"Warn\" severity

public void Warn(string message)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters_4","title":"Parameters","text":"

message String

"},{"location":"xmldocs/llama.common.llamadefaultlogger/#errorstring","title":"Error(String)","text":"

Write a log message with \"Error\" severity

public void Error(string message)\n
"},{"location":"xmldocs/llama.common.llamadefaultlogger/#parameters_5","title":"Parameters","text":"

message String

"},{"location":"xmldocs/llama.common.mirostattype/","title":"MirostatType","text":"

Namespace: LLama.Common

Type of \"mirostat\" sampling to use. https://github.com/basusourya/mirostat

public enum MirostatType\n

Inheritance Object \u2192 ValueType \u2192 Enum \u2192 MirostatType Implements IComparable, IFormattable, IConvertible

"},{"location":"xmldocs/llama.common.mirostattype/#fields","title":"Fields","text":"Name Value Description Disable 0 Disable Mirostat sampling Mirostat 1 Original mirostat algorithm Mirostat2 2 Mirostat 2.0 algorithm"},{"location":"xmldocs/llama.common.modelparams/","title":"ModelParams","text":"

Namespace: LLama.Common

The parameters for initializing a LLama model.

public class ModelParams : LLama.Abstractions.IModelParams, System.IEquatable`1[[LLama.Common.ModelParams, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ModelParams Implements IModelParams, IEquatable<ModelParams>

"},{"location":"xmldocs/llama.common.modelparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.common.modelparams/#contextsize","title":"ContextSize","text":"

Model context size (n_ctx)

public int ContextSize { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#maingpu","title":"MainGpu","text":"

the GPU that is used for scratch and small tensors

public int MainGpu { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#lowvram","title":"LowVram","text":"

if true, reduce VRAM usage at the cost of performance

public bool LowVram { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_2","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#gpulayercount","title":"GpuLayerCount","text":"

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

public int GpuLayerCount { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_3","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#seed","title":"Seed","text":"

Seed for the random number generator (seed)

public int Seed { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_4","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#usefp16memory","title":"UseFp16Memory","text":"

Use f16 instead of f32 for memory kv (memory_f16)

public bool UseFp16Memory { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_5","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#usememorymap","title":"UseMemorymap","text":"

Use mmap for faster loads (use_mmap)

public bool UseMemorymap { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_6","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#usememorylock","title":"UseMemoryLock","text":"

Use mlock to keep model in memory (use_mlock)

public bool UseMemoryLock { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_7","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#perplexity","title":"Perplexity","text":"

Compute perplexity over the prompt (perplexity)

public bool Perplexity { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_8","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#modelpath","title":"ModelPath","text":"

Model path (model)

public string ModelPath { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_9","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.modelparams/#modelalias","title":"ModelAlias","text":"

model alias

public string ModelAlias { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_10","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.modelparams/#loraadapter","title":"LoraAdapter","text":"

lora adapter path (lora_adapter)

public string LoraAdapter { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_11","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.modelparams/#lorabase","title":"LoraBase","text":"

base model path for the lora adapter (lora_base)

public string LoraBase { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_12","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.common.modelparams/#threads","title":"Threads","text":"

Number of threads (-1 = autodetect) (n_threads)

public int Threads { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_13","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#batchsize","title":"BatchSize","text":"

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

public int BatchSize { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_14","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#converteostonewline","title":"ConvertEosToNewLine","text":"

Whether to convert eos to newline during the inference.

public bool ConvertEosToNewLine { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_15","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#embeddingmode","title":"EmbeddingMode","text":"

Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore.

public bool EmbeddingMode { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_16","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#tensorsplits","title":"TensorSplits","text":"

how split tensors should be distributed across GPUs

public Single[] TensorSplits { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_17","title":"Property Value","text":"

Single[]

"},{"location":"xmldocs/llama.common.modelparams/#ropefrequencybase","title":"RopeFrequencyBase","text":"

RoPE base frequency

public float RopeFrequencyBase { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_18","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.modelparams/#ropefrequencyscale","title":"RopeFrequencyScale","text":"

RoPE frequency scaling factor

public float RopeFrequencyScale { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_19","title":"Property Value","text":"

Single

"},{"location":"xmldocs/llama.common.modelparams/#mulmatq","title":"MulMatQ","text":"

Use experimental mul_mat_q kernels

public bool MulMatQ { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_20","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#encoding","title":"Encoding","text":"

The encoding to use to convert text for the model

public Encoding Encoding { get; set; }\n
"},{"location":"xmldocs/llama.common.modelparams/#property-value_21","title":"Property Value","text":"

Encoding

"},{"location":"xmldocs/llama.common.modelparams/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.common.modelparams/#modelparamsstring","title":"ModelParams(String)","text":"
public ModelParams(string modelPath)\n
"},{"location":"xmldocs/llama.common.modelparams/#parameters","title":"Parameters","text":"

modelPath String The model path.

"},{"location":"xmldocs/llama.common.modelparams/#modelparamsstring-int32-int32-int32-boolean-boolean-boolean-boolean-string-string-int32-int32-boolean-boolean-single-single-boolean-string","title":"ModelParams(String, Int32, Int32, Int32, Boolean, Boolean, Boolean, Boolean, String, String, Int32, Int32, Boolean, Boolean, Single, Single, Boolean, String)","text":""},{"location":"xmldocs/llama.common.modelparams/#caution","title":"Caution","text":"

Use object initializer to set all optional parameters

public ModelParams(string modelPath, int contextSize, int gpuLayerCount, int seed, bool useFp16Memory, bool useMemorymap, bool useMemoryLock, bool perplexity, string loraAdapter, string loraBase, int threads, int batchSize, bool convertEosToNewLine, bool embeddingMode, float ropeFrequencyBase, float ropeFrequencyScale, bool mulMatQ, string encoding)\n
"},{"location":"xmldocs/llama.common.modelparams/#parameters_1","title":"Parameters","text":"

modelPath String The model path.

contextSize Int32 Model context size (n_ctx)

gpuLayerCount Int32 Number of layers to run in VRAM / GPU memory (n_gpu_layers)

seed Int32 Seed for the random number generator (seed)

useFp16Memory Boolean Whether to use f16 instead of f32 for memory kv (memory_f16)

useMemorymap Boolean Whether to use mmap for faster loads (use_mmap)

useMemoryLock Boolean Whether to use mlock to keep model in memory (use_mlock)

perplexity Boolean Thether to compute perplexity over the prompt (perplexity)

loraAdapter String Lora adapter path (lora_adapter)

loraBase String Base model path for the lora adapter (lora_base)

threads Int32 Number of threads (-1 = autodetect) (n_threads)

batchSize Int32 Batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

convertEosToNewLine Boolean Whether to convert eos to newline during the inference.

embeddingMode Boolean Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore.

ropeFrequencyBase Single RoPE base frequency.

ropeFrequencyScale Single RoPE frequency scaling factor

mulMatQ Boolean Use experimental mul_mat_q kernels

encoding String The encoding to use to convert text for the model

"},{"location":"xmldocs/llama.common.modelparams/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.common.modelparams/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.common.modelparams/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.common.modelparams/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.common.modelparams/#parameters_2","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.common.modelparams/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.common.modelparams/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.common.modelparams/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.common.modelparams/#parameters_3","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.common.modelparams/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#equalsmodelparams","title":"Equals(ModelParams)","text":"
public bool Equals(ModelParams other)\n
"},{"location":"xmldocs/llama.common.modelparams/#parameters_4","title":"Parameters","text":"

other ModelParams

"},{"location":"xmldocs/llama.common.modelparams/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.common.modelparams/#clone","title":"<Clone>$()","text":"
public ModelParams <Clone>$()\n
"},{"location":"xmldocs/llama.common.modelparams/#returns_5","title":"Returns","text":"

ModelParams

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/","title":"GrammarExpectedName","text":"

Namespace: LLama.Exceptions

Failed to parse a \"name\" element when one was expected

public class GrammarExpectedName : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarExpectedName Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedname/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/","title":"GrammarExpectedNext","text":"

Namespace: LLama.Exceptions

A specified string was expected when parsing

public class GrammarExpectedNext : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarExpectedNext Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectednext/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/","title":"GrammarExpectedPrevious","text":"

Namespace: LLama.Exceptions

A specified character was expected to preceded another when parsing

public class GrammarExpectedPrevious : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarExpectedPrevious Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarexpectedprevious/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/","title":"GrammarFormatException","text":"

Namespace: LLama.Exceptions

Base class for all grammar exceptions

public abstract class GrammarFormatException : System.Exception, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarformatexception/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarformatexception/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/","title":"GrammarUnexpectedCharAltElement","text":"

Namespace: LLama.Exceptions

A CHAR_ALT was created without a preceding CHAR element

public class GrammarUnexpectedCharAltElement : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnexpectedCharAltElement Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharaltelement/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/","title":"GrammarUnexpectedCharRngElement","text":"

Namespace: LLama.Exceptions

A CHAR_RNG was created without a preceding CHAR element

public class GrammarUnexpectedCharRngElement : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnexpectedCharRngElement Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedcharrngelement/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/","title":"GrammarUnexpectedEndElement","text":"

Namespace: LLama.Exceptions

An END was encountered before the last element

public class GrammarUnexpectedEndElement : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnexpectedEndElement Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendelement/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/","title":"GrammarUnexpectedEndOfInput","text":"

Namespace: LLama.Exceptions

End-of-file was encountered while parsing

public class GrammarUnexpectedEndOfInput : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnexpectedEndOfInput Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedendofinput/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/","title":"GrammarUnexpectedHexCharsCount","text":"

Namespace: LLama.Exceptions

An incorrect number of characters were encountered while parsing a hex literal

public class GrammarUnexpectedHexCharsCount : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnexpectedHexCharsCount Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunexpectedhexcharscount/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/","title":"GrammarUnknownEscapeCharacter","text":"

Namespace: LLama.Exceptions

An unexpected character was encountered after an escape sequence

public class GrammarUnknownEscapeCharacter : GrammarFormatException, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 GrammarFormatException \u2192 GrammarUnknownEscapeCharacter Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.grammarunknownescapecharacter/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.runtimeerror/","title":"RuntimeError","text":"

Namespace: LLama.Exceptions

public class RuntimeError : System.Exception, System.Runtime.Serialization.ISerializable\n

Inheritance Object \u2192 Exception \u2192 RuntimeError Implements ISerializable

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.exceptions.runtimeerror/#targetsite","title":"TargetSite","text":"
public MethodBase TargetSite { get; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value","title":"Property Value","text":"

MethodBase

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#message","title":"Message","text":"
public string Message { get; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#data","title":"Data","text":"
public IDictionary Data { get; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_2","title":"Property Value","text":"

IDictionary

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#innerexception","title":"InnerException","text":"
public Exception InnerException { get; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_3","title":"Property Value","text":"

Exception

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#helplink","title":"HelpLink","text":"
public string HelpLink { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_4","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#source","title":"Source","text":"
public string Source { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_5","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#hresult","title":"HResult","text":"
public int HResult { get; set; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#stacktrace","title":"StackTrace","text":"
public string StackTrace { get; }\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#property-value_7","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.exceptions.runtimeerror/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.exceptions.runtimeerror/#runtimeerror_1","title":"RuntimeError()","text":"
public RuntimeError()\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#runtimeerrorstring","title":"RuntimeError(String)","text":"
public RuntimeError(string message)\n
"},{"location":"xmldocs/llama.exceptions.runtimeerror/#parameters","title":"Parameters","text":"

message String

"},{"location":"xmldocs/llama.extensions.imodelparamsextensions/","title":"IModelParamsExtensions","text":"

Namespace: LLama.Extensions

Extention methods to the IModelParams interface

public static class IModelParamsExtensions\n

Inheritance Object \u2192 IModelParamsExtensions

"},{"location":"xmldocs/llama.extensions.imodelparamsextensions/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.extensions.imodelparamsextensions/#tollamacontextparamsimodelparams-llamacontextparams","title":"ToLlamaContextParams(IModelParams, LLamaContextParams&)","text":"

Convert the given IModelParams into a LLamaContextParams

public static MemoryHandle ToLlamaContextParams(IModelParams params, LLamaContextParams& result)\n
"},{"location":"xmldocs/llama.extensions.imodelparamsextensions/#parameters","title":"Parameters","text":"

params IModelParams

result LLamaContextParams&

"},{"location":"xmldocs/llama.extensions.imodelparamsextensions/#returns","title":"Returns","text":"

MemoryHandle

"},{"location":"xmldocs/llama.extensions.imodelparamsextensions/#exceptions","title":"Exceptions","text":"

FileNotFoundException

ArgumentException

"},{"location":"xmldocs/llama.extensions.keyvaluepairextensions/","title":"KeyValuePairExtensions","text":"

Namespace: LLama.Extensions

Extensions to the KeyValuePair struct

public static class KeyValuePairExtensions\n

Inheritance Object \u2192 KeyValuePairExtensions

"},{"location":"xmldocs/llama.extensions.keyvaluepairextensions/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.extensions.keyvaluepairextensions/#deconstructtkey-tvaluekeyvaluepairtkey-tvalue-tkey-tvalue","title":"Deconstruct<TKey, TValue>(KeyValuePair<TKey, TValue>, TKey&, TValue&)","text":"

Deconstruct a KeyValuePair into it's constituent parts.

public static void Deconstruct<TKey, TValue>(KeyValuePair<TKey, TValue> pair, TKey& first, TValue& second)\n
"},{"location":"xmldocs/llama.extensions.keyvaluepairextensions/#type-parameters","title":"Type Parameters","text":"

TKey Type of the Key

TValue Type of the Value

"},{"location":"xmldocs/llama.extensions.keyvaluepairextensions/#parameters","title":"Parameters","text":"

pair KeyValuePair<TKey, TValue> The KeyValuePair to deconstruct

first TKey& First element, the Key

second TValue& Second element, the Value

"},{"location":"xmldocs/llama.grammars.grammar/","title":"Grammar","text":"

Namespace: LLama.Grammars

A grammar is a set of GrammarRules for deciding which characters are valid next. Can be used to constrain output to certain formats - e.g. force the model to output JSON

public sealed class Grammar\n

Inheritance Object \u2192 Grammar

"},{"location":"xmldocs/llama.grammars.grammar/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.grammars.grammar/#startruleindex","title":"StartRuleIndex","text":"

Index of the initial rule to start from

public ulong StartRuleIndex { get; set; }\n
"},{"location":"xmldocs/llama.grammars.grammar/#property-value","title":"Property Value","text":"

UInt64

"},{"location":"xmldocs/llama.grammars.grammar/#rules","title":"Rules","text":"

The rules which make up this grammar

public IReadOnlyList<GrammarRule> Rules { get; }\n
"},{"location":"xmldocs/llama.grammars.grammar/#property-value_1","title":"Property Value","text":"

IReadOnlyList<GrammarRule>

"},{"location":"xmldocs/llama.grammars.grammar/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.grammars.grammar/#grammarireadonlylistgrammarrule-uint64","title":"Grammar(IReadOnlyList<GrammarRule>, UInt64)","text":"

Create a new grammar from a set of rules

public Grammar(IReadOnlyList<GrammarRule> rules, ulong startRuleIndex)\n
"},{"location":"xmldocs/llama.grammars.grammar/#parameters","title":"Parameters","text":"

rules IReadOnlyList<GrammarRule> The rules which make up this grammar

startRuleIndex UInt64 Index of the initial rule to start from

"},{"location":"xmldocs/llama.grammars.grammar/#exceptions","title":"Exceptions","text":"

ArgumentOutOfRangeException

"},{"location":"xmldocs/llama.grammars.grammar/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.grammars.grammar/#createinstance","title":"CreateInstance()","text":"

Create a SafeLLamaGrammarHandle instance to use for parsing

public SafeLLamaGrammarHandle CreateInstance()\n
"},{"location":"xmldocs/llama.grammars.grammar/#returns","title":"Returns","text":"

SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.grammars.grammar/#parsestring-string","title":"Parse(String, String)","text":"

Parse a string of GGML BNF into a Grammar

public static Grammar Parse(string gbnf, string startRule)\n
"},{"location":"xmldocs/llama.grammars.grammar/#parameters_1","title":"Parameters","text":"

gbnf String The string to parse

startRule String Name of the start rule of this grammar

"},{"location":"xmldocs/llama.grammars.grammar/#returns_1","title":"Returns","text":"

Grammar A Grammar which can be converted into a SafeLLamaGrammarHandle for sampling

"},{"location":"xmldocs/llama.grammars.grammar/#exceptions_1","title":"Exceptions","text":"

GrammarFormatException Thrown if input is malformed

"},{"location":"xmldocs/llama.grammars.grammar/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.grammars.grammar/#returns_2","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.grammars.grammarrule/","title":"GrammarRule","text":"

Namespace: LLama.Grammars

A single rule in a Grammar

public sealed class GrammarRule : System.IEquatable`1[[LLama.Grammars.GrammarRule, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 GrammarRule Implements IEquatable<GrammarRule>

"},{"location":"xmldocs/llama.grammars.grammarrule/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.grammars.grammarrule/#name","title":"Name","text":"

Name of this rule

public string Name { get; }\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.grammars.grammarrule/#elements","title":"Elements","text":"

The elements of this grammar rule

public IReadOnlyList<LLamaGrammarElement> Elements { get; }\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#property-value_1","title":"Property Value","text":"

IReadOnlyList<LLamaGrammarElement>

"},{"location":"xmldocs/llama.grammars.grammarrule/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.grammars.grammarrule/#grammarrulestring-ireadonlylistllamagrammarelement","title":"GrammarRule(String, IReadOnlyList<LLamaGrammarElement>)","text":"

Create a new GrammarRule containing the given elements

public GrammarRule(string name, IReadOnlyList<LLamaGrammarElement> elements)\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#parameters","title":"Parameters","text":"

name String

elements IReadOnlyList<LLamaGrammarElement>

"},{"location":"xmldocs/llama.grammars.grammarrule/#exceptions","title":"Exceptions","text":"

ArgumentException

"},{"location":"xmldocs/llama.grammars.grammarrule/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.grammars.grammarrule/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.grammars.grammarrule/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#returns_1","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.grammars.grammarrule/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#parameters_1","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.grammars.grammarrule/#returns_2","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.grammars.grammarrule/#equalsgrammarrule","title":"Equals(GrammarRule)","text":"
public bool Equals(GrammarRule other)\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#parameters_2","title":"Parameters","text":"

other GrammarRule

"},{"location":"xmldocs/llama.grammars.grammarrule/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.grammars.grammarrule/#clone","title":"<Clone>$()","text":"
public GrammarRule <Clone>$()\n
"},{"location":"xmldocs/llama.grammars.grammarrule/#returns_4","title":"Returns","text":"

GrammarRule

"},{"location":"xmldocs/llama.instructexecutor/","title":"InstructExecutor","text":"

Namespace: LLama

The LLama executor for instruct mode.

public class InstructExecutor : StatefulExecutorBase, LLama.Abstractions.ILLamaExecutor\n

Inheritance Object \u2192 StatefulExecutorBase \u2192 InstructExecutor Implements ILLamaExecutor

"},{"location":"xmldocs/llama.instructexecutor/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.instructexecutor/#context","title":"Context","text":"

The context used by the executor.

public LLamaContext Context { get; }\n
"},{"location":"xmldocs/llama.instructexecutor/#property-value","title":"Property Value","text":"

LLamaContext

"},{"location":"xmldocs/llama.instructexecutor/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.instructexecutor/#instructexecutorllamacontext-string-string","title":"InstructExecutor(LLamaContext, String, String)","text":"
public InstructExecutor(LLamaContext context, string instructionPrefix, string instructionSuffix)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters","title":"Parameters","text":"

context LLamaContext

instructionPrefix String

instructionSuffix String

"},{"location":"xmldocs/llama.instructexecutor/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.instructexecutor/#getstatedata","title":"GetStateData()","text":"
public ExecutorBaseState GetStateData()\n
"},{"location":"xmldocs/llama.instructexecutor/#returns","title":"Returns","text":"

ExecutorBaseState

"},{"location":"xmldocs/llama.instructexecutor/#loadstateexecutorbasestate","title":"LoadState(ExecutorBaseState)","text":"
public void LoadState(ExecutorBaseState data)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_1","title":"Parameters","text":"

data ExecutorBaseState

"},{"location":"xmldocs/llama.instructexecutor/#savestatestring","title":"SaveState(String)","text":"
public void SaveState(string filename)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_2","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.instructexecutor/#loadstatestring","title":"LoadState(String)","text":"
public void LoadState(string filename)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_3","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.instructexecutor/#getloopconditioninferstateargs","title":"GetLoopCondition(InferStateArgs)","text":"
protected bool GetLoopCondition(InferStateArgs args)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_4","title":"Parameters","text":"

args InferStateArgs

"},{"location":"xmldocs/llama.instructexecutor/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.instructexecutor/#preprocessinputsstring-inferstateargs","title":"PreprocessInputs(String, InferStateArgs)","text":"
protected void PreprocessInputs(string text, InferStateArgs args)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_5","title":"Parameters","text":"

text String

args InferStateArgs

"},{"location":"xmldocs/llama.instructexecutor/#postprocessiinferenceparams-inferstateargs-ienumerable1","title":"PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)","text":"
protected bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_6","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

extraOutputs IEnumerable`1&

"},{"location":"xmldocs/llama.instructexecutor/#returns_2","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.instructexecutor/#inferinternaliinferenceparams-inferstateargs","title":"InferInternal(IInferenceParams, InferStateArgs)","text":"
protected void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)\n
"},{"location":"xmldocs/llama.instructexecutor/#parameters_7","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

"},{"location":"xmldocs/llama.interactiveexecutor/","title":"InteractiveExecutor","text":"

Namespace: LLama

The LLama executor for interactive mode.

public class InteractiveExecutor : StatefulExecutorBase, LLama.Abstractions.ILLamaExecutor\n

Inheritance Object \u2192 StatefulExecutorBase \u2192 InteractiveExecutor Implements ILLamaExecutor

"},{"location":"xmldocs/llama.interactiveexecutor/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.interactiveexecutor/#context","title":"Context","text":"

The context used by the executor.

public LLamaContext Context { get; }\n
"},{"location":"xmldocs/llama.interactiveexecutor/#property-value","title":"Property Value","text":"

LLamaContext

"},{"location":"xmldocs/llama.interactiveexecutor/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.interactiveexecutor/#interactiveexecutorllamacontext","title":"InteractiveExecutor(LLamaContext)","text":"
public InteractiveExecutor(LLamaContext context)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters","title":"Parameters","text":"

context LLamaContext

"},{"location":"xmldocs/llama.interactiveexecutor/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.interactiveexecutor/#getstatedata","title":"GetStateData()","text":"
public ExecutorBaseState GetStateData()\n
"},{"location":"xmldocs/llama.interactiveexecutor/#returns","title":"Returns","text":"

ExecutorBaseState

"},{"location":"xmldocs/llama.interactiveexecutor/#loadstateexecutorbasestate","title":"LoadState(ExecutorBaseState)","text":"
public void LoadState(ExecutorBaseState data)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_1","title":"Parameters","text":"

data ExecutorBaseState

"},{"location":"xmldocs/llama.interactiveexecutor/#savestatestring","title":"SaveState(String)","text":"
public void SaveState(string filename)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_2","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.interactiveexecutor/#loadstatestring","title":"LoadState(String)","text":"
public void LoadState(string filename)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_3","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.interactiveexecutor/#getloopconditioninferstateargs","title":"GetLoopCondition(InferStateArgs)","text":"

Define whether to continue the loop to generate responses.

protected bool GetLoopCondition(InferStateArgs args)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_4","title":"Parameters","text":"

args InferStateArgs

"},{"location":"xmldocs/llama.interactiveexecutor/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.interactiveexecutor/#preprocessinputsstring-inferstateargs","title":"PreprocessInputs(String, InferStateArgs)","text":"
protected void PreprocessInputs(string text, InferStateArgs args)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_5","title":"Parameters","text":"

text String

args InferStateArgs

"},{"location":"xmldocs/llama.interactiveexecutor/#postprocessiinferenceparams-inferstateargs-ienumerable1","title":"PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)","text":"

Return whether to break the generation.

protected bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_6","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

extraOutputs IEnumerable`1&

"},{"location":"xmldocs/llama.interactiveexecutor/#returns_2","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.interactiveexecutor/#inferinternaliinferenceparams-inferstateargs","title":"InferInternal(IInferenceParams, InferStateArgs)","text":"
protected void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)\n
"},{"location":"xmldocs/llama.interactiveexecutor/#parameters_7","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

"},{"location":"xmldocs/llama.llamacontext/","title":"LLamaContext","text":"

Namespace: LLama

A llama_context, which holds all the context required to interact with a model

public sealed class LLamaContext : System.IDisposable\n

Inheritance Object \u2192 LLamaContext Implements IDisposable

"},{"location":"xmldocs/llama.llamacontext/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.llamacontext/#vocabcount","title":"VocabCount","text":"

Total number of tokens in vocabulary of this model

public int VocabCount { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamacontext/#contextsize","title":"ContextSize","text":"

Total number of tokens in the context

public int ContextSize { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamacontext/#embeddingsize","title":"EmbeddingSize","text":"

Dimension of embedding vectors

public int EmbeddingSize { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamacontext/#params","title":"Params","text":"

The model params set for this model.

public IModelParams Params { get; set; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_3","title":"Property Value","text":"

IModelParams

"},{"location":"xmldocs/llama.llamacontext/#nativehandle","title":"NativeHandle","text":"

The native handle, which is used to be passed to the native APIs

public SafeLLamaContextHandle NativeHandle { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_4","title":"Property Value","text":"

SafeLLamaContextHandle

Remarks:

Be careful how you use this!

"},{"location":"xmldocs/llama.llamacontext/#encoding","title":"Encoding","text":"

The encoding set for this model to deal with text input.

public Encoding Encoding { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_5","title":"Property Value","text":"

Encoding

"},{"location":"xmldocs/llama.llamacontext/#embeddinglength","title":"EmbeddingLength","text":"

The embedding length of the model, also known as n_embed

public int EmbeddingLength { get; }\n
"},{"location":"xmldocs/llama.llamacontext/#property-value_6","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamacontext/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.llamacontext/#llamacontextimodelparams-illamalogger","title":"LLamaContext(IModelParams, ILLamaLogger)","text":""},{"location":"xmldocs/llama.llamacontext/#caution","title":"Caution","text":"

Use the LLamaWeights.CreateContext instead

public LLamaContext(IModelParams params, ILLamaLogger logger)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters","title":"Parameters","text":"

params IModelParams Model params.

logger ILLamaLogger The logger.

"},{"location":"xmldocs/llama.llamacontext/#llamacontextllamaweights-imodelparams-illamalogger","title":"LLamaContext(LLamaWeights, IModelParams, ILLamaLogger)","text":"

Create a new LLamaContext for the given LLamaWeights

public LLamaContext(LLamaWeights model, IModelParams params, ILLamaLogger logger)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_1","title":"Parameters","text":"

model LLamaWeights

params IModelParams

logger ILLamaLogger

"},{"location":"xmldocs/llama.llamacontext/#exceptions","title":"Exceptions","text":"

ObjectDisposedException

"},{"location":"xmldocs/llama.llamacontext/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.llamacontext/#clone","title":"Clone()","text":"

Create a copy of the current state of this context

public LLamaContext Clone()\n
"},{"location":"xmldocs/llama.llamacontext/#returns","title":"Returns","text":"

LLamaContext

"},{"location":"xmldocs/llama.llamacontext/#tokenizestring-boolean","title":"Tokenize(String, Boolean)","text":"

Tokenize a string.

public Int32[] Tokenize(string text, bool addBos)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_2","title":"Parameters","text":"

text String

addBos Boolean Whether to add a bos to the text.

"},{"location":"xmldocs/llama.llamacontext/#returns_1","title":"Returns","text":"

Int32[]

"},{"location":"xmldocs/llama.llamacontext/#detokenizeienumerableint32","title":"DeTokenize(IEnumerable<Int32>)","text":"

Detokenize the tokens to text.

public string DeTokenize(IEnumerable<int> tokens)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_3","title":"Parameters","text":"

tokens IEnumerable<Int32>

"},{"location":"xmldocs/llama.llamacontext/#returns_2","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.llamacontext/#savestatestring","title":"SaveState(String)","text":"

Save the state to specified path.

public void SaveState(string filename)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_4","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.llamacontext/#getstatedata","title":"GetStateData()","text":""},{"location":"xmldocs/llama.llamacontext/#caution_1","title":"Caution","text":"

Use GetState instead, this supports larger states (over 2GB)

Get the state data as a byte array.

public Byte[] GetStateData()\n
"},{"location":"xmldocs/llama.llamacontext/#returns_3","title":"Returns","text":"

Byte[]

"},{"location":"xmldocs/llama.llamacontext/#getstate","title":"GetState()","text":"

Get the state data as an opaque handle

public State GetState()\n
"},{"location":"xmldocs/llama.llamacontext/#returns_4","title":"Returns","text":"

State

"},{"location":"xmldocs/llama.llamacontext/#loadstatestring","title":"LoadState(String)","text":"

Load the state from specified path.

public void LoadState(string filename)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_5","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.llamacontext/#exceptions_1","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#loadstatebyte","title":"LoadState(Byte[])","text":"

Load the state from memory.

public void LoadState(Byte[] stateData)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_6","title":"Parameters","text":"

stateData Byte[]

"},{"location":"xmldocs/llama.llamacontext/#exceptions_2","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#loadstatestate","title":"LoadState(State)","text":"

Load the state from memory.

public void LoadState(State state)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_7","title":"Parameters","text":"

state State

"},{"location":"xmldocs/llama.llamacontext/#exceptions_3","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#samplellamatokendataarray-nullable1-single-mirostattype-single-single-int32-single-single-single-safellamagrammarhandle","title":"Sample(LLamaTokenDataArray, Nullable`1&, Single, MirostatType, Single, Single, Int32, Single, Single, Single, SafeLLamaGrammarHandle)","text":"

Perform the sampling. Please don't use it unless you fully know what it does.

public int Sample(LLamaTokenDataArray candidates, Nullable`1& mirostat_mu, float temperature, MirostatType mirostat, float mirostatTau, float mirostatEta, int topK, float topP, float tfsZ, float typicalP, SafeLLamaGrammarHandle grammar)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_8","title":"Parameters","text":"

candidates LLamaTokenDataArray

mirostat_mu Nullable`1&

temperature Single

mirostat MirostatType

mirostatTau Single

mirostatEta Single

topK Int32

topP Single

tfsZ Single

typicalP Single

grammar SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.llamacontext/#returns_5","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.llamacontext/#applypenaltyienumerableint32-dictionaryint32-single-int32-single-single-single-boolean","title":"ApplyPenalty(IEnumerable<Int32>, Dictionary<Int32, Single>, Int32, Single, Single, Single, Boolean)","text":"

Apply the penalty for the tokens. Please don't use it unless you fully know what it does.

public LLamaTokenDataArray ApplyPenalty(IEnumerable<int> lastTokens, Dictionary<int, float> logitBias, int repeatLastTokensCount, float repeatPenalty, float alphaFrequency, float alphaPresence, bool penalizeNL)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_9","title":"Parameters","text":"

lastTokens IEnumerable<Int32>

logitBias Dictionary<Int32, Single>

repeatLastTokensCount Int32

repeatPenalty Single

alphaFrequency Single

alphaPresence Single

penalizeNL Boolean

"},{"location":"xmldocs/llama.llamacontext/#returns_6","title":"Returns","text":"

LLamaTokenDataArray

"},{"location":"xmldocs/llama.llamacontext/#evalint32-int32","title":"Eval(Int32[], Int32)","text":"
public int Eval(Int32[] tokens, int pastTokensCount)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_10","title":"Parameters","text":"

tokens Int32[]

pastTokensCount Int32

"},{"location":"xmldocs/llama.llamacontext/#returns_7","title":"Returns","text":"

Int32 The updated pastTokensCount.

"},{"location":"xmldocs/llama.llamacontext/#exceptions_4","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#evallistint32-int32","title":"Eval(List<Int32>, Int32)","text":"
public int Eval(List<int> tokens, int pastTokensCount)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_11","title":"Parameters","text":"

tokens List<Int32>

pastTokensCount Int32

"},{"location":"xmldocs/llama.llamacontext/#returns_8","title":"Returns","text":"

Int32 The updated pastTokensCount.

"},{"location":"xmldocs/llama.llamacontext/#exceptions_5","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#evalreadonlymemoryint32-int32","title":"Eval(ReadOnlyMemory<Int32>, Int32)","text":"
public int Eval(ReadOnlyMemory<int> tokens, int pastTokensCount)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_12","title":"Parameters","text":"

tokens ReadOnlyMemory<Int32>

pastTokensCount Int32

"},{"location":"xmldocs/llama.llamacontext/#returns_9","title":"Returns","text":"

Int32 The updated pastTokensCount.

"},{"location":"xmldocs/llama.llamacontext/#exceptions_6","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#evalreadonlyspanint32-int32","title":"Eval(ReadOnlySpan<Int32>, Int32)","text":"
public int Eval(ReadOnlySpan<int> tokens, int pastTokensCount)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_13","title":"Parameters","text":"

tokens ReadOnlySpan<Int32>

pastTokensCount Int32

"},{"location":"xmldocs/llama.llamacontext/#returns_10","title":"Returns","text":"

Int32 The updated pastTokensCount.

"},{"location":"xmldocs/llama.llamacontext/#exceptions_7","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamacontext/#generateresultienumerableint32","title":"GenerateResult(IEnumerable<Int32>)","text":"
internal IEnumerable<string> GenerateResult(IEnumerable<int> ids)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_14","title":"Parameters","text":"

ids IEnumerable<Int32>

"},{"location":"xmldocs/llama.llamacontext/#returns_11","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.llamacontext/#tokentostringint32","title":"TokenToString(Int32)","text":"

Convert a token into a string

public string TokenToString(int token)\n
"},{"location":"xmldocs/llama.llamacontext/#parameters_15","title":"Parameters","text":"

token Int32

"},{"location":"xmldocs/llama.llamacontext/#returns_12","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.llamacontext/#dispose","title":"Dispose()","text":"
public void Dispose()\n
"},{"location":"xmldocs/llama.llamaembedder/","title":"LLamaEmbedder","text":"

Namespace: LLama

The embedder for LLama, which supports getting embeddings from text.

public sealed class LLamaEmbedder : System.IDisposable\n

Inheritance Object \u2192 LLamaEmbedder Implements IDisposable

"},{"location":"xmldocs/llama.llamaembedder/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.llamaembedder/#embeddingsize","title":"EmbeddingSize","text":"

Dimension of embedding vectors

public int EmbeddingSize { get; }\n
"},{"location":"xmldocs/llama.llamaembedder/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamaembedder/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.llamaembedder/#llamaembedderimodelparams","title":"LLamaEmbedder(IModelParams)","text":"
public LLamaEmbedder(IModelParams params)\n
"},{"location":"xmldocs/llama.llamaembedder/#parameters","title":"Parameters","text":"

params IModelParams

"},{"location":"xmldocs/llama.llamaembedder/#llamaembedderllamaweights-imodelparams","title":"LLamaEmbedder(LLamaWeights, IModelParams)","text":"
public LLamaEmbedder(LLamaWeights weights, IModelParams params)\n
"},{"location":"xmldocs/llama.llamaembedder/#parameters_1","title":"Parameters","text":"

weights LLamaWeights

params IModelParams

"},{"location":"xmldocs/llama.llamaembedder/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.llamaembedder/#getembeddingsstring-int32-boolean-string","title":"GetEmbeddings(String, Int32, Boolean, String)","text":""},{"location":"xmldocs/llama.llamaembedder/#caution","title":"Caution","text":"

'threads' and 'encoding' parameters are no longer used

Get the embeddings of the text.

public Single[] GetEmbeddings(string text, int threads, bool addBos, string encoding)\n
"},{"location":"xmldocs/llama.llamaembedder/#parameters_2","title":"Parameters","text":"

text String

threads Int32 unused

addBos Boolean Add bos to the text.

encoding String unused

"},{"location":"xmldocs/llama.llamaembedder/#returns","title":"Returns","text":"

Single[]

"},{"location":"xmldocs/llama.llamaembedder/#exceptions","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamaembedder/#getembeddingsstring","title":"GetEmbeddings(String)","text":"

Get the embeddings of the text.

public Single[] GetEmbeddings(string text)\n
"},{"location":"xmldocs/llama.llamaembedder/#parameters_3","title":"Parameters","text":"

text String

"},{"location":"xmldocs/llama.llamaembedder/#returns_1","title":"Returns","text":"

Single[]

"},{"location":"xmldocs/llama.llamaembedder/#exceptions_1","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamaembedder/#getembeddingsstring-boolean","title":"GetEmbeddings(String, Boolean)","text":"

Get the embeddings of the text.

public Single[] GetEmbeddings(string text, bool addBos)\n
"},{"location":"xmldocs/llama.llamaembedder/#parameters_4","title":"Parameters","text":"

text String

addBos Boolean Add bos to the text.

"},{"location":"xmldocs/llama.llamaembedder/#returns_2","title":"Returns","text":"

Single[]

"},{"location":"xmldocs/llama.llamaembedder/#exceptions_2","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.llamaembedder/#dispose","title":"Dispose()","text":"
public void Dispose()\n
"},{"location":"xmldocs/llama.llamaquantizer/","title":"LLamaQuantizer","text":"

Namespace: LLama

The quantizer to quantize the model.

public static class LLamaQuantizer\n

Inheritance Object \u2192 LLamaQuantizer

"},{"location":"xmldocs/llama.llamaquantizer/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.llamaquantizer/#quantizestring-string-llamaftype-int32-boolean-boolean","title":"Quantize(String, String, LLamaFtype, Int32, Boolean, Boolean)","text":"

Quantize the model.

public static bool Quantize(string srcFileName, string dstFilename, LLamaFtype ftype, int nthread, bool allowRequantize, bool quantizeOutputTensor)\n
"},{"location":"xmldocs/llama.llamaquantizer/#parameters","title":"Parameters","text":"

srcFileName String The model file to be quantized.

dstFilename String The path to save the quantized model.

ftype LLamaFtype The type of quantization.

nthread Int32 Thread to be used during the quantization. By default it's the physical core number.

allowRequantize Boolean

quantizeOutputTensor Boolean

"},{"location":"xmldocs/llama.llamaquantizer/#returns","title":"Returns","text":"

Boolean Whether the quantization is successful.

"},{"location":"xmldocs/llama.llamaquantizer/#exceptions","title":"Exceptions","text":"

ArgumentException

"},{"location":"xmldocs/llama.llamaquantizer/#quantizestring-string-string-int32-boolean-boolean","title":"Quantize(String, String, String, Int32, Boolean, Boolean)","text":"

Quantize the model.

public static bool Quantize(string srcFileName, string dstFilename, string ftype, int nthread, bool allowRequantize, bool quantizeOutputTensor)\n
"},{"location":"xmldocs/llama.llamaquantizer/#parameters_1","title":"Parameters","text":"

srcFileName String The model file to be quantized.

dstFilename String The path to save the quantized model.

ftype String The type of quantization.

nthread Int32 Thread to be used during the quantization. By default it's the physical core number.

allowRequantize Boolean

quantizeOutputTensor Boolean

"},{"location":"xmldocs/llama.llamaquantizer/#returns_1","title":"Returns","text":"

Boolean Whether the quantization is successful.

"},{"location":"xmldocs/llama.llamaquantizer/#exceptions_1","title":"Exceptions","text":"

ArgumentException

"},{"location":"xmldocs/llama.llamatransforms/","title":"LLamaTransforms","text":"

Namespace: LLama

A class that contains all the transforms provided internally by LLama.

public class LLamaTransforms\n

Inheritance Object \u2192 LLamaTransforms

"},{"location":"xmldocs/llama.llamatransforms/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.llamatransforms/#llamatransforms_1","title":"LLamaTransforms()","text":"
public LLamaTransforms()\n
"},{"location":"xmldocs/llama.llamaweights/","title":"LLamaWeights","text":"

Namespace: LLama

A set of model weights, loaded into memory.

public sealed class LLamaWeights : System.IDisposable\n

Inheritance Object \u2192 LLamaWeights Implements IDisposable

"},{"location":"xmldocs/llama.llamaweights/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.llamaweights/#nativehandle","title":"NativeHandle","text":"

The native handle, which is used in the native APIs

public SafeLlamaModelHandle NativeHandle { get; }\n
"},{"location":"xmldocs/llama.llamaweights/#property-value","title":"Property Value","text":"

SafeLlamaModelHandle

Remarks:

Be careful how you use this!

"},{"location":"xmldocs/llama.llamaweights/#encoding","title":"Encoding","text":"

Encoding to use to convert text into bytes for the model

public Encoding Encoding { get; }\n
"},{"location":"xmldocs/llama.llamaweights/#property-value_1","title":"Property Value","text":"

Encoding

"},{"location":"xmldocs/llama.llamaweights/#vocabcount","title":"VocabCount","text":"

Total number of tokens in vocabulary of this model

public int VocabCount { get; }\n
"},{"location":"xmldocs/llama.llamaweights/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamaweights/#contextsize","title":"ContextSize","text":"

Total number of tokens in the context

public int ContextSize { get; }\n
"},{"location":"xmldocs/llama.llamaweights/#property-value_3","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamaweights/#embeddingsize","title":"EmbeddingSize","text":"

Dimension of embedding vectors

public int EmbeddingSize { get; }\n
"},{"location":"xmldocs/llama.llamaweights/#property-value_4","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.llamaweights/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.llamaweights/#loadfromfileimodelparams","title":"LoadFromFile(IModelParams)","text":"

Load weights into memory

public static LLamaWeights LoadFromFile(IModelParams params)\n
"},{"location":"xmldocs/llama.llamaweights/#parameters","title":"Parameters","text":"

params IModelParams

"},{"location":"xmldocs/llama.llamaweights/#returns","title":"Returns","text":"

LLamaWeights

"},{"location":"xmldocs/llama.llamaweights/#dispose","title":"Dispose()","text":"
public void Dispose()\n
"},{"location":"xmldocs/llama.llamaweights/#createcontextimodelparams","title":"CreateContext(IModelParams)","text":"

Create a llama_context using this model

public LLamaContext CreateContext(IModelParams params)\n
"},{"location":"xmldocs/llama.llamaweights/#parameters_1","title":"Parameters","text":"

params IModelParams

"},{"location":"xmldocs/llama.llamaweights/#returns_1","title":"Returns","text":"

LLamaContext

"},{"location":"xmldocs/llama.native.llamacontextparams/","title":"LLamaContextParams","text":"

Namespace: LLama.Native

A C# representation of the llama.cpp llama_context_params struct

public struct LLamaContextParams\n

Inheritance Object \u2192 ValueType \u2192 LLamaContextParams

"},{"location":"xmldocs/llama.native.llamacontextparams/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamacontextparams/#seed","title":"seed","text":"

RNG seed, -1 for random

public int seed;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#n_ctx","title":"n_ctx","text":"

text context

public int n_ctx;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#n_batch","title":"n_batch","text":"

prompt processing batch size

public int n_batch;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#n_gpu_layers","title":"n_gpu_layers","text":"

number of layers to store in VRAM

public int n_gpu_layers;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#main_gpu","title":"main_gpu","text":"

the GPU that is used for scratch and small tensors

public int main_gpu;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#tensor_split","title":"tensor_split","text":"

how to split layers across multiple GPUs

public IntPtr tensor_split;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#rope_freq_base","title":"rope_freq_base","text":"

ref: https://github.com/ggerganov/llama.cpp/pull/2054 RoPE base frequency

public float rope_freq_base;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#rope_freq_scale","title":"rope_freq_scale","text":"

ref: https://github.com/ggerganov/llama.cpp/pull/2054 RoPE frequency scaling factor

public float rope_freq_scale;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#progress_callback","title":"progress_callback","text":"

called with a progress value between 0 and 1, pass NULL to disable

public IntPtr progress_callback;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#progress_callback_user_data","title":"progress_callback_user_data","text":"

context pointer passed to the progress callback

public IntPtr progress_callback_user_data;\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.llamacontextparams/#low_vram","title":"low_vram","text":"

if true, reduce VRAM usage at the cost of performance

public bool low_vram { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#mul_mat_q","title":"mul_mat_q","text":"

if true, use experimental mul_mat_q kernels

public bool mul_mat_q { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_1","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#f16_kv","title":"f16_kv","text":"

use fp16 for KV cache

public bool f16_kv { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_2","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#logits_all","title":"logits_all","text":"

the llama_eval() call computes all logits, not just the last one

public bool logits_all { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_3","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#vocab_only","title":"vocab_only","text":"

only load the vocabulary, no weights

public bool vocab_only { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_4","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#use_mmap","title":"use_mmap","text":"

use mmap if possible

public bool use_mmap { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_5","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#use_mlock","title":"use_mlock","text":"

force system to keep model in RAM

public bool use_mlock { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_6","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamacontextparams/#embedding","title":"embedding","text":"

embedding mode only

public bool embedding { get; set; }\n
"},{"location":"xmldocs/llama.native.llamacontextparams/#property-value_7","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamaftype/","title":"LLamaFtype","text":"

Namespace: LLama.Native

Supported model file types

public enum LLamaFtype\n

Inheritance Object \u2192 ValueType \u2192 Enum \u2192 LLamaFtype Implements IComparable, IFormattable, IConvertible

"},{"location":"xmldocs/llama.native.llamaftype/#fields","title":"Fields","text":"Name Value Description LLAMA_FTYPE_ALL_F32 0 All f32 LLAMA_FTYPE_MOSTLY_F16 1 Mostly f16 LLAMA_FTYPE_MOSTLY_Q8_0 7 Mostly 8 bit LLAMA_FTYPE_MOSTLY_Q4_0 2 Mostly 4 bit LLAMA_FTYPE_MOSTLY_Q4_1 3 Mostly 4 bit LLAMA_FTYPE_MOSTLY_Q4_1_SOME_F16 4 Mostly 4 bit, tok_embeddings.weight and output.weight are f16 LLAMA_FTYPE_MOSTLY_Q5_0 8 Mostly 5 bit LLAMA_FTYPE_MOSTLY_Q5_1 9 Mostly 5 bit LLAMA_FTYPE_MOSTLY_Q2_K 10 K-Quant 2 bit LLAMA_FTYPE_MOSTLY_Q3_K_S 11 K-Quant 3 bit (Small) LLAMA_FTYPE_MOSTLY_Q3_K_M 12 K-Quant 3 bit (Medium) LLAMA_FTYPE_MOSTLY_Q3_K_L 13 K-Quant 3 bit (Large) LLAMA_FTYPE_MOSTLY_Q4_K_S 14 K-Quant 4 bit (Small) LLAMA_FTYPE_MOSTLY_Q4_K_M 15 K-Quant 4 bit (Medium) LLAMA_FTYPE_MOSTLY_Q5_K_S 16 K-Quant 5 bit (Small) LLAMA_FTYPE_MOSTLY_Q5_K_M 17 K-Quant 5 bit (Medium) LLAMA_FTYPE_MOSTLY_Q6_K 18 K-Quant 6 bit LLAMA_FTYPE_GUESSED 1024 File type was not specified"},{"location":"xmldocs/llama.native.llamagrammarelement/","title":"LLamaGrammarElement","text":"

Namespace: LLama.Native

An element of a grammar

public struct LLamaGrammarElement\n

Inheritance Object \u2192 ValueType \u2192 LLamaGrammarElement Implements IEquatable<LLamaGrammarElement>

"},{"location":"xmldocs/llama.native.llamagrammarelement/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamagrammarelement/#type","title":"Type","text":"

The type of this element

public LLamaGrammarElementType Type;\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#value","title":"Value","text":"

Unicode code point or rule ID

public uint Value;\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.llamagrammarelement/#llamagrammarelementllamagrammarelementtype-uint32","title":"LLamaGrammarElement(LLamaGrammarElementType, UInt32)","text":"

Construct a new LLamaGrammarElement

LLamaGrammarElement(LLamaGrammarElementType type, uint value)\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#parameters","title":"Parameters","text":"

type LLamaGrammarElementType

value UInt32

"},{"location":"xmldocs/llama.native.llamagrammarelement/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.llamagrammarelement/#equalsllamagrammarelement","title":"Equals(LLamaGrammarElement)","text":"
bool Equals(LLamaGrammarElement other)\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#parameters_1","title":"Parameters","text":"

other LLamaGrammarElement

"},{"location":"xmldocs/llama.native.llamagrammarelement/#returns","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamagrammarelement/#equalsobject","title":"Equals(Object)","text":"
bool Equals(object obj)\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.native.llamagrammarelement/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamagrammarelement/#gethashcode","title":"GetHashCode()","text":"
int GetHashCode()\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.llamagrammarelement/#ischarelement","title":"IsCharElement()","text":"
bool IsCharElement()\n
"},{"location":"xmldocs/llama.native.llamagrammarelement/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamagrammarelementtype/","title":"LLamaGrammarElementType","text":"

Namespace: LLama.Native

grammar element type

public enum LLamaGrammarElementType\n

Inheritance Object \u2192 ValueType \u2192 Enum \u2192 LLamaGrammarElementType Implements IComparable, IFormattable, IConvertible

"},{"location":"xmldocs/llama.native.llamagrammarelementtype/#fields","title":"Fields","text":"Name Value Description END 0 end of rule definition ALT 1 start of alternate definition for rule RULE_REF 2 non-terminal element: reference to rule CHAR 3 terminal element: character (code point) CHAR_NOT 4 inverse char(s) ([^a], [^a-b] [^abc]) CHAR_RNG_UPPER 5 modifies a preceding CHAR or CHAR_ALT to be an inclusive range ([a-z]) CHAR_ALT 6 modifies a preceding CHAR or CHAR_RNG_UPPER to add an alternate char to match ([ab], [a-zA])"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/","title":"LLamaModelQuantizeParams","text":"

Namespace: LLama.Native

Quantizer parameters used in the native API

public struct LLamaModelQuantizeParams\n

Inheritance Object \u2192 ValueType \u2192 LLamaModelQuantizeParams

"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#nthread","title":"nthread","text":"

number of threads to use for quantizing, if <=0 will use std::thread::hardware_concurrency()

public int nthread;\n
"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#ftype","title":"ftype","text":"

quantize to this llama_ftype

public LLamaFtype ftype;\n
"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#allow_requantize","title":"allow_requantize","text":"

allow quantizing non-f32/f16 tensors

public bool allow_requantize { get; set; }\n
"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#property-value","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#quantize_output_tensor","title":"quantize_output_tensor","text":"

quantize output.weight

public bool quantize_output_tensor { get; set; }\n
"},{"location":"xmldocs/llama.native.llamamodelquantizeparams/#property-value_1","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamatokendata/","title":"LLamaTokenData","text":"

Namespace: LLama.Native

public struct LLamaTokenData\n

Inheritance Object \u2192 ValueType \u2192 LLamaTokenData

"},{"location":"xmldocs/llama.native.llamatokendata/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamatokendata/#id","title":"id","text":"

token id

public int id;\n
"},{"location":"xmldocs/llama.native.llamatokendata/#logit","title":"logit","text":"

log-odds of the token

public float logit;\n
"},{"location":"xmldocs/llama.native.llamatokendata/#p","title":"p","text":"

probability of the token

public float p;\n
"},{"location":"xmldocs/llama.native.llamatokendata/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.llamatokendata/#llamatokendataint32-single-single","title":"LLamaTokenData(Int32, Single, Single)","text":"
LLamaTokenData(int id, float logit, float p)\n
"},{"location":"xmldocs/llama.native.llamatokendata/#parameters","title":"Parameters","text":"

id Int32

logit Single

p Single

"},{"location":"xmldocs/llama.native.llamatokendataarray/","title":"LLamaTokenDataArray","text":"

Namespace: LLama.Native

Contains an array of LLamaTokenData, potentially sorted.

public struct LLamaTokenDataArray\n

Inheritance Object \u2192 ValueType \u2192 LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.llamatokendataarray/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamatokendataarray/#data","title":"data","text":"

The LLamaTokenData

public Memory<LLamaTokenData> data;\n
"},{"location":"xmldocs/llama.native.llamatokendataarray/#sorted","title":"sorted","text":"

Indicates if data is sorted by logits in descending order. If this is false the token data is in no particular order.

public bool sorted;\n
"},{"location":"xmldocs/llama.native.llamatokendataarray/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.llamatokendataarray/#llamatokendataarraymemoryllamatokendata-boolean","title":"LLamaTokenDataArray(Memory<LLamaTokenData>, Boolean)","text":"

Create a new LLamaTokenDataArray

LLamaTokenDataArray(Memory<LLamaTokenData> tokens, bool isSorted)\n
"},{"location":"xmldocs/llama.native.llamatokendataarray/#parameters","title":"Parameters","text":"

tokens Memory<LLamaTokenData>

isSorted Boolean

"},{"location":"xmldocs/llama.native.llamatokendataarray/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.llamatokendataarray/#createreadonlyspansingle","title":"Create(ReadOnlySpan<Single>)","text":"

Create a new LLamaTokenDataArray, copying the data from the given logits

LLamaTokenDataArray Create(ReadOnlySpan<float> logits)\n
"},{"location":"xmldocs/llama.native.llamatokendataarray/#parameters_1","title":"Parameters","text":"

logits ReadOnlySpan<Single>

"},{"location":"xmldocs/llama.native.llamatokendataarray/#returns","title":"Returns","text":"

LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.llamatokendataarraynative/","title":"LLamaTokenDataArrayNative","text":"

Namespace: LLama.Native

Contains a pointer to an array of LLamaTokenData which is pinned in memory.

public struct LLamaTokenDataArrayNative\n

Inheritance Object \u2192 ValueType \u2192 LLamaTokenDataArrayNative

"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.native.llamatokendataarraynative/#data","title":"data","text":"

A pointer to an array of LlamaTokenData

public IntPtr data;\n

Remarks:

Memory must be pinned in place for all the time this LLamaTokenDataArrayNative is in use

"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#size","title":"size","text":"

Number of LLamaTokenData in the array

public ulong size;\n
"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.llamatokendataarraynative/#sorted","title":"sorted","text":"

Indicates if the items in the array are sorted

public bool sorted { get; set; }\n
"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#property-value","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.llamatokendataarraynative/#createllamatokendataarray-llamatokendataarraynative","title":"Create(LLamaTokenDataArray, LLamaTokenDataArrayNative&)","text":"

Create a new LLamaTokenDataArrayNative around the data in the LLamaTokenDataArray

MemoryHandle Create(LLamaTokenDataArray array, LLamaTokenDataArrayNative& native)\n
"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#parameters","title":"Parameters","text":"

array LLamaTokenDataArray Data source

native LLamaTokenDataArrayNative& Created native array

"},{"location":"xmldocs/llama.native.llamatokendataarraynative/#returns","title":"Returns","text":"

MemoryHandle A memory handle, pinning the data in place until disposed

"},{"location":"xmldocs/llama.native.nativeapi/","title":"NativeApi","text":"

Namespace: LLama.Native

Direct translation of the llama.cpp API

public class NativeApi\n

Inheritance Object \u2192 NativeApi

"},{"location":"xmldocs/llama.native.nativeapi/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.nativeapi/#nativeapi_1","title":"NativeApi()","text":"
public NativeApi()\n
"},{"location":"xmldocs/llama.native.nativeapi/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_token_mirostatsafellamacontexthandle-llamatokendataarraynative-single-single-int32-single","title":"llama_sample_token_mirostat(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, Single, Int32, Single&)","text":"

Mirostat 1.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

public static int llama_sample_token_mirostat(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float tau, float eta, int m, Single& mu)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& A vector of llama_token_data containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

tau Single The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

eta Single The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

m Int32 The number of tokens considered in the estimation of s_hat. This is an arbitrary value that is used to calculate s_hat, which in turn helps to calculate the value of k. In the paper, they use m = 100, but you can experiment with different values to see how it affects the performance of the algorithm.

mu Single& Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

"},{"location":"xmldocs/llama.native.nativeapi/#returns","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_token_mirostat_v2safellamacontexthandle-llamatokendataarraynative-single-single-single","title":"llama_sample_token_mirostat_v2(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, Single, Single&)","text":"

Mirostat 2.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

public static int llama_sample_token_mirostat_v2(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float tau, float eta, Single& mu)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_1","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& A vector of llama_token_data containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

tau Single The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

eta Single The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

mu Single& Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

"},{"location":"xmldocs/llama.native.nativeapi/#returns_1","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_token_greedysafellamacontexthandle-llamatokendataarraynative","title":"llama_sample_token_greedy(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)","text":"

Selects the token with the highest probability.

public static int llama_sample_token_greedy(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_2","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.nativeapi/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_tokensafellamacontexthandle-llamatokendataarraynative","title":"llama_sample_token(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)","text":"

Randomly selects a token from the candidates based on their probabilities.

public static int llama_sample_token(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_3","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.nativeapi/#returns_3","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_token_to_strsafellamacontexthandle-int32","title":"llama_token_to_str(SafeLLamaContextHandle, Int32)","text":"

Token Id -> String. Uses the vocabulary in the provided context

public static IntPtr llama_token_to_str(SafeLLamaContextHandle ctx, int token)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_4","title":"Parameters","text":"

ctx SafeLLamaContextHandle

token Int32

"},{"location":"xmldocs/llama.native.nativeapi/#returns_4","title":"Returns","text":"

IntPtr Pointer to a string.

"},{"location":"xmldocs/llama.native.nativeapi/#llama_token_bossafellamacontexthandle","title":"llama_token_bos(SafeLLamaContextHandle)","text":"

Get the \"Beginning of sentence\" token

public static int llama_token_bos(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_5","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_5","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_token_eossafellamacontexthandle","title":"llama_token_eos(SafeLLamaContextHandle)","text":"

Get the \"End of sentence\" token

public static int llama_token_eos(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_6","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_6","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_token_nlsafellamacontexthandle","title":"llama_token_nl(SafeLLamaContextHandle)","text":"

Get the \"new line\" token

public static int llama_token_nl(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_7","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_7","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_print_timingssafellamacontexthandle","title":"llama_print_timings(SafeLLamaContextHandle)","text":"

Print out timing information for this context

public static void llama_print_timings(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_8","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#llama_reset_timingssafellamacontexthandle","title":"llama_reset_timings(SafeLLamaContextHandle)","text":"

Reset all collected timing information for this context

public static void llama_reset_timings(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_9","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#llama_print_system_info","title":"llama_print_system_info()","text":"

Print system information

public static IntPtr llama_print_system_info()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_8","title":"Returns","text":"

IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_n_vocabsafellamamodelhandle","title":"llama_model_n_vocab(SafeLlamaModelHandle)","text":"

Get the number of tokens in the model vocabulary

public static int llama_model_n_vocab(SafeLlamaModelHandle model)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_10","title":"Parameters","text":"

model SafeLlamaModelHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_9","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_n_ctxsafellamamodelhandle","title":"llama_model_n_ctx(SafeLlamaModelHandle)","text":"

Get the size of the context window for the model

public static int llama_model_n_ctx(SafeLlamaModelHandle model)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_11","title":"Parameters","text":"

model SafeLlamaModelHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_10","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_n_embdsafellamamodelhandle","title":"llama_model_n_embd(SafeLlamaModelHandle)","text":"

Get the dimension of embedding vectors from this model

public static int llama_model_n_embd(SafeLlamaModelHandle model)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_12","title":"Parameters","text":"

model SafeLlamaModelHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_11","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_token_to_piece_with_modelsafellamamodelhandle-int32-byte-int32","title":"llama_token_to_piece_with_model(SafeLlamaModelHandle, Int32, Byte*, Int32)","text":"

Convert a single token into text

public static int llama_token_to_piece_with_model(SafeLlamaModelHandle model, int llamaToken, Byte* buffer, int length)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_13","title":"Parameters","text":"

model SafeLlamaModelHandle

llamaToken Int32

buffer Byte* buffer to write string into

length Int32 size of the buffer

"},{"location":"xmldocs/llama.native.nativeapi/#returns_12","title":"Returns","text":"

Int32 The length writte, or if the buffer is too small a negative that indicates the length required

"},{"location":"xmldocs/llama.native.nativeapi/#llama_tokenize_with_modelsafellamamodelhandle-byte-int32-int32-boolean","title":"llama_tokenize_with_model(SafeLlamaModelHandle, Byte, Int32, Int32, Boolean)","text":"

Convert text into tokens

public static int llama_tokenize_with_model(SafeLlamaModelHandle model, Byte* text, Int32* tokens, int n_max_tokens, bool add_bos)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_14","title":"Parameters","text":"

model SafeLlamaModelHandle

text Byte*

tokens Int32*

n_max_tokens Int32

add_bos Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#returns_13","title":"Returns","text":"

Int32 Returns the number of tokens on success, no more than n_max_tokens. Returns a negative number on failure - the number of tokens that would have been returned

"},{"location":"xmldocs/llama.native.nativeapi/#llama_log_setllamalogcallback","title":"llama_log_set(LLamaLogCallback)","text":"

Register a callback to receive llama log messages

public static void llama_log_set(LLamaLogCallback logCallback)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_15","title":"Parameters","text":"

logCallback LLamaLogCallback

"},{"location":"xmldocs/llama.native.nativeapi/#llama_grammar_initllamagrammarelement-uint64-uint64","title":"llama_grammar_init(LLamaGrammarElement, UInt64, UInt64)**","text":"

Create a new grammar from the given set of grammar rules

public static IntPtr llama_grammar_init(LLamaGrammarElement** rules, ulong n_rules, ulong start_rule_index)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_16","title":"Parameters","text":"

rules LLamaGrammarElement**

n_rules UInt64

start_rule_index UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#returns_14","title":"Returns","text":"

IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_grammar_freeintptr","title":"llama_grammar_free(IntPtr)","text":"

Free all memory from the given SafeLLamaGrammarHandle

public static void llama_grammar_free(IntPtr grammar)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_17","title":"Parameters","text":"

grammar IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_grammarsafellamacontexthandle-llamatokendataarraynative-safellamagrammarhandle","title":"llama_sample_grammar(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, SafeLLamaGrammarHandle)","text":"

Apply constraints from grammar

public static void llama_sample_grammar(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, SafeLLamaGrammarHandle grammar)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_18","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative&

grammar SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.native.nativeapi/#llama_grammar_accept_tokensafellamacontexthandle-safellamagrammarhandle-int32","title":"llama_grammar_accept_token(SafeLLamaContextHandle, SafeLLamaGrammarHandle, Int32)","text":"

Accepts the sampled token into the grammar

public static void llama_grammar_accept_token(SafeLLamaContextHandle ctx, SafeLLamaGrammarHandle grammar, int token)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_19","title":"Parameters","text":"

ctx SafeLLamaContextHandle

grammar SafeLLamaGrammarHandle

token Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_quantizestring-string-llamamodelquantizeparams","title":"llama_model_quantize(String, String, LLamaModelQuantizeParams*)","text":"

Returns 0 on success

public static int llama_model_quantize(string fname_inp, string fname_out, LLamaModelQuantizeParams* param)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_20","title":"Parameters","text":"

fname_inp String

fname_out String

param LLamaModelQuantizeParams*

"},{"location":"xmldocs/llama.native.nativeapi/#returns_15","title":"Returns","text":"

Int32 Returns 0 on success

Remarks:

not great API - very likely to change

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_classifier_free_guidancesafellamacontexthandle-llamatokendataarraynative-safellamacontexthandle-single","title":"llama_sample_classifier_free_guidance(SafeLLamaContextHandle, LLamaTokenDataArrayNative, SafeLLamaContextHandle, Single)","text":"

Apply classifier-free guidance to the logits as described in academic paper \"Stay on topic with Classifier-Free Guidance\" https://arxiv.org/abs/2306.17806

public static void llama_sample_classifier_free_guidance(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative candidates, SafeLLamaContextHandle guidanceCtx, float scale)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_21","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative A vector of llama_token_data containing the candidate tokens, the logits must be directly extracted from the original generation context without being sorted.

guidanceCtx SafeLLamaContextHandle A separate context from the same model. Other than a negative prompt at the beginning, it should have all generated and user input tokens copied from the main context.

scale Single Guidance strength. 1.0f means no guidance. Higher values mean stronger guidance.

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_repetition_penaltysafellamacontexthandle-llamatokendataarraynative-int32-uint64-single","title":"llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32*, UInt64, Single)","text":"

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, Int32* last_tokens, ulong last_tokens_size, float penalty)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_22","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

last_tokens Int32*

last_tokens_size UInt64

penalty Single

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_frequency_and_presence_penaltiessafellamacontexthandle-llamatokendataarraynative-int32-uint64-single-single","title":"llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32*, UInt64, Single, Single)","text":"

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, Int32* last_tokens, ulong last_tokens_size, float alpha_frequency, float alpha_presence)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_23","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

last_tokens Int32*

last_tokens_size UInt64

alpha_frequency Single

alpha_presence Single

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_classifier_free_guidancesafellamacontexthandle-llamatokendataarraynative-safellamacontexthandle-single_1","title":"llama_sample_classifier_free_guidance(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, SafeLLamaContextHandle, Single)","text":"

Apply classifier-free guidance to the logits as described in academic paper \"Stay on topic with Classifier-Free Guidance\" https://arxiv.org/abs/2306.17806

public static void llama_sample_classifier_free_guidance(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, SafeLLamaContextHandle guidance_ctx, float scale)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_24","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& A vector of llama_token_data containing the candidate tokens, the logits must be directly extracted from the original generation context without being sorted.

guidance_ctx SafeLLamaContextHandle A separate context from the same model. Other than a negative prompt at the beginning, it should have all generated and user input tokens copied from the main context.

scale Single Guidance strength. 1.0f means no guidance. Higher values mean stronger guidance.

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_softmaxsafellamacontexthandle-llamatokendataarraynative","title":"llama_sample_softmax(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)","text":"

Sorts candidate tokens by their logits in descending order and calculate probabilities based on logits.

public static void llama_sample_softmax(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_25","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_top_ksafellamacontexthandle-llamatokendataarraynative-int32-uint64","title":"llama_sample_top_k(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32, UInt64)","text":"

Top-K sampling described in academic paper \"The Curious Case of Neural Text Degeneration\" https://arxiv.org/abs/1904.09751

public static void llama_sample_top_k(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, int k, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_26","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

k Int32

min_keep UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_top_psafellamacontexthandle-llamatokendataarraynative-single-uint64","title":"llama_sample_top_p(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)","text":"

Nucleus sampling described in academic paper \"The Curious Case of Neural Text Degeneration\" https://arxiv.org/abs/1904.09751

public static void llama_sample_top_p(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float p, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_27","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

p Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_tail_freesafellamacontexthandle-llamatokendataarraynative-single-uint64","title":"llama_sample_tail_free(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)","text":"

Tail Free Sampling described in https://www.trentonbricken.com/Tail-Free-Sampling/.

public static void llama_sample_tail_free(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float z, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_28","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

z Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_typicalsafellamacontexthandle-llamatokendataarraynative-single-uint64","title":"llama_sample_typical(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)","text":"

Locally Typical Sampling implementation described in the paper https://arxiv.org/abs/2202.00666.

public static void llama_sample_typical(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float p, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_29","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative& Pointer to LLamaTokenDataArray

p Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#llama_sample_temperaturesafellamacontexthandle-llamatokendataarraynative-single","title":"llama_sample_temperature(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single)","text":"

Modify logits by temperature

public static void llama_sample_temperature(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float temp)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_30","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArrayNative&

temp Single

"},{"location":"xmldocs/llama.native.nativeapi/#llama_empty_call","title":"llama_empty_call()","text":"

A method that does nothing. This is a native method, calling it will force the llama native dependencies to be loaded.

public static bool llama_empty_call()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_16","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_context_default_params","title":"llama_context_default_params()","text":"

Create a LLamaContextParams with default values

public static LLamaContextParams llama_context_default_params()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_17","title":"Returns","text":"

LLamaContextParams

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_quantize_default_params","title":"llama_model_quantize_default_params()","text":"

Create a LLamaModelQuantizeParams with default values

public static LLamaModelQuantizeParams llama_model_quantize_default_params()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_18","title":"Returns","text":"

LLamaModelQuantizeParams

"},{"location":"xmldocs/llama.native.nativeapi/#llama_mmap_supported","title":"llama_mmap_supported()","text":"

Check if memory mapping is supported

public static bool llama_mmap_supported()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_19","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_mlock_supported","title":"llama_mlock_supported()","text":"

Check if memory lockingis supported

public static bool llama_mlock_supported()\n
"},{"location":"xmldocs/llama.native.nativeapi/#returns_20","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_eval_exportsafellamacontexthandle-string","title":"llama_eval_export(SafeLLamaContextHandle, String)","text":"

Export a static computation graph for context of 511 and batch size of 1 NOTE: since this functionality is mostly for debugging and demonstration purposes, we hardcode these parameters here to keep things simple IMPORTANT: do not use for anything else other than debugging and testing!

public static int llama_eval_export(SafeLLamaContextHandle ctx, string fname)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_31","title":"Parameters","text":"

ctx SafeLLamaContextHandle

fname String

"},{"location":"xmldocs/llama.native.nativeapi/#returns_21","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_load_model_from_filestring-llamacontextparams","title":"llama_load_model_from_file(String, LLamaContextParams)","text":"

Various functions for loading a ggml llama model. Allocate (almost) all memory needed for the model. Return NULL on failure

public static IntPtr llama_load_model_from_file(string path_model, LLamaContextParams params)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_32","title":"Parameters","text":"

path_model String

params LLamaContextParams

"},{"location":"xmldocs/llama.native.nativeapi/#returns_22","title":"Returns","text":"

IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_new_context_with_modelsafellamamodelhandle-llamacontextparams","title":"llama_new_context_with_model(SafeLlamaModelHandle, LLamaContextParams)","text":"

Create a new llama_context with the given model. Return value should always be wrapped in SafeLLamaContextHandle!

public static IntPtr llama_new_context_with_model(SafeLlamaModelHandle model, LLamaContextParams params)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_33","title":"Parameters","text":"

model SafeLlamaModelHandle

params LLamaContextParams

"},{"location":"xmldocs/llama.native.nativeapi/#returns_23","title":"Returns","text":"

IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_backend_initboolean","title":"llama_backend_init(Boolean)","text":"

not great API - very likely to change. Initialize the llama + ggml backend Call once at the start of the program

public static void llama_backend_init(bool numa)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_34","title":"Parameters","text":"

numa Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_freeintptr","title":"llama_free(IntPtr)","text":"

Frees all allocated memory in the given llama_context

public static void llama_free(IntPtr ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_35","title":"Parameters","text":"

ctx IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_free_modelintptr","title":"llama_free_model(IntPtr)","text":"

Frees all allocated memory associated with a model

public static void llama_free_model(IntPtr model)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_36","title":"Parameters","text":"

model IntPtr

"},{"location":"xmldocs/llama.native.nativeapi/#llama_model_apply_lora_from_filesafellamamodelhandle-string-string-int32","title":"llama_model_apply_lora_from_file(SafeLlamaModelHandle, String, String, Int32)","text":"

Apply a LoRA adapter to a loaded model path_base_model is the path to a higher quality model to use as a base for the layers modified by the adapter. Can be NULL to use the current loaded model. The model needs to be reloaded before applying a new adapter, otherwise the adapter will be applied on top of the previous one

public static int llama_model_apply_lora_from_file(SafeLlamaModelHandle model_ptr, string path_lora, string path_base_model, int n_threads)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_37","title":"Parameters","text":"

model_ptr SafeLlamaModelHandle

path_lora String

path_base_model String

n_threads Int32

"},{"location":"xmldocs/llama.native.nativeapi/#returns_24","title":"Returns","text":"

Int32 Returns 0 on success

"},{"location":"xmldocs/llama.native.nativeapi/#llama_get_kv_cache_token_countsafellamacontexthandle","title":"llama_get_kv_cache_token_count(SafeLLamaContextHandle)","text":"

Returns the number of tokens in the KV cache

public static int llama_get_kv_cache_token_count(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_38","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_25","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_set_rng_seedsafellamacontexthandle-int32","title":"llama_set_rng_seed(SafeLLamaContextHandle, Int32)","text":"

Sets the current rng seed.

public static void llama_set_rng_seed(SafeLLamaContextHandle ctx, int seed)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_39","title":"Parameters","text":"

ctx SafeLLamaContextHandle

seed Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_get_state_sizesafellamacontexthandle","title":"llama_get_state_size(SafeLLamaContextHandle)","text":"

Returns the maximum size in bytes of the state (rng, logits, embedding and kv_cache) - will often be smaller after compacting tokens

public static ulong llama_get_state_size(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_40","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_26","title":"Returns","text":"

UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#llama_copy_state_datasafellamacontexthandle-byte","title":"llama_copy_state_data(SafeLLamaContextHandle, Byte*)","text":"

Copies the state to the specified destination address. Destination needs to have allocated enough memory.

public static ulong llama_copy_state_data(SafeLLamaContextHandle ctx, Byte* dest)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_41","title":"Parameters","text":"

ctx SafeLLamaContextHandle

dest Byte*

"},{"location":"xmldocs/llama.native.nativeapi/#returns_27","title":"Returns","text":"

UInt64 the number of bytes copied

"},{"location":"xmldocs/llama.native.nativeapi/#llama_copy_state_datasafellamacontexthandle-byte_1","title":"llama_copy_state_data(SafeLLamaContextHandle, Byte[])","text":"

Copies the state to the specified destination address. Destination needs to have allocated enough memory (see llama_get_state_size)

public static ulong llama_copy_state_data(SafeLLamaContextHandle ctx, Byte[] dest)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_42","title":"Parameters","text":"

ctx SafeLLamaContextHandle

dest Byte[]

"},{"location":"xmldocs/llama.native.nativeapi/#returns_28","title":"Returns","text":"

UInt64 the number of bytes copied

"},{"location":"xmldocs/llama.native.nativeapi/#llama_set_state_datasafellamacontexthandle-byte","title":"llama_set_state_data(SafeLLamaContextHandle, Byte*)","text":"

Set the state reading from the specified address

public static ulong llama_set_state_data(SafeLLamaContextHandle ctx, Byte* src)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_43","title":"Parameters","text":"

ctx SafeLLamaContextHandle

src Byte*

"},{"location":"xmldocs/llama.native.nativeapi/#returns_29","title":"Returns","text":"

UInt64 the number of bytes read

"},{"location":"xmldocs/llama.native.nativeapi/#llama_set_state_datasafellamacontexthandle-byte_1","title":"llama_set_state_data(SafeLLamaContextHandle, Byte[])","text":"

Set the state reading from the specified address

public static ulong llama_set_state_data(SafeLLamaContextHandle ctx, Byte[] src)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_44","title":"Parameters","text":"

ctx SafeLLamaContextHandle

src Byte[]

"},{"location":"xmldocs/llama.native.nativeapi/#returns_30","title":"Returns","text":"

UInt64 the number of bytes read

"},{"location":"xmldocs/llama.native.nativeapi/#llama_load_session_filesafellamacontexthandle-string-int32-uint64-uint64","title":"llama_load_session_file(SafeLLamaContextHandle, String, Int32[], UInt64, UInt64*)","text":"

Load session file

public static bool llama_load_session_file(SafeLLamaContextHandle ctx, string path_session, Int32[] tokens_out, ulong n_token_capacity, UInt64* n_token_count_out)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_45","title":"Parameters","text":"

ctx SafeLLamaContextHandle

path_session String

tokens_out Int32[]

n_token_capacity UInt64

n_token_count_out UInt64*

"},{"location":"xmldocs/llama.native.nativeapi/#returns_31","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_save_session_filesafellamacontexthandle-string-int32-uint64","title":"llama_save_session_file(SafeLLamaContextHandle, String, Int32[], UInt64)","text":"

Save session file

public static bool llama_save_session_file(SafeLLamaContextHandle ctx, string path_session, Int32[] tokens, ulong n_token_count)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_46","title":"Parameters","text":"

ctx SafeLLamaContextHandle

path_session String

tokens Int32[]

n_token_count UInt64

"},{"location":"xmldocs/llama.native.nativeapi/#returns_32","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#llama_evalsafellamacontexthandle-int32-int32-int32-int32","title":"llama_eval(SafeLLamaContextHandle, Int32[], Int32, Int32, Int32)","text":"

Run the llama inference to obtain the logits and probabilities for the next token. tokens + n_tokens is the provided batch of new tokens to process n_past is the number of tokens to use from previous eval calls

public static int llama_eval(SafeLLamaContextHandle ctx, Int32[] tokens, int n_tokens, int n_past, int n_threads)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_47","title":"Parameters","text":"

ctx SafeLLamaContextHandle

tokens Int32[]

n_tokens Int32

n_past Int32

n_threads Int32

"},{"location":"xmldocs/llama.native.nativeapi/#returns_33","title":"Returns","text":"

Int32 Returns 0 on success

"},{"location":"xmldocs/llama.native.nativeapi/#llama_eval_with_pointersafellamacontexthandle-int32-int32-int32-int32","title":"llama_eval_with_pointer(SafeLLamaContextHandle, Int32*, Int32, Int32, Int32)","text":"

Run the llama inference to obtain the logits and probabilities for the next token. tokens + n_tokens is the provided batch of new tokens to process n_past is the number of tokens to use from previous eval calls

public static int llama_eval_with_pointer(SafeLLamaContextHandle ctx, Int32* tokens, int n_tokens, int n_past, int n_threads)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_48","title":"Parameters","text":"

ctx SafeLLamaContextHandle

tokens Int32*

n_tokens Int32

n_past Int32

n_threads Int32

"},{"location":"xmldocs/llama.native.nativeapi/#returns_34","title":"Returns","text":"

Int32 Returns 0 on success

"},{"location":"xmldocs/llama.native.nativeapi/#llama_tokenizesafellamacontexthandle-string-encoding-int32-int32-boolean","title":"llama_tokenize(SafeLLamaContextHandle, String, Encoding, Int32[], Int32, Boolean)","text":"

Convert the provided text into tokens.

public static int llama_tokenize(SafeLLamaContextHandle ctx, string text, Encoding encoding, Int32[] tokens, int n_max_tokens, bool add_bos)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_49","title":"Parameters","text":"

ctx SafeLLamaContextHandle

text String

encoding Encoding

tokens Int32[]

n_max_tokens Int32

add_bos Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#returns_35","title":"Returns","text":"

Int32 Returns the number of tokens on success, no more than n_max_tokens. Returns a negative number on failure - the number of tokens that would have been returned

"},{"location":"xmldocs/llama.native.nativeapi/#llama_tokenize_nativesafellamacontexthandle-byte-int32-int32-boolean","title":"llama_tokenize_native(SafeLLamaContextHandle, Byte, Int32, Int32, Boolean)","text":"

Convert the provided text into tokens.

public static int llama_tokenize_native(SafeLLamaContextHandle ctx, Byte* text, Int32* tokens, int n_max_tokens, bool add_bos)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_50","title":"Parameters","text":"

ctx SafeLLamaContextHandle

text Byte*

tokens Int32*

n_max_tokens Int32

add_bos Boolean

"},{"location":"xmldocs/llama.native.nativeapi/#returns_36","title":"Returns","text":"

Int32 Returns the number of tokens on success, no more than n_max_tokens. Returns a negative number on failure - the number of tokens that would have been returned

"},{"location":"xmldocs/llama.native.nativeapi/#llama_n_vocabsafellamacontexthandle","title":"llama_n_vocab(SafeLLamaContextHandle)","text":"

Get the number of tokens in the model vocabulary for this context

public static int llama_n_vocab(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_51","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_37","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_n_ctxsafellamacontexthandle","title":"llama_n_ctx(SafeLLamaContextHandle)","text":"

Get the size of the context window for the model for this context

public static int llama_n_ctx(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_52","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_38","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_n_embdsafellamacontexthandle","title":"llama_n_embd(SafeLLamaContextHandle)","text":"

Get the dimension of embedding vectors from the model for this context

public static int llama_n_embd(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_53","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_39","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.nativeapi/#llama_get_logitssafellamacontexthandle","title":"llama_get_logits(SafeLLamaContextHandle)","text":"

Token logits obtained from the last call to llama_eval() The logits for the last token are stored in the last row Can be mutated in order to change the probabilities of the next token. Rows: n_tokens Cols: n_vocab

public static Single* llama_get_logits(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_54","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_40","title":"Returns","text":"

Single*

"},{"location":"xmldocs/llama.native.nativeapi/#llama_get_embeddingssafellamacontexthandle","title":"llama_get_embeddings(SafeLLamaContextHandle)","text":"

Get the embeddings for the input shape: [n_embd] (1-dimensional)

public static Single* llama_get_embeddings(SafeLLamaContextHandle ctx)\n
"},{"location":"xmldocs/llama.native.nativeapi/#parameters_55","title":"Parameters","text":"

ctx SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.nativeapi/#returns_41","title":"Returns","text":"

Single*

"},{"location":"xmldocs/llama.native.safellamacontexthandle/","title":"SafeLLamaContextHandle","text":"

Namespace: LLama.Native

A safe wrapper around a llama_context

public sealed class SafeLLamaContextHandle : SafeLLamaHandleBase, System.IDisposable\n

Inheritance Object \u2192 CriticalFinalizerObject \u2192 SafeHandle \u2192 SafeLLamaHandleBase \u2192 SafeLLamaContextHandle Implements IDisposable

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.safellamacontexthandle/#vocabcount","title":"VocabCount","text":"

Total number of tokens in vocabulary of this model

public int VocabCount { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#contextsize","title":"ContextSize","text":"

Total number of tokens in the context

public int ContextSize { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#embeddingsize","title":"EmbeddingSize","text":"

Dimension of embedding vectors

public int EmbeddingSize { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#modelhandle","title":"ModelHandle","text":"

Get the model which this context is using

public SafeLlamaModelHandle ModelHandle { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value_3","title":"Property Value","text":"

SafeLlamaModelHandle

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#isinvalid","title":"IsInvalid","text":"
public bool IsInvalid { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value_4","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#isclosed","title":"IsClosed","text":"
public bool IsClosed { get; }\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#property-value_5","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.safellamacontexthandle/#safellamacontexthandleintptr-safellamamodelhandle","title":"SafeLLamaContextHandle(IntPtr, SafeLlamaModelHandle)","text":"

Create a new SafeLLamaContextHandle

public SafeLLamaContextHandle(IntPtr handle, SafeLlamaModelHandle model)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters","title":"Parameters","text":"

handle IntPtr pointer to an allocated llama_context

model SafeLlamaModelHandle the model which this context was created from

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.safellamacontexthandle/#releasehandle","title":"ReleaseHandle()","text":"
protected bool ReleaseHandle()\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#createsafellamamodelhandle-llamacontextparams","title":"Create(SafeLlamaModelHandle, LLamaContextParams)","text":"

Create a new llama_state for the given model

public static SafeLLamaContextHandle Create(SafeLlamaModelHandle model, LLamaContextParams lparams)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_1","title":"Parameters","text":"

model SafeLlamaModelHandle

lparams LLamaContextParams

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_1","title":"Returns","text":"

SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#exceptions","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#clonellamacontextparams","title":"Clone(LLamaContextParams)","text":"

Create a new llama context with a clone of the current llama context state

public SafeLLamaContextHandle Clone(LLamaContextParams lparams)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_2","title":"Parameters","text":"

lparams LLamaContextParams

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_2","title":"Returns","text":"

SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#tokenizestring-boolean-encoding","title":"Tokenize(String, Boolean, Encoding)","text":"

Convert the given text into tokens

public Int32[] Tokenize(string text, bool add_bos, Encoding encoding)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_3","title":"Parameters","text":"

text String The text to tokenize

add_bos Boolean Whether the \"BOS\" token should be added

encoding Encoding Encoding to use for the text

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_3","title":"Returns","text":"

Int32[]

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#exceptions_1","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#getlogits","title":"GetLogits()","text":"

Token logits obtained from the last call to llama_eval() The logits for the last token are stored in the last row Can be mutated in order to change the probabilities of the next token. Rows: n_tokens Cols: n_vocab

public Span<float> GetLogits()\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_4","title":"Returns","text":"

Span<Single>

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#tokentostringint32-encoding","title":"TokenToString(Int32, Encoding)","text":"

Convert a token into a string

public string TokenToString(int token, Encoding encoding)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_4","title":"Parameters","text":"

token Int32 Token to decode into a string

encoding Encoding

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_5","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#tokentostringint32-encoding-stringbuilder","title":"TokenToString(Int32, Encoding, StringBuilder)","text":"

Append a single llama token to a string builder

public void TokenToString(int token, Encoding encoding, StringBuilder dest)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_5","title":"Parameters","text":"

token Int32 Token to decode

encoding Encoding

dest StringBuilder string builder to append the result to

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#tokentospanint32-spanbyte","title":"TokenToSpan(Int32, Span<Byte>)","text":"

Convert a single llama token into bytes

public int TokenToSpan(int token, Span<byte> dest)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_6","title":"Parameters","text":"

token Int32 Token to decode

dest Span<Byte> A span to attempt to write into. If this is too small nothing will be written

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_6","title":"Returns","text":"

Int32 The size of this token. nothing will be written if this is larger than dest

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#evalreadonlyspanint32-int32-int32","title":"Eval(ReadOnlySpan<Int32>, Int32, Int32)","text":"

Run the llama inference to obtain the logits and probabilities for the next token.

public bool Eval(ReadOnlySpan<int> tokens, int n_past, int n_threads)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_7","title":"Parameters","text":"

tokens ReadOnlySpan<Int32> The provided batch of new tokens to process

n_past Int32 the number of tokens to use from previous eval calls

n_threads Int32

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_7","title":"Returns","text":"

Boolean Returns true on success

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#getstatesize","title":"GetStateSize()","text":"

Get the size of the state, when saved as bytes

public ulong GetStateSize()\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_8","title":"Returns","text":"

UInt64

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#getstatebyte-uint64","title":"GetState(Byte*, UInt64)","text":"

Get the raw state of this context, encoded as bytes. Data is written into the dest pointer.

public ulong GetState(Byte* dest, ulong size)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_8","title":"Parameters","text":"

dest Byte* Destination to write to

size UInt64 Number of bytes available to write to in dest (check required size with GetStateSize())

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_9","title":"Returns","text":"

UInt64 The number of bytes written to dest

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#exceptions_2","title":"Exceptions","text":"

ArgumentOutOfRangeException Thrown if dest is too small

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#getstateintptr-uint64","title":"GetState(IntPtr, UInt64)","text":"

Get the raw state of this context, encoded as bytes. Data is written into the dest pointer.

public ulong GetState(IntPtr dest, ulong size)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_9","title":"Parameters","text":"

dest IntPtr Destination to write to

size UInt64 Number of bytes available to write to in dest (check required size with GetStateSize())

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_10","title":"Returns","text":"

UInt64 The number of bytes written to dest

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#exceptions_3","title":"Exceptions","text":"

ArgumentOutOfRangeException Thrown if dest is too small

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#setstatebyte","title":"SetState(Byte*)","text":"

Set the raw state of this context

public ulong SetState(Byte* src)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_10","title":"Parameters","text":"

src Byte* The pointer to read the state from

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_11","title":"Returns","text":"

UInt64 Number of bytes read from the src pointer

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#setstateintptr","title":"SetState(IntPtr)","text":"

Set the raw state of this context

public ulong SetState(IntPtr src)\n
"},{"location":"xmldocs/llama.native.safellamacontexthandle/#parameters_11","title":"Parameters","text":"

src IntPtr The pointer to read the state from

"},{"location":"xmldocs/llama.native.safellamacontexthandle/#returns_12","title":"Returns","text":"

UInt64 Number of bytes read from the src pointer

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/","title":"SafeLLamaGrammarHandle","text":"

Namespace: LLama.Native

A safe reference to a llama_grammar

public class SafeLLamaGrammarHandle : SafeLLamaHandleBase, System.IDisposable\n

Inheritance Object \u2192 CriticalFinalizerObject \u2192 SafeHandle \u2192 SafeLLamaHandleBase \u2192 SafeLLamaGrammarHandle Implements IDisposable

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.safellamagrammarhandle/#isinvalid","title":"IsInvalid","text":"
public bool IsInvalid { get; }\n
"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#property-value","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#isclosed","title":"IsClosed","text":"
public bool IsClosed { get; }\n
"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#property-value_1","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.safellamagrammarhandle/#releasehandle","title":"ReleaseHandle()","text":"
protected bool ReleaseHandle()\n
"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#returns","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#createireadonlylistgrammarrule-uint64","title":"Create(IReadOnlyList<GrammarRule>, UInt64)","text":"

Create a new llama_grammar

public static SafeLLamaGrammarHandle Create(IReadOnlyList<GrammarRule> rules, ulong start_rule_index)\n
"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#parameters","title":"Parameters","text":"

rules IReadOnlyList<GrammarRule> A list of list of elements, each inner list makes up one grammar rule

start_rule_index UInt64 The index (in the outer list) of the start rule

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#returns_1","title":"Returns","text":"

SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#exceptions","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#createllamagrammarelement-uint64-uint64","title":"Create(LLamaGrammarElement, UInt64, UInt64)**","text":"

Create a new llama_grammar

public static SafeLLamaGrammarHandle Create(LLamaGrammarElement** rules, ulong nrules, ulong start_rule_index)\n
"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#parameters_1","title":"Parameters","text":"

rules LLamaGrammarElement** rules list, each rule is a list of rule elements (terminated by a LLamaGrammarElementType.END element)

nrules UInt64 total number of rules

start_rule_index UInt64 index of the start rule of the grammar

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#returns_2","title":"Returns","text":"

SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.native.safellamagrammarhandle/#exceptions_1","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamahandlebase/","title":"SafeLLamaHandleBase","text":"

Namespace: LLama.Native

Base class for all llama handles to native resources

public abstract class SafeLLamaHandleBase : System.Runtime.InteropServices.SafeHandle, System.IDisposable\n

Inheritance Object \u2192 CriticalFinalizerObject \u2192 SafeHandle \u2192 SafeLLamaHandleBase Implements IDisposable

"},{"location":"xmldocs/llama.native.safellamahandlebase/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.safellamahandlebase/#isinvalid","title":"IsInvalid","text":"
public bool IsInvalid { get; }\n
"},{"location":"xmldocs/llama.native.safellamahandlebase/#property-value","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamahandlebase/#isclosed","title":"IsClosed","text":"
public bool IsClosed { get; }\n
"},{"location":"xmldocs/llama.native.safellamahandlebase/#property-value_1","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamahandlebase/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.safellamahandlebase/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.native.safellamahandlebase/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.native.safellamamodelhandle/","title":"SafeLlamaModelHandle","text":"

Namespace: LLama.Native

A reference to a set of llama model weights

public sealed class SafeLlamaModelHandle : SafeLLamaHandleBase, System.IDisposable\n

Inheritance Object \u2192 CriticalFinalizerObject \u2192 SafeHandle \u2192 SafeLLamaHandleBase \u2192 SafeLlamaModelHandle Implements IDisposable

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.native.safellamamodelhandle/#vocabcount","title":"VocabCount","text":"

Total number of tokens in vocabulary of this model

public int VocabCount { get; }\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#contextsize","title":"ContextSize","text":"

Total number of tokens in the context

public int ContextSize { get; }\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#embeddingsize","title":"EmbeddingSize","text":"

Dimension of embedding vectors

public int EmbeddingSize { get; }\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#isinvalid","title":"IsInvalid","text":"
public bool IsInvalid { get; }\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#property-value_3","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#isclosed","title":"IsClosed","text":"
public bool IsClosed { get; }\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#property-value_4","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.safellamamodelhandle/#releasehandle","title":"ReleaseHandle()","text":"
protected bool ReleaseHandle()\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#loadfromfilestring-llamacontextparams","title":"LoadFromFile(String, LLamaContextParams)","text":"

Load a model from the given file path into memory

public static SafeLlamaModelHandle LoadFromFile(string modelPath, LLamaContextParams lparams)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters","title":"Parameters","text":"

modelPath String

lparams LLamaContextParams

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns_1","title":"Returns","text":"

SafeLlamaModelHandle

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#exceptions","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#applylorafromfilestring-string-int32","title":"ApplyLoraFromFile(String, String, Int32)","text":"

Apply a LoRA adapter to a loaded model

public void ApplyLoraFromFile(string lora, string modelBase, int threads)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_1","title":"Parameters","text":"

lora String

modelBase String A path to a higher quality model to use as a base for the layers modified by the adapter. Can be NULL to use the current loaded model.

threads Int32

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#exceptions_1","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#tokentospanint32-spanbyte","title":"TokenToSpan(Int32, Span<Byte>)","text":"

Convert a single llama token into bytes

public int TokenToSpan(int llama_token, Span<byte> dest)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_2","title":"Parameters","text":"

llama_token Int32 Token to decode

dest Span<Byte> A span to attempt to write into. If this is too small nothing will be written

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns_2","title":"Returns","text":"

Int32 The size of this token. nothing will be written if this is larger than dest

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#tokentostringint32-encoding","title":"TokenToString(Int32, Encoding)","text":"

Convert a single llama token into a string

public string TokenToString(int llama_token, Encoding encoding)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_3","title":"Parameters","text":"

llama_token Int32

encoding Encoding Encoding to use to decode the bytes into a string

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns_3","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#tokentostringint32-encoding-stringbuilder","title":"TokenToString(Int32, Encoding, StringBuilder)","text":"

Append a single llama token to a string builder

public void TokenToString(int llama_token, Encoding encoding, StringBuilder dest)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_4","title":"Parameters","text":"

llama_token Int32 Token to decode

encoding Encoding

dest StringBuilder string builder to append the result to

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#tokenizestring-boolean-encoding","title":"Tokenize(String, Boolean, Encoding)","text":"

Convert a string of text into tokens

public Int32[] Tokenize(string text, bool add_bos, Encoding encoding)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_5","title":"Parameters","text":"

text String

add_bos Boolean

encoding Encoding

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns_4","title":"Returns","text":"

Int32[]

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#createcontextllamacontextparams","title":"CreateContext(LLamaContextParams)","text":"

Create a new context for this model

public SafeLLamaContextHandle CreateContext(LLamaContextParams params)\n
"},{"location":"xmldocs/llama.native.safellamamodelhandle/#parameters_6","title":"Parameters","text":"

params LLamaContextParams

"},{"location":"xmldocs/llama.native.safellamamodelhandle/#returns_5","title":"Returns","text":"

SafeLLamaContextHandle

"},{"location":"xmldocs/llama.native.samplingapi/","title":"SamplingApi","text":"

Namespace: LLama.Native

Direct translation of the llama.cpp sampling API

public class SamplingApi\n

Inheritance Object \u2192 SamplingApi

"},{"location":"xmldocs/llama.native.samplingapi/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.native.samplingapi/#samplingapi_1","title":"SamplingApi()","text":"
public SamplingApi()\n
"},{"location":"xmldocs/llama.native.samplingapi/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_grammarsafellamacontexthandle-llamatokendataarray-safellamagrammarhandle","title":"llama_sample_grammar(SafeLLamaContextHandle, LLamaTokenDataArray, SafeLLamaGrammarHandle)","text":"

Apply grammar rules to candidate tokens

public static void llama_sample_grammar(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, SafeLLamaGrammarHandle grammar)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray

grammar SafeLLamaGrammarHandle

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_repetition_penaltysafellamacontexthandle-llamatokendataarray-memoryint32-uint64-single","title":"llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, UInt64, Single)","text":""},{"location":"xmldocs/llama.native.samplingapi/#caution","title":"Caution","text":"

last_tokens_size parameter is no longer needed

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, ulong last_tokens_size, float penalty)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_1","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

last_tokens Memory<Int32>

last_tokens_size UInt64

penalty Single

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_repetition_penaltysafellamacontexthandle-llamatokendataarray-memoryint32-single","title":"llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, Single)","text":"

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, float penalty)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_2","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

last_tokens Memory<Int32>

penalty Single

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_frequency_and_presence_penaltiessafellamacontexthandle-llamatokendataarray-memoryint32-uint64-single-single","title":"llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, UInt64, Single, Single)","text":""},{"location":"xmldocs/llama.native.samplingapi/#caution_1","title":"Caution","text":"

last_tokens_size parameter is no longer needed

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, ulong last_tokens_size, float alpha_frequency, float alpha_presence)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_3","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

last_tokens Memory<Int32>

last_tokens_size UInt64

alpha_frequency Single

alpha_presence Single

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_frequency_and_presence_penaltiessafellamacontexthandle-llamatokendataarray-memoryint32-single-single","title":"llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, Single, Single)","text":"

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, float alpha_frequency, float alpha_presence)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_4","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

last_tokens Memory<Int32>

alpha_frequency Single

alpha_presence Single

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_softmaxsafellamacontexthandle-llamatokendataarray","title":"llama_sample_softmax(SafeLLamaContextHandle, LLamaTokenDataArray)","text":"

Sorts candidate tokens by their logits in descending order and calculate probabilities based on logits.

public static void llama_sample_softmax(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_5","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_top_ksafellamacontexthandle-llamatokendataarray-int32-uint64","title":"llama_sample_top_k(SafeLLamaContextHandle, LLamaTokenDataArray, Int32, UInt64)","text":"

Top-K sampling described in academic paper \"The Curious Case of Neural Text Degeneration\" https://arxiv.org/abs/1904.09751

public static void llama_sample_top_k(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, int k, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_6","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

k Int32

min_keep UInt64

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_top_psafellamacontexthandle-llamatokendataarray-single-uint64","title":"llama_sample_top_p(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)","text":"

Nucleus sampling described in academic paper \"The Curious Case of Neural Text Degeneration\" https://arxiv.org/abs/1904.09751

public static void llama_sample_top_p(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float p, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_7","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

p Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_tail_freesafellamacontexthandle-llamatokendataarray-single-uint64","title":"llama_sample_tail_free(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)","text":"

Tail Free Sampling described in https://www.trentonbricken.com/Tail-Free-Sampling/.

public static void llama_sample_tail_free(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float z, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_8","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

z Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_typicalsafellamacontexthandle-llamatokendataarray-single-uint64","title":"llama_sample_typical(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)","text":"

Locally Typical Sampling implementation described in the paper https://arxiv.org/abs/2202.00666.

public static void llama_sample_typical(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float p, ulong min_keep)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_9","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

p Single

min_keep UInt64

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_temperaturesafellamacontexthandle-llamatokendataarray-single","title":"llama_sample_temperature(SafeLLamaContextHandle, LLamaTokenDataArray, Single)","text":"

Sample with temperature. As temperature increases, the prediction becomes diverse but also vulnerable to hallucinations -- generating tokens that are sensible but not factual

public static void llama_sample_temperature(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float temp)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_10","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray

temp Single

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_token_mirostatsafellamacontexthandle-llamatokendataarray-single-single-int32-single","title":"llama_sample_token_mirostat(SafeLLamaContextHandle, LLamaTokenDataArray, Single, Single, Int32, Single&)","text":"

Mirostat 1.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

public static int llama_sample_token_mirostat(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float tau, float eta, int m, Single& mu)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_11","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray A vector of LLamaTokenData containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

tau Single The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

eta Single The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

m Int32 The number of tokens considered in the estimation of s_hat. This is an arbitrary value that is used to calculate s_hat, which in turn helps to calculate the value of k. In the paper, they use m = 100, but you can experiment with different values to see how it affects the performance of the algorithm.

mu Single& Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

"},{"location":"xmldocs/llama.native.samplingapi/#returns","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_token_mirostat_v2safellamacontexthandle-llamatokendataarray-single-single-single","title":"llama_sample_token_mirostat_v2(SafeLLamaContextHandle, LLamaTokenDataArray, Single, Single, Single&)","text":"

Mirostat 2.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

public static int llama_sample_token_mirostat_v2(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float tau, float eta, Single& mu)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_12","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray A vector of LLamaTokenData containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

tau Single The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

eta Single The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

mu Single& Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

"},{"location":"xmldocs/llama.native.samplingapi/#returns_1","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_token_greedysafellamacontexthandle-llamatokendataarray","title":"llama_sample_token_greedy(SafeLLamaContextHandle, LLamaTokenDataArray)","text":"

Selects the token with the highest probability.

public static int llama_sample_token_greedy(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_13","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.samplingapi/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.native.samplingapi/#llama_sample_tokensafellamacontexthandle-llamatokendataarray","title":"llama_sample_token(SafeLLamaContextHandle, LLamaTokenDataArray)","text":"

Randomly selects a token from the candidates based on their probabilities.

public static int llama_sample_token(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)\n
"},{"location":"xmldocs/llama.native.samplingapi/#parameters_14","title":"Parameters","text":"

ctx SafeLLamaContextHandle

candidates LLamaTokenDataArray Pointer to LLamaTokenDataArray

"},{"location":"xmldocs/llama.native.samplingapi/#returns_3","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletion/","title":"ChatCompletion","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletion : System.IEquatable`1[[LLama.OldVersion.ChatCompletion, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletion Implements IEquatable<ChatCompletion>

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletion/#id","title":"Id","text":"
public string Id { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#created","title":"Created","text":"
public int Created { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#model","title":"Model","text":"
public string Model { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value_3","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#choices","title":"Choices","text":"
public ChatCompletionChoice[] Choices { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value_4","title":"Property Value","text":"

ChatCompletionChoice[]

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#usage","title":"Usage","text":"
public CompletionUsage Usage { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#property-value_5","title":"Property Value","text":"

CompletionUsage

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletion/#chatcompletionstring-string-int32-string-chatcompletionchoice-completionusage","title":"ChatCompletion(String, String, Int32, String, ChatCompletionChoice[], CompletionUsage)","text":"
public ChatCompletion(string Id, string Object, int Created, string Model, ChatCompletionChoice[] Choices, CompletionUsage Usage)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#parameters","title":"Parameters","text":"

Id String

Object String

Created Int32

Model String

Choices ChatCompletionChoice[]

Usage CompletionUsage

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletion/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#equalschatcompletion","title":"Equals(ChatCompletion)","text":"
public bool Equals(ChatCompletion other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#parameters_3","title":"Parameters","text":"

other ChatCompletion

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#clone","title":"<Clone>$()","text":"
public ChatCompletion <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#returns_5","title":"Returns","text":"

ChatCompletion

"},{"location":"xmldocs/llama.oldversion.chatcompletion/#deconstructstring-string-int32-string-chatcompletionchoice-completionusage","title":"Deconstruct(String&, String&, Int32&, String&, ChatCompletionChoice[]&, CompletionUsage&)","text":"
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, ChatCompletionChoice[]& Choices, CompletionUsage& Usage)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletion/#parameters_4","title":"Parameters","text":"

Id String&

Object String&

Created Int32&

Model String&

Choices ChatCompletionChoice[]&

Usage CompletionUsage&

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/","title":"ChatCompletionChoice","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletionChoice : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletionChoice Implements IEquatable<ChatCompletionChoice>

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#index","title":"Index","text":"
public int Index { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#message","title":"Message","text":"
public ChatCompletionMessage Message { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#property-value_1","title":"Property Value","text":"

ChatCompletionMessage

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#finishreason","title":"FinishReason","text":"
public string FinishReason { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#property-value_2","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#chatcompletionchoiceint32-chatcompletionmessage-string","title":"ChatCompletionChoice(Int32, ChatCompletionMessage, String)","text":"
public ChatCompletionChoice(int Index, ChatCompletionMessage Message, string FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#parameters","title":"Parameters","text":"

Index Int32

Message ChatCompletionMessage

FinishReason String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#equalschatcompletionchoice","title":"Equals(ChatCompletionChoice)","text":"
public bool Equals(ChatCompletionChoice other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#parameters_3","title":"Parameters","text":"

other ChatCompletionChoice

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#clone","title":"<Clone>$()","text":"
public ChatCompletionChoice <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#returns_5","title":"Returns","text":"

ChatCompletionChoice

"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#deconstructint32-chatcompletionmessage-string","title":"Deconstruct(Int32&, ChatCompletionMessage&, String&)","text":"
public void Deconstruct(Int32& Index, ChatCompletionMessage& Message, String& FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchoice/#parameters_4","title":"Parameters","text":"

Index Int32&

Message ChatCompletionMessage&

FinishReason String&

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/","title":"ChatCompletionChunk","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletionChunk : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunk, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletionChunk Implements IEquatable<ChatCompletionChunk>

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#id","title":"Id","text":"
public string Id { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#model","title":"Model","text":"
public string Model { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#property-value_2","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#created","title":"Created","text":"
public int Created { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#property-value_3","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#choices","title":"Choices","text":"
public ChatCompletionChunkChoice[] Choices { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#property-value_4","title":"Property Value","text":"

ChatCompletionChunkChoice[]

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#chatcompletionchunkstring-string-string-int32-chatcompletionchunkchoice","title":"ChatCompletionChunk(String, String, String, Int32, ChatCompletionChunkChoice[])","text":"
public ChatCompletionChunk(string Id, string Model, string Object, int Created, ChatCompletionChunkChoice[] Choices)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#parameters","title":"Parameters","text":"

Id String

Model String

Object String

Created Int32

Choices ChatCompletionChunkChoice[]

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#equalschatcompletionchunk","title":"Equals(ChatCompletionChunk)","text":"
public bool Equals(ChatCompletionChunk other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#parameters_3","title":"Parameters","text":"

other ChatCompletionChunk

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#clone","title":"<Clone>$()","text":"
public ChatCompletionChunk <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#returns_5","title":"Returns","text":"

ChatCompletionChunk

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#deconstructstring-string-string-int32-chatcompletionchunkchoice","title":"Deconstruct(String&, String&, String&, Int32&, ChatCompletionChunkChoice[]&)","text":"
public void Deconstruct(String& Id, String& Model, String& Object, Int32& Created, ChatCompletionChunkChoice[]& Choices)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunk/#parameters_4","title":"Parameters","text":"

Id String&

Model String&

Object String&

Created Int32&

Choices ChatCompletionChunkChoice[]&

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/","title":"ChatCompletionChunkChoice","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletionChunkChoice : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunkChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletionChunkChoice Implements IEquatable<ChatCompletionChunkChoice>

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#index","title":"Index","text":"
public int Index { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#delta","title":"Delta","text":"
public ChatCompletionChunkDelta Delta { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#property-value_1","title":"Property Value","text":"

ChatCompletionChunkDelta

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#finishreason","title":"FinishReason","text":"
public string FinishReason { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#property-value_2","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#chatcompletionchunkchoiceint32-chatcompletionchunkdelta-string","title":"ChatCompletionChunkChoice(Int32, ChatCompletionChunkDelta, String)","text":"
public ChatCompletionChunkChoice(int Index, ChatCompletionChunkDelta Delta, string FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#parameters","title":"Parameters","text":"

Index Int32

Delta ChatCompletionChunkDelta

FinishReason String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#equalschatcompletionchunkchoice","title":"Equals(ChatCompletionChunkChoice)","text":"
public bool Equals(ChatCompletionChunkChoice other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#parameters_3","title":"Parameters","text":"

other ChatCompletionChunkChoice

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#clone","title":"<Clone>$()","text":"
public ChatCompletionChunkChoice <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#returns_5","title":"Returns","text":"

ChatCompletionChunkChoice

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#deconstructint32-chatcompletionchunkdelta-string","title":"Deconstruct(Int32&, ChatCompletionChunkDelta&, String&)","text":"
public void Deconstruct(Int32& Index, ChatCompletionChunkDelta& Delta, String& FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkchoice/#parameters_4","title":"Parameters","text":"

Index Int32&

Delta ChatCompletionChunkDelta&

FinishReason String&

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/","title":"ChatCompletionChunkDelta","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletionChunkDelta : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunkDelta, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletionChunkDelta Implements IEquatable<ChatCompletionChunkDelta>

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#role","title":"Role","text":"
public string Role { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#content","title":"Content","text":"
public string Content { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#chatcompletionchunkdeltastring-string","title":"ChatCompletionChunkDelta(String, String)","text":"
public ChatCompletionChunkDelta(string Role, string Content)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#parameters","title":"Parameters","text":"

Role String

Content String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#equalschatcompletionchunkdelta","title":"Equals(ChatCompletionChunkDelta)","text":"
public bool Equals(ChatCompletionChunkDelta other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#parameters_3","title":"Parameters","text":"

other ChatCompletionChunkDelta

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#clone","title":"<Clone>$()","text":"
public ChatCompletionChunkDelta <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#returns_5","title":"Returns","text":"

ChatCompletionChunkDelta

"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#deconstructstring-string","title":"Deconstruct(String&, String&)","text":"
public void Deconstruct(String& Role, String& Content)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionchunkdelta/#parameters_4","title":"Parameters","text":"

Role String&

Content String&

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/","title":"ChatCompletionMessage","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatCompletionMessage : System.IEquatable`1[[LLama.OldVersion.ChatCompletionMessage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatCompletionMessage Implements IEquatable<ChatCompletionMessage>

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#role","title":"Role","text":"
public ChatRole Role { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#property-value","title":"Property Value","text":"

ChatRole

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#content","title":"Content","text":"
public string Content { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#name","title":"Name","text":"
public string Name { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#property-value_2","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#chatcompletionmessagechatrole-string-string","title":"ChatCompletionMessage(ChatRole, String, String)","text":"
public ChatCompletionMessage(ChatRole Role, string Content, string Name)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#parameters","title":"Parameters","text":"

Role ChatRole

Content String

Name String

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#equalschatcompletionmessage","title":"Equals(ChatCompletionMessage)","text":"
public bool Equals(ChatCompletionMessage other)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#parameters_3","title":"Parameters","text":"

other ChatCompletionMessage

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#clone","title":"<Clone>$()","text":"
public ChatCompletionMessage <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#returns_5","title":"Returns","text":"

ChatCompletionMessage

"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#deconstructchatrole-string-string","title":"Deconstruct(ChatRole&, String&, String&)","text":"
public void Deconstruct(ChatRole& Role, String& Content, String& Name)\n
"},{"location":"xmldocs/llama.oldversion.chatcompletionmessage/#parameters_4","title":"Parameters","text":"

Role ChatRole&

Content String&

Name String&

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/","title":"ChatMessageRecord","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatMessageRecord : System.IEquatable`1[[LLama.OldVersion.ChatMessageRecord, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 ChatMessageRecord Implements IEquatable<ChatMessageRecord>

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#message","title":"Message","text":"
public ChatCompletionMessage Message { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#property-value","title":"Property Value","text":"

ChatCompletionMessage

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#time","title":"Time","text":"
public DateTime Time { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#property-value_1","title":"Property Value","text":"

DateTime

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#chatmessagerecordchatcompletionmessage-datetime","title":"ChatMessageRecord(ChatCompletionMessage, DateTime)","text":"
public ChatMessageRecord(ChatCompletionMessage Message, DateTime Time)\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#parameters","title":"Parameters","text":"

Message ChatCompletionMessage

Time DateTime

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#equalschatmessagerecord","title":"Equals(ChatMessageRecord)","text":"
public bool Equals(ChatMessageRecord other)\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#parameters_3","title":"Parameters","text":"

other ChatMessageRecord

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#clone","title":"<Clone>$()","text":"
public ChatMessageRecord <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#returns_5","title":"Returns","text":"

ChatMessageRecord

"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#deconstructchatcompletionmessage-datetime","title":"Deconstruct(ChatCompletionMessage&, DateTime&)","text":"
public void Deconstruct(ChatCompletionMessage& Message, DateTime& Time)\n
"},{"location":"xmldocs/llama.oldversion.chatmessagerecord/#parameters_4","title":"Parameters","text":"

Message ChatCompletionMessage&

Time DateTime&

"},{"location":"xmldocs/llama.oldversion.chatrole/","title":"ChatRole","text":"

Namespace: LLama.OldVersion

public enum ChatRole\n

Inheritance Object \u2192 ValueType \u2192 Enum \u2192 ChatRole Implements IComparable, IFormattable, IConvertible

"},{"location":"xmldocs/llama.oldversion.chatrole/#fields","title":"Fields","text":"Name Value Description"},{"location":"xmldocs/llama.oldversion.chatsession-1/","title":"ChatSession<T>","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class ChatSession<T>\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#type-parameters","title":"Type Parameters","text":"

T

Inheritance Object \u2192 ChatSession<T>

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.chatsession-1/#chatsessiont_1","title":"ChatSession(T)","text":"
public ChatSession(T model)\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#parameters","title":"Parameters","text":"

model T

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.chatsession-1/#chatstring-string-string","title":"Chat(String, String, String)","text":"
public IEnumerable<string> Chat(string text, string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#parameters_1","title":"Parameters","text":"

text String

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#returns","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#withpromptstring-string","title":"WithPrompt(String, String)","text":"
public ChatSession<T> WithPrompt(string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#parameters_2","title":"Parameters","text":"

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#returns_1","title":"Returns","text":"

ChatSession<T>

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#withpromptfilestring-string","title":"WithPromptFile(String, String)","text":"
public ChatSession<T> WithPromptFile(string promptFilename, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#parameters_3","title":"Parameters","text":"

promptFilename String

encoding String

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#returns_2","title":"Returns","text":"

ChatSession<T>

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#withantipromptstring","title":"WithAntiprompt(String[])","text":"

Set the keywords to split the return value of chat AI.

public ChatSession<T> WithAntiprompt(String[] antiprompt)\n
"},{"location":"xmldocs/llama.oldversion.chatsession-1/#parameters_4","title":"Parameters","text":"

antiprompt String[]

"},{"location":"xmldocs/llama.oldversion.chatsession-1/#returns_3","title":"Returns","text":"

ChatSession<T>

"},{"location":"xmldocs/llama.oldversion.completion/","title":"Completion","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.completion/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class Completion : System.IEquatable`1[[LLama.OldVersion.Completion, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 Completion Implements IEquatable<Completion>

"},{"location":"xmldocs/llama.oldversion.completion/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.completion/#id","title":"Id","text":"
public string Id { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completion/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completion/#created","title":"Created","text":"
public int Created { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completion/#model","title":"Model","text":"
public string Model { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value_3","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completion/#choices","title":"Choices","text":"
public CompletionChoice[] Choices { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value_4","title":"Property Value","text":"

CompletionChoice[]

"},{"location":"xmldocs/llama.oldversion.completion/#usage","title":"Usage","text":"
public CompletionUsage Usage { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completion/#property-value_5","title":"Property Value","text":"

CompletionUsage

"},{"location":"xmldocs/llama.oldversion.completion/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.completion/#completionstring-string-int32-string-completionchoice-completionusage","title":"Completion(String, String, Int32, String, CompletionChoice[], CompletionUsage)","text":"
public Completion(string Id, string Object, int Created, string Model, CompletionChoice[] Choices, CompletionUsage Usage)\n
"},{"location":"xmldocs/llama.oldversion.completion/#parameters","title":"Parameters","text":"

Id String

Object String

Created Int32

Model String

Choices CompletionChoice[]

Usage CompletionUsage

"},{"location":"xmldocs/llama.oldversion.completion/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.completion/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.completion/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.completion/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.completion/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.completion/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completion/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.completion/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completion/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.completion/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.completion/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completion/#equalscompletion","title":"Equals(Completion)","text":"
public bool Equals(Completion other)\n
"},{"location":"xmldocs/llama.oldversion.completion/#parameters_3","title":"Parameters","text":"

other Completion

"},{"location":"xmldocs/llama.oldversion.completion/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completion/#clone","title":"<Clone>$()","text":"
public Completion <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.completion/#returns_5","title":"Returns","text":"

Completion

"},{"location":"xmldocs/llama.oldversion.completion/#deconstructstring-string-int32-string-completionchoice-completionusage","title":"Deconstruct(String&, String&, Int32&, String&, CompletionChoice[]&, CompletionUsage&)","text":"
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, CompletionChoice[]& Choices, CompletionUsage& Usage)\n
"},{"location":"xmldocs/llama.oldversion.completion/#parameters_4","title":"Parameters","text":"

Id String&

Object String&

Created Int32&

Model String&

Choices CompletionChoice[]&

Usage CompletionUsage&

"},{"location":"xmldocs/llama.oldversion.completionchoice/","title":"CompletionChoice","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.completionchoice/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class CompletionChoice : System.IEquatable`1[[LLama.OldVersion.CompletionChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 CompletionChoice Implements IEquatable<CompletionChoice>

"},{"location":"xmldocs/llama.oldversion.completionchoice/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.completionchoice/#text","title":"Text","text":"
public string Text { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchoice/#index","title":"Index","text":"
public int Index { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionchoice/#logprobs","title":"Logprobs","text":"
public CompletionLogprobs Logprobs { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#property-value_2","title":"Property Value","text":"

CompletionLogprobs

"},{"location":"xmldocs/llama.oldversion.completionchoice/#finishreason","title":"FinishReason","text":"
public string FinishReason { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#property-value_3","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchoice/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.completionchoice/#completionchoicestring-int32-completionlogprobs-string","title":"CompletionChoice(String, Int32, CompletionLogprobs, String)","text":"
public CompletionChoice(string Text, int Index, CompletionLogprobs Logprobs, string FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#parameters","title":"Parameters","text":"

Text String

Index Int32

Logprobs CompletionLogprobs

FinishReason String

"},{"location":"xmldocs/llama.oldversion.completionchoice/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.completionchoice/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchoice/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchoice/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionchoice/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchoice/#equalscompletionchoice","title":"Equals(CompletionChoice)","text":"
public bool Equals(CompletionChoice other)\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#parameters_3","title":"Parameters","text":"

other CompletionChoice

"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchoice/#clone","title":"<Clone>$()","text":"
public CompletionChoice <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#returns_5","title":"Returns","text":"

CompletionChoice

"},{"location":"xmldocs/llama.oldversion.completionchoice/#deconstructstring-int32-completionlogprobs-string","title":"Deconstruct(String&, Int32&, CompletionLogprobs&, String&)","text":"
public void Deconstruct(String& Text, Int32& Index, CompletionLogprobs& Logprobs, String& FinishReason)\n
"},{"location":"xmldocs/llama.oldversion.completionchoice/#parameters_4","title":"Parameters","text":"

Text String&

Index Int32&

Logprobs CompletionLogprobs&

FinishReason String&

"},{"location":"xmldocs/llama.oldversion.completionchunk/","title":"CompletionChunk","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.completionchunk/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class CompletionChunk : System.IEquatable`1[[LLama.OldVersion.CompletionChunk, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 CompletionChunk Implements IEquatable<CompletionChunk>

"},{"location":"xmldocs/llama.oldversion.completionchunk/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.completionchunk/#id","title":"Id","text":"
public string Id { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchunk/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchunk/#created","title":"Created","text":"
public int Created { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionchunk/#model","title":"Model","text":"
public string Model { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#property-value_3","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchunk/#choices","title":"Choices","text":"
public CompletionChoice[] Choices { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#property-value_4","title":"Property Value","text":"

CompletionChoice[]

"},{"location":"xmldocs/llama.oldversion.completionchunk/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.completionchunk/#completionchunkstring-string-int32-string-completionchoice","title":"CompletionChunk(String, String, Int32, String, CompletionChoice[])","text":"
public CompletionChunk(string Id, string Object, int Created, string Model, CompletionChoice[] Choices)\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#parameters","title":"Parameters","text":"

Id String

Object String

Created Int32

Model String

Choices CompletionChoice[]

"},{"location":"xmldocs/llama.oldversion.completionchunk/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.completionchunk/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionchunk/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchunk/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionchunk/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchunk/#equalscompletionchunk","title":"Equals(CompletionChunk)","text":"
public bool Equals(CompletionChunk other)\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#parameters_3","title":"Parameters","text":"

other CompletionChunk

"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionchunk/#clone","title":"<Clone>$()","text":"
public CompletionChunk <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#returns_5","title":"Returns","text":"

CompletionChunk

"},{"location":"xmldocs/llama.oldversion.completionchunk/#deconstructstring-string-int32-string-completionchoice","title":"Deconstruct(String&, String&, Int32&, String&, CompletionChoice[]&)","text":"
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, CompletionChoice[]& Choices)\n
"},{"location":"xmldocs/llama.oldversion.completionchunk/#parameters_4","title":"Parameters","text":"

Id String&

Object String&

Created Int32&

Model String&

Choices CompletionChoice[]&

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/","title":"CompletionLogprobs","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class CompletionLogprobs : System.IEquatable`1[[LLama.OldVersion.CompletionLogprobs, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 CompletionLogprobs Implements IEquatable<CompletionLogprobs>

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.completionlogprobs/#textoffset","title":"TextOffset","text":"
public Int32[] TextOffset { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#property-value","title":"Property Value","text":"

Int32[]

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#tokenlogprobs","title":"TokenLogProbs","text":"
public Single[] TokenLogProbs { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#property-value_1","title":"Property Value","text":"

Single[]

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#tokens","title":"Tokens","text":"
public String[] Tokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#property-value_2","title":"Property Value","text":"

String[]

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#toplogprobs","title":"TopLogprobs","text":"
public Dictionary`2[] TopLogprobs { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#property-value_3","title":"Property Value","text":"

Dictionary`2[]

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.completionlogprobs/#completionlogprobsint32-single-string-dictionary2","title":"CompletionLogprobs(Int32[], Single[], String[], Dictionary`2[])","text":"
public CompletionLogprobs(Int32[] TextOffset, Single[] TokenLogProbs, String[] Tokens, Dictionary`2[] TopLogprobs)\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#parameters","title":"Parameters","text":"

TextOffset Int32[]

TokenLogProbs Single[]

Tokens String[]

TopLogprobs Dictionary`2[]

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.completionlogprobs/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#equalscompletionlogprobs","title":"Equals(CompletionLogprobs)","text":"
public bool Equals(CompletionLogprobs other)\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#parameters_3","title":"Parameters","text":"

other CompletionLogprobs

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#clone","title":"<Clone>$()","text":"
public CompletionLogprobs <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#returns_5","title":"Returns","text":"

CompletionLogprobs

"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#deconstructint32-single-string-dictionary2","title":"Deconstruct(Int32[]&, Single[]&, String[]&, Dictionary`2[]&)","text":"
public void Deconstruct(Int32[]& TextOffset, Single[]& TokenLogProbs, String[]& Tokens, Dictionary`2[]& TopLogprobs)\n
"},{"location":"xmldocs/llama.oldversion.completionlogprobs/#parameters_4","title":"Parameters","text":"

TextOffset Int32[]&

TokenLogProbs Single[]&

Tokens String[]&

TopLogprobs Dictionary`2[]&

"},{"location":"xmldocs/llama.oldversion.completionusage/","title":"CompletionUsage","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.completionusage/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class CompletionUsage : System.IEquatable`1[[LLama.OldVersion.CompletionUsage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 CompletionUsage Implements IEquatable<CompletionUsage>

"},{"location":"xmldocs/llama.oldversion.completionusage/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.completionusage/#prompttokens","title":"PromptTokens","text":"
public int PromptTokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionusage/#completiontokens","title":"CompletionTokens","text":"
public int CompletionTokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionusage/#totaltokens","title":"TotalTokens","text":"
public int TotalTokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#property-value_2","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionusage/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.completionusage/#completionusageint32-int32-int32","title":"CompletionUsage(Int32, Int32, Int32)","text":"
public CompletionUsage(int PromptTokens, int CompletionTokens, int TotalTokens)\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#parameters","title":"Parameters","text":"

PromptTokens Int32

CompletionTokens Int32

TotalTokens Int32

"},{"location":"xmldocs/llama.oldversion.completionusage/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.completionusage/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.completionusage/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.completionusage/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionusage/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.completionusage/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.completionusage/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionusage/#equalscompletionusage","title":"Equals(CompletionUsage)","text":"
public bool Equals(CompletionUsage other)\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#parameters_3","title":"Parameters","text":"

other CompletionUsage

"},{"location":"xmldocs/llama.oldversion.completionusage/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.completionusage/#clone","title":"<Clone>$()","text":"
public CompletionUsage <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#returns_5","title":"Returns","text":"

CompletionUsage

"},{"location":"xmldocs/llama.oldversion.completionusage/#deconstructint32-int32-int32","title":"Deconstruct(Int32&, Int32&, Int32&)","text":"
public void Deconstruct(Int32& PromptTokens, Int32& CompletionTokens, Int32& TotalTokens)\n
"},{"location":"xmldocs/llama.oldversion.completionusage/#parameters_4","title":"Parameters","text":"

PromptTokens Int32&

CompletionTokens Int32&

TotalTokens Int32&

"},{"location":"xmldocs/llama.oldversion.embedding/","title":"Embedding","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.embedding/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class Embedding : System.IEquatable`1[[LLama.OldVersion.Embedding, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 Embedding Implements IEquatable<Embedding>

"},{"location":"xmldocs/llama.oldversion.embedding/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.embedding/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embedding/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.embedding/#model","title":"Model","text":"
public string Model { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embedding/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.embedding/#data","title":"Data","text":"
public EmbeddingData[] Data { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embedding/#property-value_2","title":"Property Value","text":"

EmbeddingData[]

"},{"location":"xmldocs/llama.oldversion.embedding/#usage","title":"Usage","text":"
public EmbeddingUsage Usage { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embedding/#property-value_3","title":"Property Value","text":"

EmbeddingUsage

"},{"location":"xmldocs/llama.oldversion.embedding/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.embedding/#embeddingstring-string-embeddingdata-embeddingusage","title":"Embedding(String, String, EmbeddingData[], EmbeddingUsage)","text":"
public Embedding(string Object, string Model, EmbeddingData[] Data, EmbeddingUsage Usage)\n
"},{"location":"xmldocs/llama.oldversion.embedding/#parameters","title":"Parameters","text":"

Object String

Model String

Data EmbeddingData[]

Usage EmbeddingUsage

"},{"location":"xmldocs/llama.oldversion.embedding/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.embedding/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.embedding/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.embedding/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.embedding/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.embedding/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embedding/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.embedding/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embedding/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.embedding/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.embedding/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embedding/#equalsembedding","title":"Equals(Embedding)","text":"
public bool Equals(Embedding other)\n
"},{"location":"xmldocs/llama.oldversion.embedding/#parameters_3","title":"Parameters","text":"

other Embedding

"},{"location":"xmldocs/llama.oldversion.embedding/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embedding/#clone","title":"<Clone>$()","text":"
public Embedding <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.embedding/#returns_5","title":"Returns","text":"

Embedding

"},{"location":"xmldocs/llama.oldversion.embedding/#deconstructstring-string-embeddingdata-embeddingusage","title":"Deconstruct(String&, String&, EmbeddingData[]&, EmbeddingUsage&)","text":"
public void Deconstruct(String& Object, String& Model, EmbeddingData[]& Data, EmbeddingUsage& Usage)\n
"},{"location":"xmldocs/llama.oldversion.embedding/#parameters_4","title":"Parameters","text":"

Object String&

Model String&

Data EmbeddingData[]&

Usage EmbeddingUsage&

"},{"location":"xmldocs/llama.oldversion.embeddingdata/","title":"EmbeddingData","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class EmbeddingData : System.IEquatable`1[[LLama.OldVersion.EmbeddingData, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 EmbeddingData Implements IEquatable<EmbeddingData>

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.embeddingdata/#index","title":"Index","text":"
public int Index { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#object","title":"Object","text":"
public string Object { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#property-value_1","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#embedding","title":"Embedding","text":"
public Single[] Embedding { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#property-value_2","title":"Property Value","text":"

Single[]

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.embeddingdata/#embeddingdataint32-string-single","title":"EmbeddingData(Int32, String, Single[])","text":"
public EmbeddingData(int Index, string Object, Single[] Embedding)\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#parameters","title":"Parameters","text":"

Index Int32

Object String

Embedding Single[]

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.embeddingdata/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#equalsembeddingdata","title":"Equals(EmbeddingData)","text":"
public bool Equals(EmbeddingData other)\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#parameters_3","title":"Parameters","text":"

other EmbeddingData

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#clone","title":"<Clone>$()","text":"
public EmbeddingData <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#returns_5","title":"Returns","text":"

EmbeddingData

"},{"location":"xmldocs/llama.oldversion.embeddingdata/#deconstructint32-string-single","title":"Deconstruct(Int32&, String&, Single[]&)","text":"
public void Deconstruct(Int32& Index, String& Object, Single[]& Embedding)\n
"},{"location":"xmldocs/llama.oldversion.embeddingdata/#parameters_4","title":"Parameters","text":"

Index Int32&

Object String&

Embedding Single[]&

"},{"location":"xmldocs/llama.oldversion.embeddingusage/","title":"EmbeddingUsage","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class EmbeddingUsage : System.IEquatable`1[[LLama.OldVersion.EmbeddingUsage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]\n

Inheritance Object \u2192 EmbeddingUsage Implements IEquatable<EmbeddingUsage>

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.embeddingusage/#prompttokens","title":"PromptTokens","text":"
public int PromptTokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#property-value","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#totaltokens","title":"TotalTokens","text":"
public int TotalTokens { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#property-value_1","title":"Property Value","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.embeddingusage/#embeddingusageint32-int32","title":"EmbeddingUsage(Int32, Int32)","text":"
public EmbeddingUsage(int PromptTokens, int TotalTokens)\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#parameters","title":"Parameters","text":"

PromptTokens Int32

TotalTokens Int32

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.embeddingusage/#tostring","title":"ToString()","text":"
public string ToString()\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#printmembersstringbuilder","title":"PrintMembers(StringBuilder)","text":"
protected bool PrintMembers(StringBuilder builder)\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#parameters_1","title":"Parameters","text":"

builder StringBuilder

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#gethashcode","title":"GetHashCode()","text":"
public int GetHashCode()\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns_2","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#equalsobject","title":"Equals(Object)","text":"
public bool Equals(object obj)\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#parameters_2","title":"Parameters","text":"

obj Object

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns_3","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#equalsembeddingusage","title":"Equals(EmbeddingUsage)","text":"
public bool Equals(EmbeddingUsage other)\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#parameters_3","title":"Parameters","text":"

other EmbeddingUsage

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns_4","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#clone","title":"<Clone>$()","text":"
public EmbeddingUsage <Clone>$()\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#returns_5","title":"Returns","text":"

EmbeddingUsage

"},{"location":"xmldocs/llama.oldversion.embeddingusage/#deconstructint32-int32","title":"Deconstruct(Int32&, Int32&)","text":"
public void Deconstruct(Int32& PromptTokens, Int32& TotalTokens)\n
"},{"location":"xmldocs/llama.oldversion.embeddingusage/#parameters_4","title":"Parameters","text":"

PromptTokens Int32&

TotalTokens Int32&

"},{"location":"xmldocs/llama.oldversion.ichatmodel/","title":"IChatModel","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.ichatmodel/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public interface IChatModel\n
"},{"location":"xmldocs/llama.oldversion.ichatmodel/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.ichatmodel/#name","title":"Name","text":"
public abstract string Name { get; }\n
"},{"location":"xmldocs/llama.oldversion.ichatmodel/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.ichatmodel/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.ichatmodel/#chatstring-string-string","title":"Chat(String, String, String)","text":"
IEnumerable<string> Chat(string text, string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.ichatmodel/#parameters","title":"Parameters","text":"

text String

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.ichatmodel/#returns","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.oldversion.ichatmodel/#initchatpromptstring-string","title":"InitChatPrompt(String, String)","text":"

Init a prompt for chat and automatically produce the next prompt during the chat.

void InitChatPrompt(string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.ichatmodel/#parameters_1","title":"Parameters","text":"

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.ichatmodel/#initchatantipromptstring","title":"InitChatAntiprompt(String[])","text":"
void InitChatAntiprompt(String[] antiprompt)\n
"},{"location":"xmldocs/llama.oldversion.ichatmodel/#parameters_2","title":"Parameters","text":"

antiprompt String[]

"},{"location":"xmldocs/llama.oldversion.llamaembedder/","title":"LLamaEmbedder","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.llamaembedder/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class LLamaEmbedder : System.IDisposable\n

Inheritance Object \u2192 LLamaEmbedder Implements IDisposable

"},{"location":"xmldocs/llama.oldversion.llamaembedder/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.llamaembedder/#llamaembedderllamaparams","title":"LLamaEmbedder(LLamaParams)","text":"
public LLamaEmbedder(LLamaParams params)\n
"},{"location":"xmldocs/llama.oldversion.llamaembedder/#parameters","title":"Parameters","text":"

params LLamaParams

"},{"location":"xmldocs/llama.oldversion.llamaembedder/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.llamaembedder/#getembeddingsstring-int32-boolean-string","title":"GetEmbeddings(String, Int32, Boolean, String)","text":"
public Single[] GetEmbeddings(string text, int n_thread, bool add_bos, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamaembedder/#parameters_1","title":"Parameters","text":"

text String

n_thread Int32

add_bos Boolean

encoding String

"},{"location":"xmldocs/llama.oldversion.llamaembedder/#returns","title":"Returns","text":"

Single[]

"},{"location":"xmldocs/llama.oldversion.llamaembedder/#dispose","title":"Dispose()","text":"
public void Dispose()\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/","title":"LLamaModel","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.llamamodel/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public class LLamaModel : IChatModel, System.IDisposable\n

Inheritance Object \u2192 LLamaModel Implements IChatModel, IDisposable

"},{"location":"xmldocs/llama.oldversion.llamamodel/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.oldversion.llamamodel/#name","title":"Name","text":"
public string Name { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#property-value","title":"Property Value","text":"

String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#verbose","title":"Verbose","text":"
public bool Verbose { get; set; }\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#property-value_1","title":"Property Value","text":"

Boolean

"},{"location":"xmldocs/llama.oldversion.llamamodel/#nativehandle","title":"NativeHandle","text":"
public SafeLLamaContextHandle NativeHandle { get; }\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#property-value_2","title":"Property Value","text":"

SafeLLamaContextHandle

"},{"location":"xmldocs/llama.oldversion.llamamodel/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.llamamodel/#llamamodelstring-string-boolean-int32-int32-int32-int32-int32-int32-int32-dictionaryint32-single-int32-single-single-single-single-single-int32-single-single-int32-single-single-string-string-string-string-liststring-string-string-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-string","title":"LLamaModel(String, String, Boolean, Int32, Int32, Int32, Int32, Int32, Int32, Int32, Dictionary<Int32, Single>, Int32, Single, Single, Single, Single, Single, Int32, Single, Single, Int32, Single, Single, String, String, String, String, List<String>, String, String, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, String)","text":"

Please refer LLamaParams to find the meanings of each arg. Be sure to have set the n_gpu_layers, otherwise it will load 20 layers to gpu by default.

public LLamaModel(string model_path, string model_name, bool verbose, int seed, int n_threads, int n_predict, int n_ctx, int n_batch, int n_keep, int n_gpu_layers, Dictionary<int, float> logit_bias, int top_k, float top_p, float tfs_z, float typical_p, float temp, float repeat_penalty, int repeat_last_n, float frequency_penalty, float presence_penalty, int mirostat, float mirostat_tau, float mirostat_eta, string prompt, string path_session, string input_prefix, string input_suffix, List<string> antiprompt, string lora_adapter, string lora_base, bool memory_f16, bool random_prompt, bool use_color, bool interactive, bool embedding, bool interactive_first, bool prompt_cache_all, bool instruct, bool penalize_nl, bool perplexity, bool use_mmap, bool use_mlock, bool mem_test, bool verbose_prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters","title":"Parameters","text":"

model_path String The model file path.

model_name String The model name.

verbose Boolean Whether to print details when running the model.

seed Int32

n_threads Int32

n_predict Int32

n_ctx Int32

n_batch Int32

n_keep Int32

n_gpu_layers Int32

logit_bias Dictionary<Int32, Single>

top_k Int32

top_p Single

tfs_z Single

typical_p Single

temp Single

repeat_penalty Single

repeat_last_n Int32

frequency_penalty Single

presence_penalty Single

mirostat Int32

mirostat_tau Single

mirostat_eta Single

prompt String

path_session String

input_prefix String

input_suffix String

antiprompt List<String>

lora_adapter String

lora_base String

memory_f16 Boolean

random_prompt Boolean

use_color Boolean

interactive Boolean

embedding Boolean

interactive_first Boolean

prompt_cache_all Boolean

instruct Boolean

penalize_nl Boolean

perplexity Boolean

use_mmap Boolean

use_mlock Boolean

mem_test Boolean

verbose_prompt Boolean

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#llamamodelllamaparams-string-boolean-string","title":"LLamaModel(LLamaParams, String, Boolean, String)","text":"

Please refer LLamaParams to find the meanings of each arg. Be sure to have set the n_gpu_layers, otherwise it will load 20 layers to gpu by default.

public LLamaModel(LLamaParams params, string name, bool verbose, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_1","title":"Parameters","text":"

params LLamaParams The LLamaModel params

name String Model name

verbose Boolean Whether to output the detailed info.

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.oldversion.llamamodel/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.oldversion.llamamodel/#withpromptstring-string","title":"WithPrompt(String, String)","text":"

Apply a prompt to the model.

public LLamaModel WithPrompt(string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_2","title":"Parameters","text":"

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns","title":"Returns","text":"

LLamaModel

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions_1","title":"Exceptions","text":"

ArgumentException

"},{"location":"xmldocs/llama.oldversion.llamamodel/#withpromptfilestring","title":"WithPromptFile(String)","text":"

Apply the prompt file to the model.

public LLamaModel WithPromptFile(string promptFileName)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_3","title":"Parameters","text":"

promptFileName String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns_1","title":"Returns","text":"

LLamaModel

"},{"location":"xmldocs/llama.oldversion.llamamodel/#initchatpromptstring-string","title":"InitChatPrompt(String, String)","text":"
public void InitChatPrompt(string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_4","title":"Parameters","text":"

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#initchatantipromptstring","title":"InitChatAntiprompt(String[])","text":"
public void InitChatAntiprompt(String[] antiprompt)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_5","title":"Parameters","text":"

antiprompt String[]

"},{"location":"xmldocs/llama.oldversion.llamamodel/#chatstring-string-string","title":"Chat(String, String, String)","text":"

Chat with the LLaMa model under interactive mode.

public IEnumerable<string> Chat(string text, string prompt, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_6","title":"Parameters","text":"

text String

prompt String

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns_2","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions_2","title":"Exceptions","text":"

ArgumentException

"},{"location":"xmldocs/llama.oldversion.llamamodel/#savestatestring","title":"SaveState(String)","text":"

Save the state to specified path.

public void SaveState(string filename)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_7","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#loadstatestring-boolean","title":"LoadState(String, Boolean)","text":"

Load the state from specified path.

public void LoadState(string filename, bool clearPreviousEmbed)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_8","title":"Parameters","text":"

filename String

clearPreviousEmbed Boolean Whether to clear previous footprints of this model.

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions_3","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.oldversion.llamamodel/#tokenizestring-string","title":"Tokenize(String, String)","text":"

Tokenize a string.

public List<int> Tokenize(string text, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_9","title":"Parameters","text":"

text String The utf-8 encoded string to tokenize.

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns_3","title":"Returns","text":"

List<Int32> A list of tokens.

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions_4","title":"Exceptions","text":"

RuntimeError If the tokenization failed.

"},{"location":"xmldocs/llama.oldversion.llamamodel/#detokenizeienumerableint32","title":"DeTokenize(IEnumerable<Int32>)","text":"

Detokenize a list of tokens.

public string DeTokenize(IEnumerable<int> tokens)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_10","title":"Parameters","text":"

tokens IEnumerable<Int32> The list of tokens to detokenize.

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns_4","title":"Returns","text":"

String The detokenized string.

"},{"location":"xmldocs/llama.oldversion.llamamodel/#callstring-string","title":"Call(String, String)","text":"

Call the model to run inference.

public IEnumerable<string> Call(string text, string encoding)\n
"},{"location":"xmldocs/llama.oldversion.llamamodel/#parameters_11","title":"Parameters","text":"

text String

encoding String

"},{"location":"xmldocs/llama.oldversion.llamamodel/#returns_5","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.oldversion.llamamodel/#exceptions_5","title":"Exceptions","text":"

RuntimeError

"},{"location":"xmldocs/llama.oldversion.llamamodel/#dispose","title":"Dispose()","text":"
public void Dispose()\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/","title":"LLamaParams","text":"

Namespace: LLama.OldVersion

"},{"location":"xmldocs/llama.oldversion.llamaparams/#caution","title":"Caution","text":"

The entire LLama.OldVersion namespace will be removed

public struct LLamaParams\n

Inheritance Object \u2192 ValueType \u2192 LLamaParams

"},{"location":"xmldocs/llama.oldversion.llamaparams/#fields","title":"Fields","text":""},{"location":"xmldocs/llama.oldversion.llamaparams/#seed","title":"seed","text":"
public int seed;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_threads","title":"n_threads","text":"
public int n_threads;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_predict","title":"n_predict","text":"
public int n_predict;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_ctx","title":"n_ctx","text":"
public int n_ctx;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_batch","title":"n_batch","text":"
public int n_batch;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_keep","title":"n_keep","text":"
public int n_keep;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#n_gpu_layers","title":"n_gpu_layers","text":"
public int n_gpu_layers;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#logit_bias","title":"logit_bias","text":"
public Dictionary<int, float> logit_bias;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#top_k","title":"top_k","text":"
public int top_k;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#top_p","title":"top_p","text":"
public float top_p;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#tfs_z","title":"tfs_z","text":"
public float tfs_z;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#typical_p","title":"typical_p","text":"
public float typical_p;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#temp","title":"temp","text":"
public float temp;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#repeat_penalty","title":"repeat_penalty","text":"
public float repeat_penalty;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#repeat_last_n","title":"repeat_last_n","text":"
public int repeat_last_n;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#frequency_penalty","title":"frequency_penalty","text":"
public float frequency_penalty;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#presence_penalty","title":"presence_penalty","text":"
public float presence_penalty;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#mirostat","title":"mirostat","text":"
public int mirostat;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#mirostat_tau","title":"mirostat_tau","text":"
public float mirostat_tau;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#mirostat_eta","title":"mirostat_eta","text":"
public float mirostat_eta;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#model","title":"model","text":"
public string model;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#prompt","title":"prompt","text":"
public string prompt;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#path_session","title":"path_session","text":"
public string path_session;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#input_prefix","title":"input_prefix","text":"
public string input_prefix;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#input_suffix","title":"input_suffix","text":"
public string input_suffix;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#antiprompt","title":"antiprompt","text":"
public List<string> antiprompt;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#lora_adapter","title":"lora_adapter","text":"
public string lora_adapter;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#lora_base","title":"lora_base","text":"
public string lora_base;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#memory_f16","title":"memory_f16","text":"
public bool memory_f16;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#random_prompt","title":"random_prompt","text":"
public bool random_prompt;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#use_color","title":"use_color","text":"
public bool use_color;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#interactive","title":"interactive","text":"
public bool interactive;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#prompt_cache_all","title":"prompt_cache_all","text":"
public bool prompt_cache_all;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#embedding","title":"embedding","text":"
public bool embedding;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#interactive_first","title":"interactive_first","text":"
public bool interactive_first;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#instruct","title":"instruct","text":"
public bool instruct;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#penalize_nl","title":"penalize_nl","text":"
public bool penalize_nl;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#perplexity","title":"perplexity","text":"
public bool perplexity;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#use_mmap","title":"use_mmap","text":"
public bool use_mmap;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#use_mlock","title":"use_mlock","text":"
public bool use_mlock;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#mem_test","title":"mem_test","text":"
public bool mem_test;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#verbose_prompt","title":"verbose_prompt","text":"
public bool verbose_prompt;\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.oldversion.llamaparams/#llamaparamsint32-int32-int32-int32-int32-int32-int32-dictionaryint32-single-int32-single-single-single-single-single-int32-single-single-int32-single-single-string-string-string-string-string-liststring-string-string-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean-boolean","title":"LLamaParams(Int32, Int32, Int32, Int32, Int32, Int32, Int32, Dictionary<Int32, Single>, Int32, Single, Single, Single, Single, Single, Int32, Single, Single, Int32, Single, Single, String, String, String, String, String, List<String>, String, String, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean)","text":"
LLamaParams(int seed, int n_threads, int n_predict, int n_ctx, int n_batch, int n_keep, int n_gpu_layers, Dictionary<int, float> logit_bias, int top_k, float top_p, float tfs_z, float typical_p, float temp, float repeat_penalty, int repeat_last_n, float frequency_penalty, float presence_penalty, int mirostat, float mirostat_tau, float mirostat_eta, string model, string prompt, string path_session, string input_prefix, string input_suffix, List<string> antiprompt, string lora_adapter, string lora_base, bool memory_f16, bool random_prompt, bool use_color, bool interactive, bool prompt_cache_all, bool embedding, bool interactive_first, bool instruct, bool penalize_nl, bool perplexity, bool use_mmap, bool use_mlock, bool mem_test, bool verbose_prompt)\n
"},{"location":"xmldocs/llama.oldversion.llamaparams/#parameters","title":"Parameters","text":"

seed Int32

n_threads Int32

n_predict Int32

n_ctx Int32

n_batch Int32

n_keep Int32

n_gpu_layers Int32

logit_bias Dictionary<Int32, Single>

top_k Int32

top_p Single

tfs_z Single

typical_p Single

temp Single

repeat_penalty Single

repeat_last_n Int32

frequency_penalty Single

presence_penalty Single

mirostat Int32

mirostat_tau Single

mirostat_eta Single

model String

prompt String

path_session String

input_prefix String

input_suffix String

antiprompt List<String>

lora_adapter String

lora_base String

memory_f16 Boolean

random_prompt Boolean

use_color Boolean

interactive Boolean

prompt_cache_all Boolean

embedding Boolean

interactive_first Boolean

instruct Boolean

penalize_nl Boolean

perplexity Boolean

use_mmap Boolean

use_mlock Boolean

mem_test Boolean

verbose_prompt Boolean

"},{"location":"xmldocs/llama.statefulexecutorbase/","title":"StatefulExecutorBase","text":"

Namespace: LLama

The base class for stateful LLama executors.

public abstract class StatefulExecutorBase : LLama.Abstractions.ILLamaExecutor\n

Inheritance Object \u2192 StatefulExecutorBase Implements ILLamaExecutor

"},{"location":"xmldocs/llama.statefulexecutorbase/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.statefulexecutorbase/#context","title":"Context","text":"

The context used by the executor.

public LLamaContext Context { get; }\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#property-value","title":"Property Value","text":"

LLamaContext

"},{"location":"xmldocs/llama.statefulexecutorbase/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.statefulexecutorbase/#withsessionfilestring","title":"WithSessionFile(String)","text":"

This API is currently not verified.

public StatefulExecutorBase WithSessionFile(string filename)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.statefulexecutorbase/#returns","title":"Returns","text":"

StatefulExecutorBase

"},{"location":"xmldocs/llama.statefulexecutorbase/#exceptions","title":"Exceptions","text":"

ArgumentNullException

RuntimeError

"},{"location":"xmldocs/llama.statefulexecutorbase/#savesessionfilestring","title":"SaveSessionFile(String)","text":"

This API has not been verified currently.

public void SaveSessionFile(string filename)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_1","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.statefulexecutorbase/#handlerunoutofcontextint32","title":"HandleRunOutOfContext(Int32)","text":"

After running out of the context, take some tokens from the original prompt and recompute the logits in batches.

protected void HandleRunOutOfContext(int tokensToKeep)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_2","title":"Parameters","text":"

tokensToKeep Int32

"},{"location":"xmldocs/llama.statefulexecutorbase/#tryreusemathingprefix","title":"TryReuseMathingPrefix()","text":"

Try to reuse the matching prefix from the session file.

protected void TryReuseMathingPrefix()\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#getloopconditioninferstateargs","title":"GetLoopCondition(InferStateArgs)","text":"

Decide whether to continue the loop.

protected abstract bool GetLoopCondition(InferStateArgs args)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_3","title":"Parameters","text":"

args InferStateArgs

"},{"location":"xmldocs/llama.statefulexecutorbase/#returns_1","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.statefulexecutorbase/#preprocessinputsstring-inferstateargs","title":"PreprocessInputs(String, InferStateArgs)","text":"

Preprocess the inputs before the inference.

protected abstract void PreprocessInputs(string text, InferStateArgs args)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_4","title":"Parameters","text":"

text String

args InferStateArgs

"},{"location":"xmldocs/llama.statefulexecutorbase/#postprocessiinferenceparams-inferstateargs-ienumerable1","title":"PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)","text":"

Do some post processing after the inference.

protected abstract bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_5","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

extraOutputs IEnumerable`1&

"},{"location":"xmldocs/llama.statefulexecutorbase/#returns_2","title":"Returns","text":"

Boolean

"},{"location":"xmldocs/llama.statefulexecutorbase/#inferinternaliinferenceparams-inferstateargs","title":"InferInternal(IInferenceParams, InferStateArgs)","text":"

The core inference logic.

protected abstract void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_6","title":"Parameters","text":"

inferenceParams IInferenceParams

args InferStateArgs

"},{"location":"xmldocs/llama.statefulexecutorbase/#savestatestring","title":"SaveState(String)","text":"

Save the current state to a file.

public abstract void SaveState(string filename)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_7","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.statefulexecutorbase/#getstatedata","title":"GetStateData()","text":"

Get the current state data.

public abstract ExecutorBaseState GetStateData()\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#returns_3","title":"Returns","text":"

ExecutorBaseState

"},{"location":"xmldocs/llama.statefulexecutorbase/#loadstateexecutorbasestate","title":"LoadState(ExecutorBaseState)","text":"

Load the state from data.

public abstract void LoadState(ExecutorBaseState data)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_8","title":"Parameters","text":"

data ExecutorBaseState

"},{"location":"xmldocs/llama.statefulexecutorbase/#loadstatestring","title":"LoadState(String)","text":"

Load the state from a file.

public abstract void LoadState(string filename)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_9","title":"Parameters","text":"

filename String

"},{"location":"xmldocs/llama.statefulexecutorbase/#inferstring-iinferenceparams-cancellationtoken","title":"Infer(String, IInferenceParams, CancellationToken)","text":"

Execute the inference.

public IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_10","title":"Parameters","text":"

text String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.statefulexecutorbase/#returns_4","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.statefulexecutorbase/#inferasyncstring-iinferenceparams-cancellationtoken","title":"InferAsync(String, IInferenceParams, CancellationToken)","text":"

Execute the inference asynchronously.

public IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.statefulexecutorbase/#parameters_11","title":"Parameters","text":"

text String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.statefulexecutorbase/#returns_5","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.statelessexecutor/","title":"StatelessExecutor","text":"

Namespace: LLama

This executor infer the input as one-time job. Previous inputs won't impact on the response to current input.

public class StatelessExecutor : LLama.Abstractions.ILLamaExecutor\n

Inheritance Object \u2192 StatelessExecutor Implements ILLamaExecutor

"},{"location":"xmldocs/llama.statelessexecutor/#properties","title":"Properties","text":""},{"location":"xmldocs/llama.statelessexecutor/#context","title":"Context","text":"

The context used by the executor when running the inference.

public LLamaContext Context { get; private set; }\n
"},{"location":"xmldocs/llama.statelessexecutor/#property-value","title":"Property Value","text":"

LLamaContext

"},{"location":"xmldocs/llama.statelessexecutor/#constructors","title":"Constructors","text":""},{"location":"xmldocs/llama.statelessexecutor/#statelessexecutorllamaweights-imodelparams","title":"StatelessExecutor(LLamaWeights, IModelParams)","text":"

Create a new stateless executor which will use the given model

public StatelessExecutor(LLamaWeights weights, IModelParams params)\n
"},{"location":"xmldocs/llama.statelessexecutor/#parameters","title":"Parameters","text":"

weights LLamaWeights

params IModelParams

"},{"location":"xmldocs/llama.statelessexecutor/#statelessexecutorllamacontext","title":"StatelessExecutor(LLamaContext)","text":""},{"location":"xmldocs/llama.statelessexecutor/#caution","title":"Caution","text":"

Use the constructor which automatically creates contexts using the LLamaWeights

Create a new stateless executor which will use the model used to create the given context

public StatelessExecutor(LLamaContext context)\n
"},{"location":"xmldocs/llama.statelessexecutor/#parameters_1","title":"Parameters","text":"

context LLamaContext

"},{"location":"xmldocs/llama.statelessexecutor/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.statelessexecutor/#inferstring-iinferenceparams-cancellationtoken","title":"Infer(String, IInferenceParams, CancellationToken)","text":"
public IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.statelessexecutor/#parameters_2","title":"Parameters","text":"

text String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.statelessexecutor/#returns","title":"Returns","text":"

IEnumerable<String>

"},{"location":"xmldocs/llama.statelessexecutor/#inferasyncstring-iinferenceparams-cancellationtoken","title":"InferAsync(String, IInferenceParams, CancellationToken)","text":"
public IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)\n
"},{"location":"xmldocs/llama.statelessexecutor/#parameters_3","title":"Parameters","text":"

text String

inferenceParams IInferenceParams

cancellationToken CancellationToken

"},{"location":"xmldocs/llama.statelessexecutor/#returns_1","title":"Returns","text":"

IAsyncEnumerable<String>

"},{"location":"xmldocs/llama.utils/","title":"Utils","text":"

Namespace: LLama

Assorted llama utilities

public static class Utils\n

Inheritance Object \u2192 Utils

"},{"location":"xmldocs/llama.utils/#methods","title":"Methods","text":""},{"location":"xmldocs/llama.utils/#initllamacontextfrommodelparamsimodelparams","title":"InitLLamaContextFromModelParams(IModelParams)","text":""},{"location":"xmldocs/llama.utils/#caution","title":"Caution","text":"

Use LLamaWeights.LoadFromFile and LLamaWeights.CreateContext instead

public static SafeLLamaContextHandle InitLLamaContextFromModelParams(IModelParams params)\n
"},{"location":"xmldocs/llama.utils/#parameters","title":"Parameters","text":"

params IModelParams

"},{"location":"xmldocs/llama.utils/#returns","title":"Returns","text":"

SafeLLamaContextHandle

"},{"location":"xmldocs/llama.utils/#tokenizesafellamacontexthandle-string-boolean-encoding","title":"Tokenize(SafeLLamaContextHandle, String, Boolean, Encoding)","text":""},{"location":"xmldocs/llama.utils/#caution_1","title":"Caution","text":"

Use SafeLLamaContextHandle Tokenize method instead

public static IEnumerable<int> Tokenize(SafeLLamaContextHandle ctx, string text, bool add_bos, Encoding encoding)\n
"},{"location":"xmldocs/llama.utils/#parameters_1","title":"Parameters","text":"

ctx SafeLLamaContextHandle

text String

add_bos Boolean

encoding Encoding

"},{"location":"xmldocs/llama.utils/#returns_1","title":"Returns","text":"

IEnumerable<Int32>

"},{"location":"xmldocs/llama.utils/#getlogitssafellamacontexthandle-int32","title":"GetLogits(SafeLLamaContextHandle, Int32)","text":""},{"location":"xmldocs/llama.utils/#caution_2","title":"Caution","text":"

Use SafeLLamaContextHandle GetLogits method instead

public static Span<float> GetLogits(SafeLLamaContextHandle ctx, int length)\n
"},{"location":"xmldocs/llama.utils/#parameters_2","title":"Parameters","text":"

ctx SafeLLamaContextHandle

length Int32

"},{"location":"xmldocs/llama.utils/#returns_2","title":"Returns","text":"

Span<Single>

"},{"location":"xmldocs/llama.utils/#evalsafellamacontexthandle-int32-int32-int32-int32-int32","title":"Eval(SafeLLamaContextHandle, Int32[], Int32, Int32, Int32, Int32)","text":""},{"location":"xmldocs/llama.utils/#caution_3","title":"Caution","text":"

Use SafeLLamaContextHandle Eval method instead

public static int Eval(SafeLLamaContextHandle ctx, Int32[] tokens, int startIndex, int n_tokens, int n_past, int n_threads)\n
"},{"location":"xmldocs/llama.utils/#parameters_3","title":"Parameters","text":"

ctx SafeLLamaContextHandle

tokens Int32[]

startIndex Int32

n_tokens Int32

n_past Int32

n_threads Int32

"},{"location":"xmldocs/llama.utils/#returns_3","title":"Returns","text":"

Int32

"},{"location":"xmldocs/llama.utils/#tokentostringint32-safellamacontexthandle-encoding","title":"TokenToString(Int32, SafeLLamaContextHandle, Encoding)","text":""},{"location":"xmldocs/llama.utils/#caution_4","title":"Caution","text":"

Use SafeLLamaContextHandle TokenToString method instead

public static string TokenToString(int token, SafeLLamaContextHandle ctx, Encoding encoding)\n
"},{"location":"xmldocs/llama.utils/#parameters_4","title":"Parameters","text":"

token Int32

ctx SafeLLamaContextHandle

encoding Encoding

"},{"location":"xmldocs/llama.utils/#returns_4","title":"Returns","text":"

String

"},{"location":"xmldocs/llama.utils/#ptrtostringintptr-encoding","title":"PtrToString(IntPtr, Encoding)","text":""},{"location":"xmldocs/llama.utils/#caution_5","title":"Caution","text":"

No longer used internally by LlamaSharp

public static string PtrToString(IntPtr ptr, Encoding encoding)\n
"},{"location":"xmldocs/llama.utils/#parameters_5","title":"Parameters","text":"

ptr IntPtr

encoding Encoding

"},{"location":"xmldocs/llama.utils/#returns_5","title":"Returns","text":"

String

"}]} \ No newline at end of file diff --git a/0.5/sitemap.xml b/0.5/sitemap.xml new file mode 100755 index 00000000..0f8724ef --- /dev/null +++ b/0.5/sitemap.xml @@ -0,0 +1,3 @@ + + + \ No newline at end of file diff --git a/0.5/sitemap.xml.gz b/0.5/sitemap.xml.gz new file mode 100755 index 00000000..e9e72e98 Binary files /dev/null and b/0.5/sitemap.xml.gz differ diff --git a/0.5/xmldocs/index.html b/0.5/xmldocs/index.html new file mode 100755 index 00000000..218ef0ae --- /dev/null +++ b/0.5/xmldocs/index.html @@ -0,0 +1,2267 @@ + + + + + + + + + + + + + + + + + + + + + + index - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaSharp

+

LLama

+

ChatSession

+

InstructExecutor

+

InteractiveExecutor

+

LLamaContext

+

LLamaEmbedder

+

LLamaQuantizer

+

LLamaTransforms

+

LLamaWeights

+

StatefulExecutorBase

+

StatelessExecutor

+

Utils

+

LLama.Abstractions

+

IHistoryTransform

+

IInferenceParams

+

ILLamaExecutor

+

IModelParams

+

ITextStreamTransform

+

ITextTransform

+

LLama.Common

+

AuthorRole

+

ChatHistory

+

FixedSizeQueue<T>

+

ILLamaLogger

+

InferenceParams

+

LLamaDefaultLogger

+

MirostatType

+

ModelParams

+

LLama.Exceptions

+

GrammarExpectedName

+

GrammarExpectedNext

+

GrammarExpectedPrevious

+

GrammarFormatException

+

GrammarUnexpectedCharAltElement

+

GrammarUnexpectedCharRngElement

+

GrammarUnexpectedEndElement

+

GrammarUnexpectedEndOfInput

+

GrammarUnexpectedHexCharsCount

+

GrammarUnknownEscapeCharacter

+

RuntimeError

+

LLama.Extensions

+

IModelParamsExtensions

+

KeyValuePairExtensions

+

LLama.Grammars

+

Grammar

+

GrammarRule

+

LLama.Native

+

LLamaContextParams

+

LLamaFtype

+

LLamaGrammarElement

+

LLamaGrammarElementType

+

LLamaModelQuantizeParams

+

LLamaTokenData

+

LLamaTokenDataArray

+

LLamaTokenDataArrayNative

+

NativeApi

+

SafeLLamaContextHandle

+

SafeLLamaGrammarHandle

+

SafeLLamaHandleBase

+

SafeLlamaModelHandle

+

SamplingApi

+

LLama.OldVersion

+

ChatCompletion

+

ChatCompletionChoice

+

ChatCompletionChunk

+

ChatCompletionChunkChoice

+

ChatCompletionChunkDelta

+

ChatCompletionMessage

+

ChatMessageRecord

+

ChatRole

+

ChatSession<T>

+

Completion

+

CompletionChoice

+

CompletionChunk

+

CompletionLogprobs

+

CompletionUsage

+

Embedding

+

EmbeddingData

+

EmbeddingUsage

+

IChatModel

+

LLamaEmbedder

+

LLamaModel

+

LLamaParams

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.ihistorytransform/index.html b/0.5/xmldocs/llama.abstractions.ihistorytransform/index.html new file mode 100755 index 00000000..0d7e0194 --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.ihistorytransform/index.html @@ -0,0 +1,2232 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.ihistorytransform - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

IHistoryTransform

+

Namespace: LLama.Abstractions

+

Transform history to plain text and vice versa.

+
public interface IHistoryTransform
+
+

Methods

+

HistoryToText(ChatHistory)

+

Convert a ChatHistory instance to plain text.

+
string HistoryToText(ChatHistory history)
+
+

Parameters

+

history ChatHistory
+The ChatHistory instance

+

Returns

+

String

+

TextToHistory(AuthorRole, String)

+

Converts plain text to a ChatHistory instance.

+
ChatHistory TextToHistory(AuthorRole role, string text)
+
+

Parameters

+

role AuthorRole
+The role for the author.

+

text String
+The chat history as plain text.

+

Returns

+

ChatHistory
+The updated history.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.iinferenceparams/index.html b/0.5/xmldocs/llama.abstractions.iinferenceparams/index.html new file mode 100755 index 00000000..f7cf1f88 --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.iinferenceparams/index.html @@ -0,0 +1,3074 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.iinferenceparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

IInferenceParams

+

Namespace: LLama.Abstractions

+

The paramters used for inference.

+
public interface IInferenceParams
+
+

Properties

+

TokensKeep

+

number of tokens to keep from initial prompt

+
public abstract int TokensKeep { get; set; }
+
+

Property Value

+

Int32

+

MaxTokens

+

how many new tokens to predict (n_predict), set to -1 to inifinitely generate response + until it complete.

+
public abstract int MaxTokens { get; set; }
+
+

Property Value

+

Int32

+

LogitBias

+

logit bias for specific tokens

+
public abstract Dictionary<int, float> LogitBias { get; set; }
+
+

Property Value

+

Dictionary<Int32, Single>

+

AntiPrompts

+

Sequences where the model will stop generating further tokens.

+
public abstract IEnumerable<string> AntiPrompts { get; set; }
+
+

Property Value

+

IEnumerable<String>

+

PathSession

+

path to file for saving/loading model eval state

+
public abstract string PathSession { get; set; }
+
+

Property Value

+

String

+

InputSuffix

+

string to suffix user inputs with

+
public abstract string InputSuffix { get; set; }
+
+

Property Value

+

String

+

InputPrefix

+

string to prefix user inputs with

+
public abstract string InputPrefix { get; set; }
+
+

Property Value

+

String

+

TopK

+

0 or lower to use vocab size

+
public abstract int TopK { get; set; }
+
+

Property Value

+

Int32

+

TopP

+

1.0 = disabled

+
public abstract float TopP { get; set; }
+
+

Property Value

+

Single

+

TfsZ

+

1.0 = disabled

+
public abstract float TfsZ { get; set; }
+
+

Property Value

+

Single

+

TypicalP

+

1.0 = disabled

+
public abstract float TypicalP { get; set; }
+
+

Property Value

+

Single

+

Temperature

+

1.0 = disabled

+
public abstract float Temperature { get; set; }
+
+

Property Value

+

Single

+

RepeatPenalty

+

1.0 = disabled

+
public abstract float RepeatPenalty { get; set; }
+
+

Property Value

+

Single

+

RepeatLastTokensCount

+

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

+
public abstract int RepeatLastTokensCount { get; set; }
+
+

Property Value

+

Int32

+

FrequencyPenalty

+

frequency penalty coefficient + 0.0 = disabled

+
public abstract float FrequencyPenalty { get; set; }
+
+

Property Value

+

Single

+

PresencePenalty

+

presence penalty coefficient + 0.0 = disabled

+
public abstract float PresencePenalty { get; set; }
+
+

Property Value

+

Single

+

Mirostat

+

Mirostat uses tokens instead of words. + algorithm described in the paper https://arxiv.org/abs/2007.14966. + 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

+
public abstract MirostatType Mirostat { get; set; }
+
+

Property Value

+

MirostatType

+

MirostatTau

+

target entropy

+
public abstract float MirostatTau { get; set; }
+
+

Property Value

+

Single

+

MirostatEta

+

learning rate

+
public abstract float MirostatEta { get; set; }
+
+

Property Value

+

Single

+

PenalizeNL

+

consider newlines as a repeatable token (penalize_nl)

+
public abstract bool PenalizeNL { get; set; }
+
+

Property Value

+

Boolean

+

Grammar

+

Grammar to constrain possible tokens

+
public abstract SafeLLamaGrammarHandle Grammar { get; set; }
+
+

Property Value

+

SafeLLamaGrammarHandle

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.illamaexecutor/index.html b/0.5/xmldocs/llama.abstractions.illamaexecutor/index.html new file mode 100755 index 00000000..b05ee54e --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.illamaexecutor/index.html @@ -0,0 +1,2310 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.illamaexecutor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ILLamaExecutor

+

Namespace: LLama.Abstractions

+

A high level interface for LLama models.

+
public interface ILLamaExecutor
+
+

Properties

+

Context

+

The loaded context for this executor.

+
public abstract LLamaContext Context { get; }
+
+

Property Value

+

LLamaContext

+

Methods

+

Infer(String, IInferenceParams, CancellationToken)

+

Infers a response from the model.

+
IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken token)
+
+

Parameters

+

text String
+Your prompt

+

inferenceParams IInferenceParams
+Any additional parameters

+

token CancellationToken
+A cancellation token.

+

Returns

+

IEnumerable<String>

+

InferAsync(String, IInferenceParams, CancellationToken)

+

Asynchronously infers a response from the model.

+
IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken token)
+
+

Parameters

+

text String
+Your prompt

+

inferenceParams IInferenceParams
+Any additional parameters

+

token CancellationToken
+A cancellation token.

+

Returns

+

IAsyncEnumerable<String>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.imodelparams/index.html b/0.5/xmldocs/llama.abstractions.imodelparams/index.html new file mode 100755 index 00000000..94aacf6f --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.imodelparams/index.html @@ -0,0 +1,3116 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.imodelparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

IModelParams

+

Namespace: LLama.Abstractions

+

The parameters for initializing a LLama model.

+
public interface IModelParams
+
+

Properties

+

ContextSize

+

Model context size (n_ctx)

+
public abstract int ContextSize { get; set; }
+
+

Property Value

+

Int32

+

MainGpu

+

the GPU that is used for scratch and small tensors

+
public abstract int MainGpu { get; set; }
+
+

Property Value

+

Int32

+

LowVram

+

if true, reduce VRAM usage at the cost of performance

+
public abstract bool LowVram { get; set; }
+
+

Property Value

+

Boolean

+

GpuLayerCount

+

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

+
public abstract int GpuLayerCount { get; set; }
+
+

Property Value

+

Int32

+

Seed

+

Seed for the random number generator (seed)

+
public abstract int Seed { get; set; }
+
+

Property Value

+

Int32

+

UseFp16Memory

+

Use f16 instead of f32 for memory kv (memory_f16)

+
public abstract bool UseFp16Memory { get; set; }
+
+

Property Value

+

Boolean

+

UseMemorymap

+

Use mmap for faster loads (use_mmap)

+
public abstract bool UseMemorymap { get; set; }
+
+

Property Value

+

Boolean

+

UseMemoryLock

+

Use mlock to keep model in memory (use_mlock)

+
public abstract bool UseMemoryLock { get; set; }
+
+

Property Value

+

Boolean

+

Perplexity

+

Compute perplexity over the prompt (perplexity)

+
public abstract bool Perplexity { get; set; }
+
+

Property Value

+

Boolean

+

ModelPath

+

Model path (model)

+
public abstract string ModelPath { get; set; }
+
+

Property Value

+

String

+

ModelAlias

+

model alias

+
public abstract string ModelAlias { get; set; }
+
+

Property Value

+

String

+

LoraAdapter

+

lora adapter path (lora_adapter)

+
public abstract string LoraAdapter { get; set; }
+
+

Property Value

+

String

+

LoraBase

+

base model path for the lora adapter (lora_base)

+
public abstract string LoraBase { get; set; }
+
+

Property Value

+

String

+

Threads

+

Number of threads (-1 = autodetect) (n_threads)

+
public abstract int Threads { get; set; }
+
+

Property Value

+

Int32

+

BatchSize

+

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

+
public abstract int BatchSize { get; set; }
+
+

Property Value

+

Int32

+

ConvertEosToNewLine

+

Whether to convert eos to newline during the inference.

+
public abstract bool ConvertEosToNewLine { get; set; }
+
+

Property Value

+

Boolean

+

EmbeddingMode

+

Whether to use embedding mode. (embedding) Note that if this is set to true, + The LLamaModel won't produce text response anymore.

+
public abstract bool EmbeddingMode { get; set; }
+
+

Property Value

+

Boolean

+

TensorSplits

+

how split tensors should be distributed across GPUs

+
public abstract Single[] TensorSplits { get; set; }
+
+

Property Value

+

Single[]

+

RopeFrequencyBase

+

RoPE base frequency

+
public abstract float RopeFrequencyBase { get; set; }
+
+

Property Value

+

Single

+

RopeFrequencyScale

+

RoPE frequency scaling factor

+
public abstract float RopeFrequencyScale { get; set; }
+
+

Property Value

+

Single

+

MulMatQ

+

Use experimental mul_mat_q kernels

+
public abstract bool MulMatQ { get; set; }
+
+

Property Value

+

Boolean

+

Encoding

+

The encoding to use for models

+
public abstract Encoding Encoding { get; set; }
+
+

Property Value

+

Encoding

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.itextstreamtransform/index.html b/0.5/xmldocs/llama.abstractions.itextstreamtransform/index.html new file mode 100755 index 00000000..2b3ab231 --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.itextstreamtransform/index.html @@ -0,0 +1,2227 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.itextstreamtransform - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ITextStreamTransform

+

Namespace: LLama.Abstractions

+

Takes a stream of tokens and transforms them.

+
public interface ITextStreamTransform
+
+

Methods

+

Transform(IEnumerable<String>)

+

Takes a stream of tokens and transforms them, returning a new stream of tokens.

+
IEnumerable<string> Transform(IEnumerable<string> tokens)
+
+

Parameters

+

tokens IEnumerable<String>

+

Returns

+

IEnumerable<String>

+

TransformAsync(IAsyncEnumerable<String>)

+

Takes a stream of tokens and transforms them, returning a new stream of tokens asynchronously.

+
IAsyncEnumerable<string> TransformAsync(IAsyncEnumerable<string> tokens)
+
+

Parameters

+

tokens IAsyncEnumerable<String>

+

Returns

+

IAsyncEnumerable<String>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.abstractions.itexttransform/index.html b/0.5/xmldocs/llama.abstractions.itexttransform/index.html new file mode 100755 index 00000000..a9ee5d91 --- /dev/null +++ b/0.5/xmldocs/llama.abstractions.itexttransform/index.html @@ -0,0 +1,2171 @@ + + + + + + + + + + + + + + + + + + + + + + llama.abstractions.itexttransform - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ITextTransform

+

Namespace: LLama.Abstractions

+

An interface for text transformations. + These can be used to compose a pipeline of text transformations, such as: + - Tokenization + - Lowercasing + - Punctuation removal + - Trimming + - etc.

+
public interface ITextTransform
+
+

Methods

+

Transform(String)

+

Takes a string and transforms it.

+
string Transform(string text)
+
+

Parameters

+

text String

+

Returns

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.chatsession/index.html b/0.5/xmldocs/llama.chatsession/index.html new file mode 100755 index 00000000..f25d5d84 --- /dev/null +++ b/0.5/xmldocs/llama.chatsession/index.html @@ -0,0 +1,2968 @@ + + + + + + + + + + + + + + + + + + + + + + llama.chatsession - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatSession

+

Namespace: LLama

+

The main chat session class.

+
public class ChatSession
+
+

Inheritance ObjectChatSession

+

Fields

+

OutputTransform

+

The output transform used in this session.

+
public ITextStreamTransform OutputTransform;
+
+

Properties

+

Executor

+

The executor for this session.

+
public ILLamaExecutor Executor { get; }
+
+

Property Value

+

ILLamaExecutor

+

History

+

The chat history for this session.

+
public ChatHistory History { get; }
+
+

Property Value

+

ChatHistory

+

HistoryTransform

+

The history transform used in this session.

+
public IHistoryTransform HistoryTransform { get; set; }
+
+

Property Value

+

IHistoryTransform

+

InputTransformPipeline

+

The input transform pipeline used in this session.

+
public List<ITextTransform> InputTransformPipeline { get; set; }
+
+

Property Value

+

List<ITextTransform>

+

Constructors

+

ChatSession(ILLamaExecutor)

+
public ChatSession(ILLamaExecutor executor)
+
+

Parameters

+

executor ILLamaExecutor
+The executor for this session

+

Methods

+

WithHistoryTransform(IHistoryTransform)

+

Use a custom history transform.

+
public ChatSession WithHistoryTransform(IHistoryTransform transform)
+
+

Parameters

+

transform IHistoryTransform

+

Returns

+

ChatSession

+

AddInputTransform(ITextTransform)

+

Add a text transform to the input transform pipeline.

+
public ChatSession AddInputTransform(ITextTransform transform)
+
+

Parameters

+

transform ITextTransform

+

Returns

+

ChatSession

+

WithOutputTransform(ITextStreamTransform)

+

Use a custom output transform.

+
public ChatSession WithOutputTransform(ITextStreamTransform transform)
+
+

Parameters

+

transform ITextStreamTransform

+

Returns

+

ChatSession

+

SaveSession(String)

+
public void SaveSession(string path)
+
+

Parameters

+

path String
+The directory name to save the session. If the directory does not exist, a new directory will be created.

+

LoadSession(String)

+
public void LoadSession(string path)
+
+

Parameters

+

path String
+The directory name to load the session.

+

Chat(ChatHistory, IInferenceParams, CancellationToken)

+

Get the response from the LLama model with chat histories.

+
public IEnumerable<string> Chat(ChatHistory history, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

history ChatHistory

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IEnumerable<String>

+

Chat(String, IInferenceParams, CancellationToken)

+

Get the response from the LLama model. Note that prompt could not only be the preset words, + but also the question you want to ask.

+
public IEnumerable<string> Chat(string prompt, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

prompt String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IEnumerable<String>

+

ChatAsync(ChatHistory, IInferenceParams, CancellationToken)

+

Get the response from the LLama model with chat histories.

+
public IAsyncEnumerable<string> ChatAsync(ChatHistory history, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

history ChatHistory

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IAsyncEnumerable<String>

+

ChatAsync(String, IInferenceParams, CancellationToken)

+

Get the response from the LLama model with chat histories asynchronously.

+
public IAsyncEnumerable<string> ChatAsync(string prompt, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

prompt String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IAsyncEnumerable<String>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.authorrole/index.html b/0.5/xmldocs/llama.common.authorrole/index.html new file mode 100755 index 00000000..d1b0b2d1 --- /dev/null +++ b/0.5/xmldocs/llama.common.authorrole/index.html @@ -0,0 +1,2124 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.authorrole - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

AuthorRole

+

Namespace: LLama.Common

+

Role of the message author, e.g. user/assistant/system

+
public enum AuthorRole
+
+

Inheritance ObjectValueTypeEnumAuthorRole
+Implements IComparable, IFormattable, IConvertible

+

Fields

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameValueDescription
Unknown-1Role is unknown
System0Message comes from a "system" prompt, not written by a user or language model
User1Message comes from the user
Assistant2Messages was generated by the language model
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.chathistory/index.html b/0.5/xmldocs/llama.common.chathistory/index.html new file mode 100755 index 00000000..4751ccf7 --- /dev/null +++ b/0.5/xmldocs/llama.common.chathistory/index.html @@ -0,0 +1,2271 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.chathistory - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatHistory

+

Namespace: LLama.Common

+

The chat history class

+
public class ChatHistory
+
+

Inheritance ObjectChatHistory

+

Properties

+

Messages

+

List of messages in the chat

+
public List<Message> Messages { get; }
+
+

Property Value

+

List<Message>

+

Constructors

+

ChatHistory()

+

Create a new instance of the chat content class

+
public ChatHistory()
+
+

Methods

+

AddMessage(AuthorRole, String)

+

Add a message to the chat history

+
public void AddMessage(AuthorRole authorRole, string content)
+
+

Parameters

+

authorRole AuthorRole
+Role of the message author

+

content String
+Message content

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.fixedsizequeue-1/index.html b/0.5/xmldocs/llama.common.fixedsizequeue-1/index.html new file mode 100755 index 00000000..9e8307f6 --- /dev/null +++ b/0.5/xmldocs/llama.common.fixedsizequeue-1/index.html @@ -0,0 +1,2517 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.fixedsizequeue-1 - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

FixedSizeQueue<T>

+

Namespace: LLama.Common

+

A queue with fixed storage size. + Currently it's only a naive implementation and needs to be further optimized in the future.

+
public class FixedSizeQueue<T> : , System.Collections.IEnumerable
+
+

Type Parameters

+

T

+

Inheritance ObjectFixedSizeQueue<T>
+Implements IEnumerable<T>, IEnumerable

+

Properties

+

Count

+

Number of items in this queue

+
public int Count { get; }
+
+

Property Value

+

Int32

+

Capacity

+

Maximum number of items allowed in this queue

+
public int Capacity { get; }
+
+

Property Value

+

Int32

+

Constructors

+

FixedSizeQueue(Int32)

+

Create a new queue

+
public FixedSizeQueue(int size)
+
+

Parameters

+

size Int32
+the maximum number of items to store in this queue

+

FixedSizeQueue(Int32, IEnumerable<T>)

+

Fill the quene with the data. Please ensure that data.Count <= size

+
public FixedSizeQueue(int size, IEnumerable<T> data)
+
+

Parameters

+

size Int32

+

data IEnumerable<T>

+

Methods

+

FillWith(T)

+

Replace every item in the queue with the given value

+
public FixedSizeQueue<T> FillWith(T value)
+
+

Parameters

+

value T
+The value to replace all items with

+

Returns

+

FixedSizeQueue<T>
+returns this

+

Enqueue(T)

+

Enquene an element.

+
public void Enqueue(T item)
+
+

Parameters

+

item T

+

GetEnumerator()

+
public IEnumerator<T> GetEnumerator()
+
+

Returns

+

IEnumerator<T>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.illamalogger/index.html b/0.5/xmldocs/llama.common.illamalogger/index.html new file mode 100755 index 00000000..1ac5cf30 --- /dev/null +++ b/0.5/xmldocs/llama.common.illamalogger/index.html @@ -0,0 +1,2154 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.illamalogger - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ILLamaLogger

+

Namespace: LLama.Common

+

receives log messages from LLamaSharp

+
public interface ILLamaLogger
+
+

Methods

+

Log(String, String, LogLevel)

+

Write the log in customized way

+
void Log(string source, string message, LogLevel level)
+
+

Parameters

+

source String
+The source of the log. It may be a method name or class name.

+

message String
+The message.

+

level LogLevel
+The log level.

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.inferenceparams/index.html b/0.5/xmldocs/llama.common.inferenceparams/index.html new file mode 100755 index 00000000..de6f6736 --- /dev/null +++ b/0.5/xmldocs/llama.common.inferenceparams/index.html @@ -0,0 +1,3120 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.inferenceparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

InferenceParams

+

Namespace: LLama.Common

+

The paramters used for inference.

+
public class InferenceParams : LLama.Abstractions.IInferenceParams
+
+

Inheritance ObjectInferenceParams
+Implements IInferenceParams

+

Properties

+

TokensKeep

+

number of tokens to keep from initial prompt

+
public int TokensKeep { get; set; }
+
+

Property Value

+

Int32

+

MaxTokens

+

how many new tokens to predict (n_predict), set to -1 to inifinitely generate response + until it complete.

+
public int MaxTokens { get; set; }
+
+

Property Value

+

Int32

+

LogitBias

+

logit bias for specific tokens

+
public Dictionary<int, float> LogitBias { get; set; }
+
+

Property Value

+

Dictionary<Int32, Single>

+

AntiPrompts

+

Sequences where the model will stop generating further tokens.

+
public IEnumerable<string> AntiPrompts { get; set; }
+
+

Property Value

+

IEnumerable<String>

+

PathSession

+

path to file for saving/loading model eval state

+
public string PathSession { get; set; }
+
+

Property Value

+

String

+

InputSuffix

+

string to suffix user inputs with

+
public string InputSuffix { get; set; }
+
+

Property Value

+

String

+

InputPrefix

+

string to prefix user inputs with

+
public string InputPrefix { get; set; }
+
+

Property Value

+

String

+

TopK

+

0 or lower to use vocab size

+
public int TopK { get; set; }
+
+

Property Value

+

Int32

+

TopP

+

1.0 = disabled

+
public float TopP { get; set; }
+
+

Property Value

+

Single

+

TfsZ

+

1.0 = disabled

+
public float TfsZ { get; set; }
+
+

Property Value

+

Single

+

TypicalP

+

1.0 = disabled

+
public float TypicalP { get; set; }
+
+

Property Value

+

Single

+

Temperature

+

1.0 = disabled

+
public float Temperature { get; set; }
+
+

Property Value

+

Single

+

RepeatPenalty

+

1.0 = disabled

+
public float RepeatPenalty { get; set; }
+
+

Property Value

+

Single

+

RepeatLastTokensCount

+

last n tokens to penalize (0 = disable penalty, -1 = context size) (repeat_last_n)

+
public int RepeatLastTokensCount { get; set; }
+
+

Property Value

+

Int32

+

FrequencyPenalty

+

frequency penalty coefficient + 0.0 = disabled

+
public float FrequencyPenalty { get; set; }
+
+

Property Value

+

Single

+

PresencePenalty

+

presence penalty coefficient + 0.0 = disabled

+
public float PresencePenalty { get; set; }
+
+

Property Value

+

Single

+

Mirostat

+

Mirostat uses tokens instead of words. + algorithm described in the paper https://arxiv.org/abs/2007.14966. + 0 = disabled, 1 = mirostat, 2 = mirostat 2.0

+
public MirostatType Mirostat { get; set; }
+
+

Property Value

+

MirostatType

+

MirostatTau

+

target entropy

+
public float MirostatTau { get; set; }
+
+

Property Value

+

Single

+

MirostatEta

+

learning rate

+
public float MirostatEta { get; set; }
+
+

Property Value

+

Single

+

PenalizeNL

+

consider newlines as a repeatable token (penalize_nl)

+
public bool PenalizeNL { get; set; }
+
+

Property Value

+

Boolean

+

Grammar

+

A grammar to constrain the possible tokens

+
public SafeLLamaGrammarHandle Grammar { get; set; }
+
+

Property Value

+

SafeLLamaGrammarHandle

+

Constructors

+

InferenceParams()

+
public InferenceParams()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.llamadefaultlogger/index.html b/0.5/xmldocs/llama.common.llamadefaultlogger/index.html new file mode 100755 index 00000000..c4f38404 --- /dev/null +++ b/0.5/xmldocs/llama.common.llamadefaultlogger/index.html @@ -0,0 +1,2695 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.llamadefaultlogger - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaDefaultLogger

+

Namespace: LLama.Common

+

The default logger of LLamaSharp. On default it write to console. Use methods of LLamaLogger.Default to change the behavior. + It's recommended to inherit ILLamaLogger to customize the behavior.

+
public sealed class LLamaDefaultLogger : ILLamaLogger
+
+

Inheritance ObjectLLamaDefaultLogger
+Implements ILLamaLogger

+

Properties

+

Default

+

Get the default logger instance

+
public static LLamaDefaultLogger Default { get; }
+
+

Property Value

+

LLamaDefaultLogger

+

Methods

+

EnableNative()

+

Enable logging output from llama.cpp

+
public LLamaDefaultLogger EnableNative()
+
+

Returns

+

LLamaDefaultLogger

+

EnableConsole()

+

Enable writing log messages to console

+
public LLamaDefaultLogger EnableConsole()
+
+

Returns

+

LLamaDefaultLogger

+

DisableConsole()

+

Disable writing messages to console

+
public LLamaDefaultLogger DisableConsole()
+
+

Returns

+

LLamaDefaultLogger

+

EnableFile(String, FileMode)

+

Enable writing log messages to file

+
public LLamaDefaultLogger EnableFile(string filename, FileMode mode)
+
+

Parameters

+

filename String

+

mode FileMode

+

Returns

+

LLamaDefaultLogger

+

DisableFile(String)

+

Caution

+

Use DisableFile method without 'filename' parameter

+
+

Disable writing log messages to file

+
public LLamaDefaultLogger DisableFile(string filename)
+
+

Parameters

+

filename String
+unused!

+

Returns

+

LLamaDefaultLogger

+

DisableFile()

+

Disable writing log messages to file

+
public LLamaDefaultLogger DisableFile()
+
+

Returns

+

LLamaDefaultLogger

+

Log(String, String, LogLevel)

+

Log a message

+
public void Log(string source, string message, LogLevel level)
+
+

Parameters

+

source String
+The source of this message (e.g. class name)

+

message String
+The message to log

+

level LogLevel
+Severity level of this message

+

Info(String)

+

Write a log message with "Info" severity

+
public void Info(string message)
+
+

Parameters

+

message String

+

Warn(String)

+

Write a log message with "Warn" severity

+
public void Warn(string message)
+
+

Parameters

+

message String

+

Error(String)

+

Write a log message with "Error" severity

+
public void Error(string message)
+
+

Parameters

+

message String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.mirostattype/index.html b/0.5/xmldocs/llama.common.mirostattype/index.html new file mode 100755 index 00000000..f78111c7 --- /dev/null +++ b/0.5/xmldocs/llama.common.mirostattype/index.html @@ -0,0 +1,2120 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.mirostattype - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

MirostatType

+

Namespace: LLama.Common

+

Type of "mirostat" sampling to use. + https://github.com/basusourya/mirostat

+
public enum MirostatType
+
+

Inheritance ObjectValueTypeEnumMirostatType
+Implements IComparable, IFormattable, IConvertible

+

Fields

+ + + + + + + + + + + + + + + + + + + + + + + + + +
NameValueDescription
Disable0Disable Mirostat sampling
Mirostat1Original mirostat algorithm
Mirostat22Mirostat 2.0 algorithm
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.common.modelparams/index.html b/0.5/xmldocs/llama.common.modelparams/index.html new file mode 100755 index 00000000..df119e3c --- /dev/null +++ b/0.5/xmldocs/llama.common.modelparams/index.html @@ -0,0 +1,3633 @@ + + + + + + + + + + + + + + + + + + + + + + llama.common.modelparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ModelParams

+

Namespace: LLama.Common

+

The parameters for initializing a LLama model.

+
public class ModelParams : LLama.Abstractions.IModelParams, System.IEquatable`1[[LLama.Common.ModelParams, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectModelParams
+Implements IModelParams, IEquatable<ModelParams>

+

Properties

+

ContextSize

+

Model context size (n_ctx)

+
public int ContextSize { get; set; }
+
+

Property Value

+

Int32

+

MainGpu

+

the GPU that is used for scratch and small tensors

+
public int MainGpu { get; set; }
+
+

Property Value

+

Int32

+

LowVram

+

if true, reduce VRAM usage at the cost of performance

+
public bool LowVram { get; set; }
+
+

Property Value

+

Boolean

+

GpuLayerCount

+

Number of layers to run in VRAM / GPU memory (n_gpu_layers)

+
public int GpuLayerCount { get; set; }
+
+

Property Value

+

Int32

+

Seed

+

Seed for the random number generator (seed)

+
public int Seed { get; set; }
+
+

Property Value

+

Int32

+

UseFp16Memory

+

Use f16 instead of f32 for memory kv (memory_f16)

+
public bool UseFp16Memory { get; set; }
+
+

Property Value

+

Boolean

+

UseMemorymap

+

Use mmap for faster loads (use_mmap)

+
public bool UseMemorymap { get; set; }
+
+

Property Value

+

Boolean

+

UseMemoryLock

+

Use mlock to keep model in memory (use_mlock)

+
public bool UseMemoryLock { get; set; }
+
+

Property Value

+

Boolean

+

Perplexity

+

Compute perplexity over the prompt (perplexity)

+
public bool Perplexity { get; set; }
+
+

Property Value

+

Boolean

+

ModelPath

+

Model path (model)

+
public string ModelPath { get; set; }
+
+

Property Value

+

String

+

ModelAlias

+

model alias

+
public string ModelAlias { get; set; }
+
+

Property Value

+

String

+

LoraAdapter

+

lora adapter path (lora_adapter)

+
public string LoraAdapter { get; set; }
+
+

Property Value

+

String

+

LoraBase

+

base model path for the lora adapter (lora_base)

+
public string LoraBase { get; set; }
+
+

Property Value

+

String

+

Threads

+

Number of threads (-1 = autodetect) (n_threads)

+
public int Threads { get; set; }
+
+

Property Value

+

Int32

+

BatchSize

+

batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

+
public int BatchSize { get; set; }
+
+

Property Value

+

Int32

+

ConvertEosToNewLine

+

Whether to convert eos to newline during the inference.

+
public bool ConvertEosToNewLine { get; set; }
+
+

Property Value

+

Boolean

+

EmbeddingMode

+

Whether to use embedding mode. (embedding) Note that if this is set to true, + The LLamaModel won't produce text response anymore.

+
public bool EmbeddingMode { get; set; }
+
+

Property Value

+

Boolean

+

TensorSplits

+

how split tensors should be distributed across GPUs

+
public Single[] TensorSplits { get; set; }
+
+

Property Value

+

Single[]

+

RopeFrequencyBase

+

RoPE base frequency

+
public float RopeFrequencyBase { get; set; }
+
+

Property Value

+

Single

+

RopeFrequencyScale

+

RoPE frequency scaling factor

+
public float RopeFrequencyScale { get; set; }
+
+

Property Value

+

Single

+

MulMatQ

+

Use experimental mul_mat_q kernels

+
public bool MulMatQ { get; set; }
+
+

Property Value

+

Boolean

+

Encoding

+

The encoding to use to convert text for the model

+
public Encoding Encoding { get; set; }
+
+

Property Value

+

Encoding

+

Constructors

+

ModelParams(String)

+
public ModelParams(string modelPath)
+
+

Parameters

+

modelPath String
+The model path.

+

ModelParams(String, Int32, Int32, Int32, Boolean, Boolean, Boolean, Boolean, String, String, Int32, Int32, Boolean, Boolean, Single, Single, Boolean, String)

+

Caution

+

Use object initializer to set all optional parameters

+
+
public ModelParams(string modelPath, int contextSize, int gpuLayerCount, int seed, bool useFp16Memory, bool useMemorymap, bool useMemoryLock, bool perplexity, string loraAdapter, string loraBase, int threads, int batchSize, bool convertEosToNewLine, bool embeddingMode, float ropeFrequencyBase, float ropeFrequencyScale, bool mulMatQ, string encoding)
+
+

Parameters

+

modelPath String
+The model path.

+

contextSize Int32
+Model context size (n_ctx)

+

gpuLayerCount Int32
+Number of layers to run in VRAM / GPU memory (n_gpu_layers)

+

seed Int32
+Seed for the random number generator (seed)

+

useFp16Memory Boolean
+Whether to use f16 instead of f32 for memory kv (memory_f16)

+

useMemorymap Boolean
+Whether to use mmap for faster loads (use_mmap)

+

useMemoryLock Boolean
+Whether to use mlock to keep model in memory (use_mlock)

+

perplexity Boolean
+Thether to compute perplexity over the prompt (perplexity)

+

loraAdapter String
+Lora adapter path (lora_adapter)

+

loraBase String
+Base model path for the lora adapter (lora_base)

+

threads Int32
+Number of threads (-1 = autodetect) (n_threads)

+

batchSize Int32
+Batch size for prompt processing (must be >=32 to use BLAS) (n_batch)

+

convertEosToNewLine Boolean
+Whether to convert eos to newline during the inference.

+

embeddingMode Boolean
+Whether to use embedding mode. (embedding) Note that if this is set to true, The LLamaModel won't produce text response anymore.

+

ropeFrequencyBase Single
+RoPE base frequency.

+

ropeFrequencyScale Single
+RoPE frequency scaling factor

+

mulMatQ Boolean
+Use experimental mul_mat_q kernels

+

encoding String
+The encoding to use to convert text for the model

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ModelParams)

+
public bool Equals(ModelParams other)
+
+

Parameters

+

other ModelParams

+

Returns

+

Boolean

+

<Clone>$()

+
public ModelParams <Clone>$()
+
+

Returns

+

ModelParams

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarexpectedname/index.html b/0.5/xmldocs/llama.exceptions.grammarexpectedname/index.html new file mode 100755 index 00000000..b683447d --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarexpectedname/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarexpectedname - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarExpectedName

+

Namespace: LLama.Exceptions

+

Failed to parse a "name" element when one was expected

+
public class GrammarExpectedName : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarExpectedName
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarexpectednext/index.html b/0.5/xmldocs/llama.exceptions.grammarexpectednext/index.html new file mode 100755 index 00000000..4303527b --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarexpectednext/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarexpectednext - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarExpectedNext

+

Namespace: LLama.Exceptions

+

A specified string was expected when parsing

+
public class GrammarExpectedNext : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarExpectedNext
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarexpectedprevious/index.html b/0.5/xmldocs/llama.exceptions.grammarexpectedprevious/index.html new file mode 100755 index 00000000..05cbb42e --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarexpectedprevious/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarexpectedprevious - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarExpectedPrevious

+

Namespace: LLama.Exceptions

+

A specified character was expected to preceded another when parsing

+
public class GrammarExpectedPrevious : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarExpectedPrevious
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarformatexception/index.html b/0.5/xmldocs/llama.exceptions.grammarformatexception/index.html new file mode 100755 index 00000000..d10e6d03 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarformatexception/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarformatexception - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarFormatException

+

Namespace: LLama.Exceptions

+

Base class for all grammar exceptions

+
public abstract class GrammarFormatException : System.Exception, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatException
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html b/0.5/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html new file mode 100755 index 00000000..08943cd7 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunexpectedcharaltelement - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnexpectedCharAltElement

+

Namespace: LLama.Exceptions

+

A CHAR_ALT was created without a preceding CHAR element

+
public class GrammarUnexpectedCharAltElement : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnexpectedCharAltElement
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html b/0.5/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html new file mode 100755 index 00000000..ba20cd25 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunexpectedcharrngelement - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnexpectedCharRngElement

+

Namespace: LLama.Exceptions

+

A CHAR_RNG was created without a preceding CHAR element

+
public class GrammarUnexpectedCharRngElement : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnexpectedCharRngElement
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html b/0.5/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html new file mode 100755 index 00000000..2e17d90f --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunexpectedendelement - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnexpectedEndElement

+

Namespace: LLama.Exceptions

+

An END was encountered before the last element

+
public class GrammarUnexpectedEndElement : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnexpectedEndElement
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html b/0.5/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html new file mode 100755 index 00000000..99de7030 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunexpectedendofinput - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnexpectedEndOfInput

+

Namespace: LLama.Exceptions

+

End-of-file was encountered while parsing

+
public class GrammarUnexpectedEndOfInput : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnexpectedEndOfInput
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html b/0.5/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html new file mode 100755 index 00000000..32c7be22 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunexpectedhexcharscount - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnexpectedHexCharsCount

+

Namespace: LLama.Exceptions

+

An incorrect number of characters were encountered while parsing a hex literal

+
public class GrammarUnexpectedHexCharsCount : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnexpectedHexCharsCount
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html b/0.5/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html new file mode 100755 index 00000000..c032f81c --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html @@ -0,0 +1,2465 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.grammarunknownescapecharacter - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarUnknownEscapeCharacter

+

Namespace: LLama.Exceptions

+

An unexpected character was encountered after an escape sequence

+
public class GrammarUnknownEscapeCharacter : GrammarFormatException, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionGrammarFormatExceptionGrammarUnknownEscapeCharacter
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.exceptions.runtimeerror/index.html b/0.5/xmldocs/llama.exceptions.runtimeerror/index.html new file mode 100755 index 00000000..fa98e4a0 --- /dev/null +++ b/0.5/xmldocs/llama.exceptions.runtimeerror/index.html @@ -0,0 +1,2553 @@ + + + + + + + + + + + + + + + + + + + + + + llama.exceptions.runtimeerror - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

RuntimeError

+

Namespace: LLama.Exceptions

+
public class RuntimeError : System.Exception, System.Runtime.Serialization.ISerializable
+
+

Inheritance ObjectExceptionRuntimeError
+Implements ISerializable

+

Properties

+

TargetSite

+
public MethodBase TargetSite { get; }
+
+

Property Value

+

MethodBase

+

Message

+
public string Message { get; }
+
+

Property Value

+

String

+

Data

+
public IDictionary Data { get; }
+
+

Property Value

+

IDictionary

+

InnerException

+
public Exception InnerException { get; }
+
+

Property Value

+

Exception

+ +
public string HelpLink { get; set; }
+
+

Property Value

+

String

+

Source

+
public string Source { get; set; }
+
+

Property Value

+

String

+

HResult

+
public int HResult { get; set; }
+
+

Property Value

+

Int32

+

StackTrace

+
public string StackTrace { get; }
+
+

Property Value

+

String

+

Constructors

+

RuntimeError()

+
public RuntimeError()
+
+

RuntimeError(String)

+
public RuntimeError(string message)
+
+

Parameters

+

message String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.extensions.imodelparamsextensions/index.html b/0.5/xmldocs/llama.extensions.imodelparamsextensions/index.html new file mode 100755 index 00000000..1d61bb4b --- /dev/null +++ b/0.5/xmldocs/llama.extensions.imodelparamsextensions/index.html @@ -0,0 +1,2184 @@ + + + + + + + + + + + + + + + + + + + + + + llama.extensions.imodelparamsextensions - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

IModelParamsExtensions

+

Namespace: LLama.Extensions

+

Extention methods to the IModelParams interface

+
public static class IModelParamsExtensions
+
+

Inheritance ObjectIModelParamsExtensions

+

Methods

+

ToLlamaContextParams(IModelParams, LLamaContextParams&)

+

Convert the given IModelParams into a LLamaContextParams

+
public static MemoryHandle ToLlamaContextParams(IModelParams params, LLamaContextParams& result)
+
+

Parameters

+

params IModelParams

+

result LLamaContextParams&

+

Returns

+

MemoryHandle

+

Exceptions

+

FileNotFoundException

+

ArgumentException

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.extensions.keyvaluepairextensions/index.html b/0.5/xmldocs/llama.extensions.keyvaluepairextensions/index.html new file mode 100755 index 00000000..ad9f338b --- /dev/null +++ b/0.5/xmldocs/llama.extensions.keyvaluepairextensions/index.html @@ -0,0 +1,2174 @@ + + + + + + + + + + + + + + + + + + + + + + llama.extensions.keyvaluepairextensions - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

KeyValuePairExtensions

+

Namespace: LLama.Extensions

+

Extensions to the KeyValuePair struct

+
public static class KeyValuePairExtensions
+
+

Inheritance ObjectKeyValuePairExtensions

+

Methods

+

Deconstruct<TKey, TValue>(KeyValuePair<TKey, TValue>, TKey&, TValue&)

+

Deconstruct a KeyValuePair into it's constituent parts.

+
public static void Deconstruct<TKey, TValue>(KeyValuePair<TKey, TValue> pair, TKey& first, TValue& second)
+
+

Type Parameters

+

TKey
+Type of the Key

+

TValue
+Type of the Value

+

Parameters

+

pair KeyValuePair<TKey, TValue>
+The KeyValuePair to deconstruct

+

first TKey&
+First element, the Key

+

second TValue&
+Second element, the Value

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.grammars.grammar/index.html b/0.5/xmldocs/llama.grammars.grammar/index.html new file mode 100755 index 00000000..f05901a4 --- /dev/null +++ b/0.5/xmldocs/llama.grammars.grammar/index.html @@ -0,0 +1,2490 @@ + + + + + + + + + + + + + + + + + + + + + + llama.grammars.grammar - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Grammar

+

Namespace: LLama.Grammars

+

A grammar is a set of GrammarRules for deciding which characters are valid next. Can be used to constrain + output to certain formats - e.g. force the model to output JSON

+
public sealed class Grammar
+
+

Inheritance ObjectGrammar

+

Properties

+

StartRuleIndex

+

Index of the initial rule to start from

+
public ulong StartRuleIndex { get; set; }
+
+

Property Value

+

UInt64

+

Rules

+

The rules which make up this grammar

+
public IReadOnlyList<GrammarRule> Rules { get; }
+
+

Property Value

+

IReadOnlyList<GrammarRule>

+

Constructors

+

Grammar(IReadOnlyList<GrammarRule>, UInt64)

+

Create a new grammar from a set of rules

+
public Grammar(IReadOnlyList<GrammarRule> rules, ulong startRuleIndex)
+
+

Parameters

+

rules IReadOnlyList<GrammarRule>
+The rules which make up this grammar

+

startRuleIndex UInt64
+Index of the initial rule to start from

+

Exceptions

+

ArgumentOutOfRangeException

+

Methods

+

CreateInstance()

+

Create a SafeLLamaGrammarHandle instance to use for parsing

+
public SafeLLamaGrammarHandle CreateInstance()
+
+

Returns

+

SafeLLamaGrammarHandle

+

Parse(String, String)

+

Parse a string of GGML BNF into a Grammar

+
public static Grammar Parse(string gbnf, string startRule)
+
+

Parameters

+

gbnf String
+The string to parse

+

startRule String
+Name of the start rule of this grammar

+

Returns

+

Grammar
+A Grammar which can be converted into a SafeLLamaGrammarHandle for sampling

+

Exceptions

+

GrammarFormatException
+Thrown if input is malformed

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.grammars.grammarrule/index.html b/0.5/xmldocs/llama.grammars.grammarrule/index.html new file mode 100755 index 00000000..13ca143a --- /dev/null +++ b/0.5/xmldocs/llama.grammars.grammarrule/index.html @@ -0,0 +1,2571 @@ + + + + + + + + + + + + + + + + + + + + + + llama.grammars.grammarrule - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

GrammarRule

+

Namespace: LLama.Grammars

+

A single rule in a Grammar

+
public sealed class GrammarRule : System.IEquatable`1[[LLama.Grammars.GrammarRule, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectGrammarRule
+Implements IEquatable<GrammarRule>

+

Properties

+

Name

+

Name of this rule

+
public string Name { get; }
+
+

Property Value

+

String

+

Elements

+

The elements of this grammar rule

+
public IReadOnlyList<LLamaGrammarElement> Elements { get; }
+
+

Property Value

+

IReadOnlyList<LLamaGrammarElement>

+

Constructors

+

GrammarRule(String, IReadOnlyList<LLamaGrammarElement>)

+

Create a new GrammarRule containing the given elements

+
public GrammarRule(string name, IReadOnlyList<LLamaGrammarElement> elements)
+
+

Parameters

+

name String

+

elements IReadOnlyList<LLamaGrammarElement>

+

Exceptions

+

ArgumentException

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(GrammarRule)

+
public bool Equals(GrammarRule other)
+
+

Parameters

+

other GrammarRule

+

Returns

+

Boolean

+

<Clone>$()

+
public GrammarRule <Clone>$()
+
+

Returns

+

GrammarRule

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.instructexecutor/index.html b/0.5/xmldocs/llama.instructexecutor/index.html new file mode 100755 index 00000000..b299f184 --- /dev/null +++ b/0.5/xmldocs/llama.instructexecutor/index.html @@ -0,0 +1,2648 @@ + + + + + + + + + + + + + + + + + + + + + + llama.instructexecutor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

InstructExecutor

+

Namespace: LLama

+

The LLama executor for instruct mode.

+
public class InstructExecutor : StatefulExecutorBase, LLama.Abstractions.ILLamaExecutor
+
+

Inheritance ObjectStatefulExecutorBaseInstructExecutor
+Implements ILLamaExecutor

+

Properties

+

Context

+

The context used by the executor.

+
public LLamaContext Context { get; }
+
+

Property Value

+

LLamaContext

+

Constructors

+

InstructExecutor(LLamaContext, String, String)

+
public InstructExecutor(LLamaContext context, string instructionPrefix, string instructionSuffix)
+
+

Parameters

+

context LLamaContext

+

instructionPrefix String

+

instructionSuffix String

+

Methods

+

GetStateData()

+
public ExecutorBaseState GetStateData()
+
+

Returns

+

ExecutorBaseState

+

LoadState(ExecutorBaseState)

+
public void LoadState(ExecutorBaseState data)
+
+

Parameters

+

data ExecutorBaseState

+

SaveState(String)

+
public void SaveState(string filename)
+
+

Parameters

+

filename String

+

LoadState(String)

+
public void LoadState(string filename)
+
+

Parameters

+

filename String

+

GetLoopCondition(InferStateArgs)

+
protected bool GetLoopCondition(InferStateArgs args)
+
+

Parameters

+

args InferStateArgs

+

Returns

+

Boolean

+

PreprocessInputs(String, InferStateArgs)

+
protected void PreprocessInputs(string text, InferStateArgs args)
+
+

Parameters

+

text String

+

args InferStateArgs

+

PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)

+
protected bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+

extraOutputs IEnumerable`1&

+

Returns

+

Boolean

+

InferInternal(IInferenceParams, InferStateArgs)

+
protected void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.interactiveexecutor/index.html b/0.5/xmldocs/llama.interactiveexecutor/index.html new file mode 100755 index 00000000..f6136c6f --- /dev/null +++ b/0.5/xmldocs/llama.interactiveexecutor/index.html @@ -0,0 +1,2648 @@ + + + + + + + + + + + + + + + + + + + + + + llama.interactiveexecutor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

InteractiveExecutor

+

Namespace: LLama

+

The LLama executor for interactive mode.

+
public class InteractiveExecutor : StatefulExecutorBase, LLama.Abstractions.ILLamaExecutor
+
+

Inheritance ObjectStatefulExecutorBaseInteractiveExecutor
+Implements ILLamaExecutor

+

Properties

+

Context

+

The context used by the executor.

+
public LLamaContext Context { get; }
+
+

Property Value

+

LLamaContext

+

Constructors

+

InteractiveExecutor(LLamaContext)

+
public InteractiveExecutor(LLamaContext context)
+
+

Parameters

+

context LLamaContext

+

Methods

+

GetStateData()

+
public ExecutorBaseState GetStateData()
+
+

Returns

+

ExecutorBaseState

+

LoadState(ExecutorBaseState)

+
public void LoadState(ExecutorBaseState data)
+
+

Parameters

+

data ExecutorBaseState

+

SaveState(String)

+
public void SaveState(string filename)
+
+

Parameters

+

filename String

+

LoadState(String)

+
public void LoadState(string filename)
+
+

Parameters

+

filename String

+

GetLoopCondition(InferStateArgs)

+

Define whether to continue the loop to generate responses.

+
protected bool GetLoopCondition(InferStateArgs args)
+
+

Parameters

+

args InferStateArgs

+

Returns

+

Boolean

+

PreprocessInputs(String, InferStateArgs)

+
protected void PreprocessInputs(string text, InferStateArgs args)
+
+

Parameters

+

text String

+

args InferStateArgs

+

PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)

+

Return whether to break the generation.

+
protected bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+

extraOutputs IEnumerable`1&

+

Returns

+

Boolean

+

InferInternal(IInferenceParams, InferStateArgs)

+
protected void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.llamacontext/index.html b/0.5/xmldocs/llama.llamacontext/index.html new file mode 100755 index 00000000..d5b41f74 --- /dev/null +++ b/0.5/xmldocs/llama.llamacontext/index.html @@ -0,0 +1,3721 @@ + + + + + + + + + + + + + + + + + + + + + + llama.llamacontext - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaContext

+

Namespace: LLama

+

A llama_context, which holds all the context required to interact with a model

+
public sealed class LLamaContext : System.IDisposable
+
+

Inheritance ObjectLLamaContext
+Implements IDisposable

+

Properties

+

VocabCount

+

Total number of tokens in vocabulary of this model

+
public int VocabCount { get; }
+
+

Property Value

+

Int32

+

ContextSize

+

Total number of tokens in the context

+
public int ContextSize { get; }
+
+

Property Value

+

Int32

+

EmbeddingSize

+

Dimension of embedding vectors

+
public int EmbeddingSize { get; }
+
+

Property Value

+

Int32

+

Params

+

The model params set for this model.

+
public IModelParams Params { get; set; }
+
+

Property Value

+

IModelParams

+

NativeHandle

+

The native handle, which is used to be passed to the native APIs

+
public SafeLLamaContextHandle NativeHandle { get; }
+
+

Property Value

+

SafeLLamaContextHandle

+

Remarks:

+

Be careful how you use this!

+

Encoding

+

The encoding set for this model to deal with text input.

+
public Encoding Encoding { get; }
+
+

Property Value

+

Encoding

+

EmbeddingLength

+

The embedding length of the model, also known as n_embed

+
public int EmbeddingLength { get; }
+
+

Property Value

+

Int32

+

Constructors

+

LLamaContext(IModelParams, ILLamaLogger)

+

Caution

+

Use the LLamaWeights.CreateContext instead

+
+
public LLamaContext(IModelParams params, ILLamaLogger logger)
+
+

Parameters

+

params IModelParams
+Model params.

+

logger ILLamaLogger
+The logger.

+

LLamaContext(LLamaWeights, IModelParams, ILLamaLogger)

+

Create a new LLamaContext for the given LLamaWeights

+
public LLamaContext(LLamaWeights model, IModelParams params, ILLamaLogger logger)
+
+

Parameters

+

model LLamaWeights

+

params IModelParams

+

logger ILLamaLogger

+

Exceptions

+

ObjectDisposedException

+

Methods

+

Clone()

+

Create a copy of the current state of this context

+
public LLamaContext Clone()
+
+

Returns

+

LLamaContext

+

Tokenize(String, Boolean)

+

Tokenize a string.

+
public Int32[] Tokenize(string text, bool addBos)
+
+

Parameters

+

text String

+

addBos Boolean
+Whether to add a bos to the text.

+

Returns

+

Int32[]

+

DeTokenize(IEnumerable<Int32>)

+

Detokenize the tokens to text.

+
public string DeTokenize(IEnumerable<int> tokens)
+
+

Parameters

+

tokens IEnumerable<Int32>

+

Returns

+

String

+

SaveState(String)

+

Save the state to specified path.

+
public void SaveState(string filename)
+
+

Parameters

+

filename String

+

GetStateData()

+

Caution

+

Use GetState instead, this supports larger states (over 2GB)

+
+

Get the state data as a byte array.

+
public Byte[] GetStateData()
+
+

Returns

+

Byte[]

+

GetState()

+

Get the state data as an opaque handle

+
public State GetState()
+
+

Returns

+

State

+

LoadState(String)

+

Load the state from specified path.

+
public void LoadState(string filename)
+
+

Parameters

+

filename String

+

Exceptions

+

RuntimeError

+

LoadState(Byte[])

+

Load the state from memory.

+
public void LoadState(Byte[] stateData)
+
+

Parameters

+

stateData Byte[]

+

Exceptions

+

RuntimeError

+

LoadState(State)

+

Load the state from memory.

+
public void LoadState(State state)
+
+

Parameters

+

state State

+

Exceptions

+

RuntimeError

+

Sample(LLamaTokenDataArray, Nullable`1&, Single, MirostatType, Single, Single, Int32, Single, Single, Single, SafeLLamaGrammarHandle)

+

Perform the sampling. Please don't use it unless you fully know what it does.

+
public int Sample(LLamaTokenDataArray candidates, Nullable`1& mirostat_mu, float temperature, MirostatType mirostat, float mirostatTau, float mirostatEta, int topK, float topP, float tfsZ, float typicalP, SafeLLamaGrammarHandle grammar)
+
+

Parameters

+

candidates LLamaTokenDataArray

+

mirostat_mu Nullable`1&

+

temperature Single

+

mirostat MirostatType

+

mirostatTau Single

+

mirostatEta Single

+

topK Int32

+

topP Single

+

tfsZ Single

+

typicalP Single

+

grammar SafeLLamaGrammarHandle

+

Returns

+

Int32

+

ApplyPenalty(IEnumerable<Int32>, Dictionary<Int32, Single>, Int32, Single, Single, Single, Boolean)

+

Apply the penalty for the tokens. Please don't use it unless you fully know what it does.

+
public LLamaTokenDataArray ApplyPenalty(IEnumerable<int> lastTokens, Dictionary<int, float> logitBias, int repeatLastTokensCount, float repeatPenalty, float alphaFrequency, float alphaPresence, bool penalizeNL)
+
+

Parameters

+

lastTokens IEnumerable<Int32>

+

logitBias Dictionary<Int32, Single>

+

repeatLastTokensCount Int32

+

repeatPenalty Single

+

alphaFrequency Single

+

alphaPresence Single

+

penalizeNL Boolean

+

Returns

+

LLamaTokenDataArray

+

Eval(Int32[], Int32)

+
public int Eval(Int32[] tokens, int pastTokensCount)
+
+

Parameters

+

tokens Int32[]

+

pastTokensCount Int32

+

Returns

+

Int32
+The updated pastTokensCount.

+

Exceptions

+

RuntimeError

+

Eval(List<Int32>, Int32)

+
public int Eval(List<int> tokens, int pastTokensCount)
+
+

Parameters

+

tokens List<Int32>

+

pastTokensCount Int32

+

Returns

+

Int32
+The updated pastTokensCount.

+

Exceptions

+

RuntimeError

+

Eval(ReadOnlyMemory<Int32>, Int32)

+
public int Eval(ReadOnlyMemory<int> tokens, int pastTokensCount)
+
+

Parameters

+

tokens ReadOnlyMemory<Int32>

+

pastTokensCount Int32

+

Returns

+

Int32
+The updated pastTokensCount.

+

Exceptions

+

RuntimeError

+

Eval(ReadOnlySpan<Int32>, Int32)

+
public int Eval(ReadOnlySpan<int> tokens, int pastTokensCount)
+
+

Parameters

+

tokens ReadOnlySpan<Int32>

+

pastTokensCount Int32

+

Returns

+

Int32
+The updated pastTokensCount.

+

Exceptions

+

RuntimeError

+

GenerateResult(IEnumerable<Int32>)

+
internal IEnumerable<string> GenerateResult(IEnumerable<int> ids)
+
+

Parameters

+

ids IEnumerable<Int32>

+

Returns

+

IEnumerable<String>

+

TokenToString(Int32)

+

Convert a token into a string

+
public string TokenToString(int token)
+
+

Parameters

+

token Int32

+

Returns

+

String

+

Dispose()

+
public void Dispose()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.llamaembedder/index.html b/0.5/xmldocs/llama.llamaembedder/index.html new file mode 100755 index 00000000..ad4b4360 --- /dev/null +++ b/0.5/xmldocs/llama.llamaembedder/index.html @@ -0,0 +1,2572 @@ + + + + + + + + + + + + + + + + + + + + + + llama.llamaembedder - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaEmbedder

+

Namespace: LLama

+

The embedder for LLama, which supports getting embeddings from text.

+
public sealed class LLamaEmbedder : System.IDisposable
+
+

Inheritance ObjectLLamaEmbedder
+Implements IDisposable

+

Properties

+

EmbeddingSize

+

Dimension of embedding vectors

+
public int EmbeddingSize { get; }
+
+

Property Value

+

Int32

+

Constructors

+

LLamaEmbedder(IModelParams)

+
public LLamaEmbedder(IModelParams params)
+
+

Parameters

+

params IModelParams

+

LLamaEmbedder(LLamaWeights, IModelParams)

+
public LLamaEmbedder(LLamaWeights weights, IModelParams params)
+
+

Parameters

+

weights LLamaWeights

+

params IModelParams

+

Methods

+

GetEmbeddings(String, Int32, Boolean, String)

+

Caution

+

'threads' and 'encoding' parameters are no longer used

+
+

Get the embeddings of the text.

+
public Single[] GetEmbeddings(string text, int threads, bool addBos, string encoding)
+
+

Parameters

+

text String

+

threads Int32
+unused

+

addBos Boolean
+Add bos to the text.

+

encoding String
+unused

+

Returns

+

Single[]

+

Exceptions

+

RuntimeError

+

GetEmbeddings(String)

+

Get the embeddings of the text.

+
public Single[] GetEmbeddings(string text)
+
+

Parameters

+

text String

+

Returns

+

Single[]

+

Exceptions

+

RuntimeError

+

GetEmbeddings(String, Boolean)

+

Get the embeddings of the text.

+
public Single[] GetEmbeddings(string text, bool addBos)
+
+

Parameters

+

text String

+

addBos Boolean
+Add bos to the text.

+

Returns

+

Single[]

+

Exceptions

+

RuntimeError

+

Dispose()

+
public void Dispose()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.llamaquantizer/index.html b/0.5/xmldocs/llama.llamaquantizer/index.html new file mode 100755 index 00000000..1403157c --- /dev/null +++ b/0.5/xmldocs/llama.llamaquantizer/index.html @@ -0,0 +1,2280 @@ + + + + + + + + + + + + + + + + + + + + + + llama.llamaquantizer - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaQuantizer

+

Namespace: LLama

+

The quantizer to quantize the model.

+
public static class LLamaQuantizer
+
+

Inheritance ObjectLLamaQuantizer

+

Methods

+

Quantize(String, String, LLamaFtype, Int32, Boolean, Boolean)

+

Quantize the model.

+
public static bool Quantize(string srcFileName, string dstFilename, LLamaFtype ftype, int nthread, bool allowRequantize, bool quantizeOutputTensor)
+
+

Parameters

+

srcFileName String
+The model file to be quantized.

+

dstFilename String
+The path to save the quantized model.

+

ftype LLamaFtype
+The type of quantization.

+

nthread Int32
+Thread to be used during the quantization. By default it's the physical core number.

+

allowRequantize Boolean

+

quantizeOutputTensor Boolean

+

Returns

+

Boolean
+Whether the quantization is successful.

+

Exceptions

+

ArgumentException

+

Quantize(String, String, String, Int32, Boolean, Boolean)

+

Quantize the model.

+
public static bool Quantize(string srcFileName, string dstFilename, string ftype, int nthread, bool allowRequantize, bool quantizeOutputTensor)
+
+

Parameters

+

srcFileName String
+The model file to be quantized.

+

dstFilename String
+The path to save the quantized model.

+

ftype String
+The type of quantization.

+

nthread Int32
+Thread to be used during the quantization. By default it's the physical core number.

+

allowRequantize Boolean

+

quantizeOutputTensor Boolean

+

Returns

+

Boolean
+Whether the quantization is successful.

+

Exceptions

+

ArgumentException

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.llamatransforms/index.html b/0.5/xmldocs/llama.llamatransforms/index.html new file mode 100755 index 00000000..68150578 --- /dev/null +++ b/0.5/xmldocs/llama.llamatransforms/index.html @@ -0,0 +1,2121 @@ + + + + + + + + + + + + + + + + + + + + + + llama.llamatransforms - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaTransforms

+

Namespace: LLama

+

A class that contains all the transforms provided internally by LLama.

+
public class LLamaTransforms
+
+

Inheritance ObjectLLamaTransforms

+

Constructors

+

LLamaTransforms()

+
public LLamaTransforms()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.llamaweights/index.html b/0.5/xmldocs/llama.llamaweights/index.html new file mode 100755 index 00000000..99d6c540 --- /dev/null +++ b/0.5/xmldocs/llama.llamaweights/index.html @@ -0,0 +1,2505 @@ + + + + + + + + + + + + + + + + + + + + + + llama.llamaweights - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaWeights

+

Namespace: LLama

+

A set of model weights, loaded into memory.

+
public sealed class LLamaWeights : System.IDisposable
+
+

Inheritance ObjectLLamaWeights
+Implements IDisposable

+

Properties

+

NativeHandle

+

The native handle, which is used in the native APIs

+
public SafeLlamaModelHandle NativeHandle { get; }
+
+

Property Value

+

SafeLlamaModelHandle

+

Remarks:

+

Be careful how you use this!

+

Encoding

+

Encoding to use to convert text into bytes for the model

+
public Encoding Encoding { get; }
+
+

Property Value

+

Encoding

+

VocabCount

+

Total number of tokens in vocabulary of this model

+
public int VocabCount { get; }
+
+

Property Value

+

Int32

+

ContextSize

+

Total number of tokens in the context

+
public int ContextSize { get; }
+
+

Property Value

+

Int32

+

EmbeddingSize

+

Dimension of embedding vectors

+
public int EmbeddingSize { get; }
+
+

Property Value

+

Int32

+

Methods

+

LoadFromFile(IModelParams)

+

Load weights into memory

+
public static LLamaWeights LoadFromFile(IModelParams params)
+
+

Parameters

+

params IModelParams

+

Returns

+

LLamaWeights

+

Dispose()

+
public void Dispose()
+
+

CreateContext(IModelParams)

+

Create a llama_context using this model

+
public LLamaContext CreateContext(IModelParams params)
+
+

Parameters

+

params IModelParams

+

Returns

+

LLamaContext

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamacontextparams/index.html b/0.5/xmldocs/llama.native.llamacontextparams/index.html new file mode 100755 index 00000000..0d9015d9 --- /dev/null +++ b/0.5/xmldocs/llama.native.llamacontextparams/index.html @@ -0,0 +1,2681 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamacontextparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaContextParams

+

Namespace: LLama.Native

+

A C# representation of the llama.cpp llama_context_params struct

+
public struct LLamaContextParams
+
+

Inheritance ObjectValueTypeLLamaContextParams

+

Fields

+

seed

+

RNG seed, -1 for random

+
public int seed;
+
+

n_ctx

+

text context

+
public int n_ctx;
+
+

n_batch

+

prompt processing batch size

+
public int n_batch;
+
+

n_gpu_layers

+

number of layers to store in VRAM

+
public int n_gpu_layers;
+
+

main_gpu

+

the GPU that is used for scratch and small tensors

+
public int main_gpu;
+
+

tensor_split

+

how to split layers across multiple GPUs

+
public IntPtr tensor_split;
+
+

rope_freq_base

+

ref: https://github.com/ggerganov/llama.cpp/pull/2054 + RoPE base frequency

+
public float rope_freq_base;
+
+

rope_freq_scale

+

ref: https://github.com/ggerganov/llama.cpp/pull/2054 + RoPE frequency scaling factor

+
public float rope_freq_scale;
+
+

progress_callback

+

called with a progress value between 0 and 1, pass NULL to disable

+
public IntPtr progress_callback;
+
+

progress_callback_user_data

+

context pointer passed to the progress callback

+
public IntPtr progress_callback_user_data;
+
+

Properties

+

low_vram

+

if true, reduce VRAM usage at the cost of performance

+
public bool low_vram { get; set; }
+
+

Property Value

+

Boolean

+

mul_mat_q

+

if true, use experimental mul_mat_q kernels

+
public bool mul_mat_q { get; set; }
+
+

Property Value

+

Boolean

+

f16_kv

+

use fp16 for KV cache

+
public bool f16_kv { get; set; }
+
+

Property Value

+

Boolean

+

logits_all

+

the llama_eval() call computes all logits, not just the last one

+
public bool logits_all { get; set; }
+
+

Property Value

+

Boolean

+

vocab_only

+

only load the vocabulary, no weights

+
public bool vocab_only { get; set; }
+
+

Property Value

+

Boolean

+

use_mmap

+

use mmap if possible

+
public bool use_mmap { get; set; }
+
+

Property Value

+

Boolean

+

use_mlock

+

force system to keep model in RAM

+
public bool use_mlock { get; set; }
+
+

Property Value

+

Boolean

+

embedding

+

embedding mode only

+
public bool embedding { get; set; }
+
+

Property Value

+

Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamaftype/index.html b/0.5/xmldocs/llama.native.llamaftype/index.html new file mode 100755 index 00000000..07dd7fcf --- /dev/null +++ b/0.5/xmldocs/llama.native.llamaftype/index.html @@ -0,0 +1,2194 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamaftype - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaFtype

+

Namespace: LLama.Native

+

Supported model file types

+
public enum LLamaFtype
+
+

Inheritance ObjectValueTypeEnumLLamaFtype
+Implements IComparable, IFormattable, IConvertible

+

Fields

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameValueDescription
LLAMA_FTYPE_ALL_F320All f32
LLAMA_FTYPE_MOSTLY_F161Mostly f16
LLAMA_FTYPE_MOSTLY_Q8_07Mostly 8 bit
LLAMA_FTYPE_MOSTLY_Q4_02Mostly 4 bit
LLAMA_FTYPE_MOSTLY_Q4_13Mostly 4 bit
LLAMA_FTYPE_MOSTLY_Q4_1_SOME_F164Mostly 4 bit, tok_embeddings.weight and output.weight are f16
LLAMA_FTYPE_MOSTLY_Q5_08Mostly 5 bit
LLAMA_FTYPE_MOSTLY_Q5_19Mostly 5 bit
LLAMA_FTYPE_MOSTLY_Q2_K10K-Quant 2 bit
LLAMA_FTYPE_MOSTLY_Q3_K_S11K-Quant 3 bit (Small)
LLAMA_FTYPE_MOSTLY_Q3_K_M12K-Quant 3 bit (Medium)
LLAMA_FTYPE_MOSTLY_Q3_K_L13K-Quant 3 bit (Large)
LLAMA_FTYPE_MOSTLY_Q4_K_S14K-Quant 4 bit (Small)
LLAMA_FTYPE_MOSTLY_Q4_K_M15K-Quant 4 bit (Medium)
LLAMA_FTYPE_MOSTLY_Q5_K_S16K-Quant 5 bit (Small)
LLAMA_FTYPE_MOSTLY_Q5_K_M17K-Quant 5 bit (Medium)
LLAMA_FTYPE_MOSTLY_Q6_K18K-Quant 6 bit
LLAMA_FTYPE_GUESSED1024File type was not specified
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamagrammarelement/index.html b/0.5/xmldocs/llama.native.llamagrammarelement/index.html new file mode 100755 index 00000000..96e35633 --- /dev/null +++ b/0.5/xmldocs/llama.native.llamagrammarelement/index.html @@ -0,0 +1,2454 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamagrammarelement - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaGrammarElement

+

Namespace: LLama.Native

+

An element of a grammar

+
public struct LLamaGrammarElement
+
+

Inheritance ObjectValueTypeLLamaGrammarElement
+Implements IEquatable<LLamaGrammarElement>

+

Fields

+

Type

+

The type of this element

+
public LLamaGrammarElementType Type;
+
+

Value

+

Unicode code point or rule ID

+
public uint Value;
+
+

Constructors

+

LLamaGrammarElement(LLamaGrammarElementType, UInt32)

+

Construct a new LLamaGrammarElement

+
LLamaGrammarElement(LLamaGrammarElementType type, uint value)
+
+

Parameters

+

type LLamaGrammarElementType

+

value UInt32

+

Methods

+

Equals(LLamaGrammarElement)

+
bool Equals(LLamaGrammarElement other)
+
+

Parameters

+

other LLamaGrammarElement

+

Returns

+

Boolean

+

Equals(Object)

+
bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

GetHashCode()

+
int GetHashCode()
+
+

Returns

+

Int32

+

IsCharElement()

+
bool IsCharElement()
+
+

Returns

+

Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamagrammarelementtype/index.html b/0.5/xmldocs/llama.native.llamagrammarelementtype/index.html new file mode 100755 index 00000000..dfa8e27f --- /dev/null +++ b/0.5/xmldocs/llama.native.llamagrammarelementtype/index.html @@ -0,0 +1,2139 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamagrammarelementtype - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaGrammarElementType

+

Namespace: LLama.Native

+

grammar element type

+
public enum LLamaGrammarElementType
+
+

Inheritance ObjectValueTypeEnumLLamaGrammarElementType
+Implements IComparable, IFormattable, IConvertible

+

Fields

+ + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
NameValueDescription
END0end of rule definition
ALT1start of alternate definition for rule
RULE_REF2non-terminal element: reference to rule
CHAR3terminal element: character (code point)
CHAR_NOT4inverse char(s) ([^a], [^a-b] [^abc])
CHAR_RNG_UPPER5modifies a preceding CHAR or CHAR_ALT to be an inclusive range ([a-z])
CHAR_ALT6modifies a preceding CHAR or CHAR_RNG_UPPER to add an alternate char to match ([ab], [a-zA])
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamamodelquantizeparams/index.html b/0.5/xmldocs/llama.native.llamamodelquantizeparams/index.html new file mode 100755 index 00000000..a4a7b35a --- /dev/null +++ b/0.5/xmldocs/llama.native.llamamodelquantizeparams/index.html @@ -0,0 +1,2259 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamamodelquantizeparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaModelQuantizeParams

+

Namespace: LLama.Native

+

Quantizer parameters used in the native API

+
public struct LLamaModelQuantizeParams
+
+

Inheritance ObjectValueTypeLLamaModelQuantizeParams

+

Fields

+

nthread

+

number of threads to use for quantizing, if <=0 will use std::thread::hardware_concurrency()

+
public int nthread;
+
+

ftype

+

quantize to this llama_ftype

+
public LLamaFtype ftype;
+
+

Properties

+

allow_requantize

+

allow quantizing non-f32/f16 tensors

+
public bool allow_requantize { get; set; }
+
+

Property Value

+

Boolean

+

quantize_output_tensor

+

quantize output.weight

+
public bool quantize_output_tensor { get; set; }
+
+

Property Value

+

Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamatokendata/index.html b/0.5/xmldocs/llama.native.llamatokendata/index.html new file mode 100755 index 00000000..b78726a3 --- /dev/null +++ b/0.5/xmldocs/llama.native.llamatokendata/index.html @@ -0,0 +1,2231 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamatokendata - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaTokenData

+

Namespace: LLama.Native

+
public struct LLamaTokenData
+
+

Inheritance ObjectValueTypeLLamaTokenData

+

Fields

+

id

+

token id

+
public int id;
+
+

logit

+

log-odds of the token

+
public float logit;
+
+

p

+

probability of the token

+
public float p;
+
+

Constructors

+

LLamaTokenData(Int32, Single, Single)

+
LLamaTokenData(int id, float logit, float p)
+
+

Parameters

+

id Int32

+

logit Single

+

p Single

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamatokendataarray/index.html b/0.5/xmldocs/llama.native.llamatokendataarray/index.html new file mode 100755 index 00000000..8782b497 --- /dev/null +++ b/0.5/xmldocs/llama.native.llamatokendataarray/index.html @@ -0,0 +1,2303 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamatokendataarray - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaTokenDataArray

+

Namespace: LLama.Native

+

Contains an array of LLamaTokenData, potentially sorted.

+
public struct LLamaTokenDataArray
+
+

Inheritance ObjectValueTypeLLamaTokenDataArray

+

Fields

+

data

+

The LLamaTokenData

+
public Memory<LLamaTokenData> data;
+
+

sorted

+

Indicates if data is sorted by logits in descending order. If this is false the token data is in no particular order.

+
public bool sorted;
+
+

Constructors

+

LLamaTokenDataArray(Memory<LLamaTokenData>, Boolean)

+

Create a new LLamaTokenDataArray

+
LLamaTokenDataArray(Memory<LLamaTokenData> tokens, bool isSorted)
+
+

Parameters

+

tokens Memory<LLamaTokenData>

+

isSorted Boolean

+

Methods

+

Create(ReadOnlySpan<Single>)

+

Create a new LLamaTokenDataArray, copying the data from the given logits

+
LLamaTokenDataArray Create(ReadOnlySpan<float> logits)
+
+

Parameters

+

logits ReadOnlySpan<Single>

+

Returns

+

LLamaTokenDataArray

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.llamatokendataarraynative/index.html b/0.5/xmldocs/llama.native.llamatokendataarraynative/index.html new file mode 100755 index 00000000..813013c7 --- /dev/null +++ b/0.5/xmldocs/llama.native.llamatokendataarraynative/index.html @@ -0,0 +1,2308 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.llamatokendataarraynative - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaTokenDataArrayNative

+

Namespace: LLama.Native

+

Contains a pointer to an array of LLamaTokenData which is pinned in memory.

+
public struct LLamaTokenDataArrayNative
+
+

Inheritance ObjectValueTypeLLamaTokenDataArrayNative

+

Fields

+

data

+

A pointer to an array of LlamaTokenData

+
public IntPtr data;
+
+

Remarks:

+

Memory must be pinned in place for all the time this LLamaTokenDataArrayNative is in use

+

size

+

Number of LLamaTokenData in the array

+
public ulong size;
+
+

Properties

+

sorted

+

Indicates if the items in the array are sorted

+
public bool sorted { get; set; }
+
+

Property Value

+

Boolean

+

Methods

+

Create(LLamaTokenDataArray, LLamaTokenDataArrayNative&)

+

Create a new LLamaTokenDataArrayNative around the data in the LLamaTokenDataArray

+
MemoryHandle Create(LLamaTokenDataArray array, LLamaTokenDataArrayNative& native)
+
+

Parameters

+

array LLamaTokenDataArray
+Data source

+

native LLamaTokenDataArrayNative&
+Created native array

+

Returns

+

MemoryHandle
+A memory handle, pinning the data in place until disposed

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.nativeapi/index.html b/0.5/xmldocs/llama.native.nativeapi/index.html new file mode 100755 index 00000000..a11eb299 --- /dev/null +++ b/0.5/xmldocs/llama.native.nativeapi/index.html @@ -0,0 +1,5736 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.nativeapi - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

NativeApi

+

Namespace: LLama.Native

+

Direct translation of the llama.cpp API

+
public class NativeApi
+
+

Inheritance ObjectNativeApi

+

Constructors

+

NativeApi()

+
public NativeApi()
+
+

Methods

+

llama_sample_token_mirostat(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, Single, Int32, Single&)

+

Mirostat 1.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

+
public static int llama_sample_token_mirostat(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float tau, float eta, int m, Single& mu)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+A vector of llama_token_data containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

+

tau Single
+The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

+

eta Single
+The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

+

m Int32
+The number of tokens considered in the estimation of s_hat. This is an arbitrary value that is used to calculate s_hat, which in turn helps to calculate the value of k. In the paper, they use m = 100, but you can experiment with different values to see how it affects the performance of the algorithm.

+

mu Single&
+Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

+

Returns

+

Int32

+

llama_sample_token_mirostat_v2(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, Single, Single&)

+

Mirostat 2.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

+
public static int llama_sample_token_mirostat_v2(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float tau, float eta, Single& mu)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+A vector of llama_token_data containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

+

tau Single
+The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

+

eta Single
+The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

+

mu Single&
+Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

+

Returns

+

Int32

+

llama_sample_token_greedy(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)

+

Selects the token with the highest probability.

+
public static int llama_sample_token_greedy(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

Returns

+

Int32

+

llama_sample_token(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)

+

Randomly selects a token from the candidates based on their probabilities.

+
public static int llama_sample_token(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

Returns

+

Int32

+

llama_token_to_str(SafeLLamaContextHandle, Int32)

+

Token Id -> String. Uses the vocabulary in the provided context

+
public static IntPtr llama_token_to_str(SafeLLamaContextHandle ctx, int token)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

token Int32

+

Returns

+

IntPtr
+Pointer to a string.

+

llama_token_bos(SafeLLamaContextHandle)

+

Get the "Beginning of sentence" token

+
public static int llama_token_bos(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_token_eos(SafeLLamaContextHandle)

+

Get the "End of sentence" token

+
public static int llama_token_eos(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_token_nl(SafeLLamaContextHandle)

+

Get the "new line" token

+
public static int llama_token_nl(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_print_timings(SafeLLamaContextHandle)

+

Print out timing information for this context

+
public static void llama_print_timings(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

llama_reset_timings(SafeLLamaContextHandle)

+

Reset all collected timing information for this context

+
public static void llama_reset_timings(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

llama_print_system_info()

+

Print system information

+
public static IntPtr llama_print_system_info()
+
+

Returns

+

IntPtr

+

llama_model_n_vocab(SafeLlamaModelHandle)

+

Get the number of tokens in the model vocabulary

+
public static int llama_model_n_vocab(SafeLlamaModelHandle model)
+
+

Parameters

+

model SafeLlamaModelHandle

+

Returns

+

Int32

+

llama_model_n_ctx(SafeLlamaModelHandle)

+

Get the size of the context window for the model

+
public static int llama_model_n_ctx(SafeLlamaModelHandle model)
+
+

Parameters

+

model SafeLlamaModelHandle

+

Returns

+

Int32

+

llama_model_n_embd(SafeLlamaModelHandle)

+

Get the dimension of embedding vectors from this model

+
public static int llama_model_n_embd(SafeLlamaModelHandle model)
+
+

Parameters

+

model SafeLlamaModelHandle

+

Returns

+

Int32

+

llama_token_to_piece_with_model(SafeLlamaModelHandle, Int32, Byte*, Int32)

+

Convert a single token into text

+
public static int llama_token_to_piece_with_model(SafeLlamaModelHandle model, int llamaToken, Byte* buffer, int length)
+
+

Parameters

+

model SafeLlamaModelHandle

+

llamaToken Int32

+

buffer Byte*
+buffer to write string into

+

length Int32
+size of the buffer

+

Returns

+

Int32
+The length writte, or if the buffer is too small a negative that indicates the length required

+

llama_tokenize_with_model(SafeLlamaModelHandle, Byte, Int32, Int32, Boolean)

+

Convert text into tokens

+
public static int llama_tokenize_with_model(SafeLlamaModelHandle model, Byte* text, Int32* tokens, int n_max_tokens, bool add_bos)
+
+

Parameters

+

model SafeLlamaModelHandle

+

text Byte*

+

tokens Int32*

+

n_max_tokens Int32

+

add_bos Boolean

+

Returns

+

Int32
+Returns the number of tokens on success, no more than n_max_tokens. + Returns a negative number on failure - the number of tokens that would have been returned

+

llama_log_set(LLamaLogCallback)

+

Register a callback to receive llama log messages

+
public static void llama_log_set(LLamaLogCallback logCallback)
+
+

Parameters

+

logCallback LLamaLogCallback

+

llama_grammar_init(LLamaGrammarElement, UInt64, UInt64)**

+

Create a new grammar from the given set of grammar rules

+
public static IntPtr llama_grammar_init(LLamaGrammarElement** rules, ulong n_rules, ulong start_rule_index)
+
+

Parameters

+

rules LLamaGrammarElement**

+

n_rules UInt64

+

start_rule_index UInt64

+

Returns

+

IntPtr

+

llama_grammar_free(IntPtr)

+

Free all memory from the given SafeLLamaGrammarHandle

+
public static void llama_grammar_free(IntPtr grammar)
+
+

Parameters

+

grammar IntPtr

+

llama_sample_grammar(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, SafeLLamaGrammarHandle)

+

Apply constraints from grammar

+
public static void llama_sample_grammar(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, SafeLLamaGrammarHandle grammar)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&

+

grammar SafeLLamaGrammarHandle

+

llama_grammar_accept_token(SafeLLamaContextHandle, SafeLLamaGrammarHandle, Int32)

+

Accepts the sampled token into the grammar

+
public static void llama_grammar_accept_token(SafeLLamaContextHandle ctx, SafeLLamaGrammarHandle grammar, int token)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

grammar SafeLLamaGrammarHandle

+

token Int32

+

llama_model_quantize(String, String, LLamaModelQuantizeParams*)

+

Returns 0 on success

+
public static int llama_model_quantize(string fname_inp, string fname_out, LLamaModelQuantizeParams* param)
+
+

Parameters

+

fname_inp String

+

fname_out String

+

param LLamaModelQuantizeParams*

+

Returns

+

Int32
+Returns 0 on success

+

Remarks:

+

not great API - very likely to change

+

llama_sample_classifier_free_guidance(SafeLLamaContextHandle, LLamaTokenDataArrayNative, SafeLLamaContextHandle, Single)

+

Apply classifier-free guidance to the logits as described in academic paper "Stay on topic with Classifier-Free Guidance" https://arxiv.org/abs/2306.17806

+
public static void llama_sample_classifier_free_guidance(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative candidates, SafeLLamaContextHandle guidanceCtx, float scale)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative
+A vector of llama_token_data containing the candidate tokens, the logits must be directly extracted from the original generation context without being sorted.

+

guidanceCtx SafeLLamaContextHandle
+A separate context from the same model. Other than a negative prompt at the beginning, it should have all generated and user input tokens copied from the main context.

+

scale Single
+Guidance strength. 1.0f means no guidance. Higher values mean stronger guidance.

+

llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32*, UInt64, Single)

+

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

+
public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, Int32* last_tokens, ulong last_tokens_size, float penalty)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

last_tokens Int32*

+

last_tokens_size UInt64

+

penalty Single

+

llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32*, UInt64, Single, Single)

+

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

+
public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, Int32* last_tokens, ulong last_tokens_size, float alpha_frequency, float alpha_presence)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

last_tokens Int32*

+

last_tokens_size UInt64

+

alpha_frequency Single

+

alpha_presence Single

+

llama_sample_classifier_free_guidance(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, SafeLLamaContextHandle, Single)

+

Apply classifier-free guidance to the logits as described in academic paper "Stay on topic with Classifier-Free Guidance" https://arxiv.org/abs/2306.17806

+
public static void llama_sample_classifier_free_guidance(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, SafeLLamaContextHandle guidance_ctx, float scale)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+A vector of llama_token_data containing the candidate tokens, the logits must be directly extracted from the original generation context without being sorted.

+

guidance_ctx SafeLLamaContextHandle
+A separate context from the same model. Other than a negative prompt at the beginning, it should have all generated and user input tokens copied from the main context.

+

scale Single
+Guidance strength. 1.0f means no guidance. Higher values mean stronger guidance.

+

llama_sample_softmax(SafeLLamaContextHandle, LLamaTokenDataArrayNative&)

+

Sorts candidate tokens by their logits in descending order and calculate probabilities based on logits.

+
public static void llama_sample_softmax(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

llama_sample_top_k(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Int32, UInt64)

+

Top-K sampling described in academic paper "The Curious Case of Neural Text Degeneration" https://arxiv.org/abs/1904.09751

+
public static void llama_sample_top_k(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, int k, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

k Int32

+

min_keep UInt64

+

llama_sample_top_p(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)

+

Nucleus sampling described in academic paper "The Curious Case of Neural Text Degeneration" https://arxiv.org/abs/1904.09751

+
public static void llama_sample_top_p(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float p, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

p Single

+

min_keep UInt64

+

llama_sample_tail_free(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)

+

Tail Free Sampling described in https://www.trentonbricken.com/Tail-Free-Sampling/.

+
public static void llama_sample_tail_free(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float z, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

z Single

+

min_keep UInt64

+

llama_sample_typical(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single, UInt64)

+

Locally Typical Sampling implementation described in the paper https://arxiv.org/abs/2202.00666.

+
public static void llama_sample_typical(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float p, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&
+Pointer to LLamaTokenDataArray

+

p Single

+

min_keep UInt64

+

llama_sample_temperature(SafeLLamaContextHandle, LLamaTokenDataArrayNative&, Single)

+

Modify logits by temperature

+
public static void llama_sample_temperature(SafeLLamaContextHandle ctx, LLamaTokenDataArrayNative& candidates, float temp)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArrayNative&

+

temp Single

+

llama_empty_call()

+

A method that does nothing. This is a native method, calling it will force the llama native dependencies to be loaded.

+
public static bool llama_empty_call()
+
+

Returns

+

Boolean

+

llama_context_default_params()

+

Create a LLamaContextParams with default values

+
public static LLamaContextParams llama_context_default_params()
+
+

Returns

+

LLamaContextParams

+

llama_model_quantize_default_params()

+

Create a LLamaModelQuantizeParams with default values

+
public static LLamaModelQuantizeParams llama_model_quantize_default_params()
+
+

Returns

+

LLamaModelQuantizeParams

+

llama_mmap_supported()

+

Check if memory mapping is supported

+
public static bool llama_mmap_supported()
+
+

Returns

+

Boolean

+

llama_mlock_supported()

+

Check if memory lockingis supported

+
public static bool llama_mlock_supported()
+
+

Returns

+

Boolean

+

llama_eval_export(SafeLLamaContextHandle, String)

+

Export a static computation graph for context of 511 and batch size of 1 + NOTE: since this functionality is mostly for debugging and demonstration purposes, we hardcode these + parameters here to keep things simple + IMPORTANT: do not use for anything else other than debugging and testing!

+
public static int llama_eval_export(SafeLLamaContextHandle ctx, string fname)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

fname String

+

Returns

+

Int32

+

llama_load_model_from_file(String, LLamaContextParams)

+

Various functions for loading a ggml llama model. + Allocate (almost) all memory needed for the model. + Return NULL on failure

+
public static IntPtr llama_load_model_from_file(string path_model, LLamaContextParams params)
+
+

Parameters

+

path_model String

+

params LLamaContextParams

+

Returns

+

IntPtr

+

llama_new_context_with_model(SafeLlamaModelHandle, LLamaContextParams)

+

Create a new llama_context with the given model. + Return value should always be wrapped in SafeLLamaContextHandle!

+
public static IntPtr llama_new_context_with_model(SafeLlamaModelHandle model, LLamaContextParams params)
+
+

Parameters

+

model SafeLlamaModelHandle

+

params LLamaContextParams

+

Returns

+

IntPtr

+

llama_backend_init(Boolean)

+

not great API - very likely to change. + Initialize the llama + ggml backend + Call once at the start of the program

+
public static void llama_backend_init(bool numa)
+
+

Parameters

+

numa Boolean

+

llama_free(IntPtr)

+

Frees all allocated memory in the given llama_context

+
public static void llama_free(IntPtr ctx)
+
+

Parameters

+

ctx IntPtr

+

llama_free_model(IntPtr)

+

Frees all allocated memory associated with a model

+
public static void llama_free_model(IntPtr model)
+
+

Parameters

+

model IntPtr

+

llama_model_apply_lora_from_file(SafeLlamaModelHandle, String, String, Int32)

+

Apply a LoRA adapter to a loaded model + path_base_model is the path to a higher quality model to use as a base for + the layers modified by the adapter. Can be NULL to use the current loaded model. + The model needs to be reloaded before applying a new adapter, otherwise the adapter + will be applied on top of the previous one

+
public static int llama_model_apply_lora_from_file(SafeLlamaModelHandle model_ptr, string path_lora, string path_base_model, int n_threads)
+
+

Parameters

+

model_ptr SafeLlamaModelHandle

+

path_lora String

+

path_base_model String

+

n_threads Int32

+

Returns

+

Int32
+Returns 0 on success

+

llama_get_kv_cache_token_count(SafeLLamaContextHandle)

+

Returns the number of tokens in the KV cache

+
public static int llama_get_kv_cache_token_count(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_set_rng_seed(SafeLLamaContextHandle, Int32)

+

Sets the current rng seed.

+
public static void llama_set_rng_seed(SafeLLamaContextHandle ctx, int seed)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

seed Int32

+

llama_get_state_size(SafeLLamaContextHandle)

+

Returns the maximum size in bytes of the state (rng, logits, embedding + and kv_cache) - will often be smaller after compacting tokens

+
public static ulong llama_get_state_size(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

UInt64

+

llama_copy_state_data(SafeLLamaContextHandle, Byte*)

+

Copies the state to the specified destination address. + Destination needs to have allocated enough memory.

+
public static ulong llama_copy_state_data(SafeLLamaContextHandle ctx, Byte* dest)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

dest Byte*

+

Returns

+

UInt64
+the number of bytes copied

+

llama_copy_state_data(SafeLLamaContextHandle, Byte[])

+

Copies the state to the specified destination address. + Destination needs to have allocated enough memory (see llama_get_state_size)

+
public static ulong llama_copy_state_data(SafeLLamaContextHandle ctx, Byte[] dest)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

dest Byte[]

+

Returns

+

UInt64
+the number of bytes copied

+

llama_set_state_data(SafeLLamaContextHandle, Byte*)

+

Set the state reading from the specified address

+
public static ulong llama_set_state_data(SafeLLamaContextHandle ctx, Byte* src)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

src Byte*

+

Returns

+

UInt64
+the number of bytes read

+

llama_set_state_data(SafeLLamaContextHandle, Byte[])

+

Set the state reading from the specified address

+
public static ulong llama_set_state_data(SafeLLamaContextHandle ctx, Byte[] src)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

src Byte[]

+

Returns

+

UInt64
+the number of bytes read

+

llama_load_session_file(SafeLLamaContextHandle, String, Int32[], UInt64, UInt64*)

+

Load session file

+
public static bool llama_load_session_file(SafeLLamaContextHandle ctx, string path_session, Int32[] tokens_out, ulong n_token_capacity, UInt64* n_token_count_out)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

path_session String

+

tokens_out Int32[]

+

n_token_capacity UInt64

+

n_token_count_out UInt64*

+

Returns

+

Boolean

+

llama_save_session_file(SafeLLamaContextHandle, String, Int32[], UInt64)

+

Save session file

+
public static bool llama_save_session_file(SafeLLamaContextHandle ctx, string path_session, Int32[] tokens, ulong n_token_count)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

path_session String

+

tokens Int32[]

+

n_token_count UInt64

+

Returns

+

Boolean

+

llama_eval(SafeLLamaContextHandle, Int32[], Int32, Int32, Int32)

+

Run the llama inference to obtain the logits and probabilities for the next token. + tokens + n_tokens is the provided batch of new tokens to process + n_past is the number of tokens to use from previous eval calls

+
public static int llama_eval(SafeLLamaContextHandle ctx, Int32[] tokens, int n_tokens, int n_past, int n_threads)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

tokens Int32[]

+

n_tokens Int32

+

n_past Int32

+

n_threads Int32

+

Returns

+

Int32
+Returns 0 on success

+

llama_eval_with_pointer(SafeLLamaContextHandle, Int32*, Int32, Int32, Int32)

+

Run the llama inference to obtain the logits and probabilities for the next token. + tokens + n_tokens is the provided batch of new tokens to process + n_past is the number of tokens to use from previous eval calls

+
public static int llama_eval_with_pointer(SafeLLamaContextHandle ctx, Int32* tokens, int n_tokens, int n_past, int n_threads)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

tokens Int32*

+

n_tokens Int32

+

n_past Int32

+

n_threads Int32

+

Returns

+

Int32
+Returns 0 on success

+

llama_tokenize(SafeLLamaContextHandle, String, Encoding, Int32[], Int32, Boolean)

+

Convert the provided text into tokens.

+
public static int llama_tokenize(SafeLLamaContextHandle ctx, string text, Encoding encoding, Int32[] tokens, int n_max_tokens, bool add_bos)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

text String

+

encoding Encoding

+

tokens Int32[]

+

n_max_tokens Int32

+

add_bos Boolean

+

Returns

+

Int32
+Returns the number of tokens on success, no more than n_max_tokens. + Returns a negative number on failure - the number of tokens that would have been returned

+

llama_tokenize_native(SafeLLamaContextHandle, Byte, Int32, Int32, Boolean)

+

Convert the provided text into tokens.

+
public static int llama_tokenize_native(SafeLLamaContextHandle ctx, Byte* text, Int32* tokens, int n_max_tokens, bool add_bos)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

text Byte*

+

tokens Int32*

+

n_max_tokens Int32

+

add_bos Boolean

+

Returns

+

Int32
+Returns the number of tokens on success, no more than n_max_tokens. + Returns a negative number on failure - the number of tokens that would have been returned

+

llama_n_vocab(SafeLLamaContextHandle)

+

Get the number of tokens in the model vocabulary for this context

+
public static int llama_n_vocab(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_n_ctx(SafeLLamaContextHandle)

+

Get the size of the context window for the model for this context

+
public static int llama_n_ctx(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_n_embd(SafeLLamaContextHandle)

+

Get the dimension of embedding vectors from the model for this context

+
public static int llama_n_embd(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Int32

+

llama_get_logits(SafeLLamaContextHandle)

+

Token logits obtained from the last call to llama_eval() + The logits for the last token are stored in the last row + Can be mutated in order to change the probabilities of the next token.
+ Rows: n_tokens
+ Cols: n_vocab

+
public static Single* llama_get_logits(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Single*

+

llama_get_embeddings(SafeLLamaContextHandle)

+

Get the embeddings for the input + shape: [n_embd] (1-dimensional)

+
public static Single* llama_get_embeddings(SafeLLamaContextHandle ctx)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

Returns

+

Single*

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.safellamacontexthandle/index.html b/0.5/xmldocs/llama.native.safellamacontexthandle/index.html new file mode 100755 index 00000000..5678a29d --- /dev/null +++ b/0.5/xmldocs/llama.native.safellamacontexthandle/index.html @@ -0,0 +1,3388 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.safellamacontexthandle - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

SafeLLamaContextHandle

+

Namespace: LLama.Native

+

A safe wrapper around a llama_context

+
public sealed class SafeLLamaContextHandle : SafeLLamaHandleBase, System.IDisposable
+
+

Inheritance ObjectCriticalFinalizerObjectSafeHandleSafeLLamaHandleBaseSafeLLamaContextHandle
+Implements IDisposable

+

Properties

+

VocabCount

+

Total number of tokens in vocabulary of this model

+
public int VocabCount { get; }
+
+

Property Value

+

Int32

+

ContextSize

+

Total number of tokens in the context

+
public int ContextSize { get; }
+
+

Property Value

+

Int32

+

EmbeddingSize

+

Dimension of embedding vectors

+
public int EmbeddingSize { get; }
+
+

Property Value

+

Int32

+

ModelHandle

+

Get the model which this context is using

+
public SafeLlamaModelHandle ModelHandle { get; }
+
+

Property Value

+

SafeLlamaModelHandle

+

IsInvalid

+
public bool IsInvalid { get; }
+
+

Property Value

+

Boolean

+

IsClosed

+
public bool IsClosed { get; }
+
+

Property Value

+

Boolean

+

Constructors

+

SafeLLamaContextHandle(IntPtr, SafeLlamaModelHandle)

+

Create a new SafeLLamaContextHandle

+
public SafeLLamaContextHandle(IntPtr handle, SafeLlamaModelHandle model)
+
+

Parameters

+

handle IntPtr
+pointer to an allocated llama_context

+

model SafeLlamaModelHandle
+the model which this context was created from

+

Methods

+

ReleaseHandle()

+
protected bool ReleaseHandle()
+
+

Returns

+

Boolean

+

Create(SafeLlamaModelHandle, LLamaContextParams)

+

Create a new llama_state for the given model

+
public static SafeLLamaContextHandle Create(SafeLlamaModelHandle model, LLamaContextParams lparams)
+
+

Parameters

+

model SafeLlamaModelHandle

+

lparams LLamaContextParams

+

Returns

+

SafeLLamaContextHandle

+

Exceptions

+

RuntimeError

+

Clone(LLamaContextParams)

+

Create a new llama context with a clone of the current llama context state

+
public SafeLLamaContextHandle Clone(LLamaContextParams lparams)
+
+

Parameters

+

lparams LLamaContextParams

+

Returns

+

SafeLLamaContextHandle

+

Tokenize(String, Boolean, Encoding)

+

Convert the given text into tokens

+
public Int32[] Tokenize(string text, bool add_bos, Encoding encoding)
+
+

Parameters

+

text String
+The text to tokenize

+

add_bos Boolean
+Whether the "BOS" token should be added

+

encoding Encoding
+Encoding to use for the text

+

Returns

+

Int32[]

+

Exceptions

+

RuntimeError

+

GetLogits()

+

Token logits obtained from the last call to llama_eval() + The logits for the last token are stored in the last row + Can be mutated in order to change the probabilities of the next token.
+ Rows: n_tokens
+ Cols: n_vocab

+
public Span<float> GetLogits()
+
+

Returns

+

Span<Single>

+

TokenToString(Int32, Encoding)

+

Convert a token into a string

+
public string TokenToString(int token, Encoding encoding)
+
+

Parameters

+

token Int32
+Token to decode into a string

+

encoding Encoding

+

Returns

+

String

+

TokenToString(Int32, Encoding, StringBuilder)

+

Append a single llama token to a string builder

+
public void TokenToString(int token, Encoding encoding, StringBuilder dest)
+
+

Parameters

+

token Int32
+Token to decode

+

encoding Encoding

+

dest StringBuilder
+string builder to append the result to

+

TokenToSpan(Int32, Span<Byte>)

+

Convert a single llama token into bytes

+
public int TokenToSpan(int token, Span<byte> dest)
+
+

Parameters

+

token Int32
+Token to decode

+

dest Span<Byte>
+A span to attempt to write into. If this is too small nothing will be written

+

Returns

+

Int32
+The size of this token. nothing will be written if this is larger than dest

+

Eval(ReadOnlySpan<Int32>, Int32, Int32)

+

Run the llama inference to obtain the logits and probabilities for the next token.

+
public bool Eval(ReadOnlySpan<int> tokens, int n_past, int n_threads)
+
+

Parameters

+

tokens ReadOnlySpan<Int32>
+The provided batch of new tokens to process

+

n_past Int32
+the number of tokens to use from previous eval calls

+

n_threads Int32

+

Returns

+

Boolean
+Returns true on success

+

GetStateSize()

+

Get the size of the state, when saved as bytes

+
public ulong GetStateSize()
+
+

Returns

+

UInt64

+

GetState(Byte*, UInt64)

+

Get the raw state of this context, encoded as bytes. Data is written into the dest pointer.

+
public ulong GetState(Byte* dest, ulong size)
+
+

Parameters

+

dest Byte*
+Destination to write to

+

size UInt64
+Number of bytes available to write to in dest (check required size with GetStateSize())

+

Returns

+

UInt64
+The number of bytes written to dest

+

Exceptions

+

ArgumentOutOfRangeException
+Thrown if dest is too small

+

GetState(IntPtr, UInt64)

+

Get the raw state of this context, encoded as bytes. Data is written into the dest pointer.

+
public ulong GetState(IntPtr dest, ulong size)
+
+

Parameters

+

dest IntPtr
+Destination to write to

+

size UInt64
+Number of bytes available to write to in dest (check required size with GetStateSize())

+

Returns

+

UInt64
+The number of bytes written to dest

+

Exceptions

+

ArgumentOutOfRangeException
+Thrown if dest is too small

+

SetState(Byte*)

+

Set the raw state of this context

+
public ulong SetState(Byte* src)
+
+

Parameters

+

src Byte*
+The pointer to read the state from

+

Returns

+

UInt64
+Number of bytes read from the src pointer

+

SetState(IntPtr)

+

Set the raw state of this context

+
public ulong SetState(IntPtr src)
+
+

Parameters

+

src IntPtr
+The pointer to read the state from

+

Returns

+

UInt64
+Number of bytes read from the src pointer

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.safellamagrammarhandle/index.html b/0.5/xmldocs/llama.native.safellamagrammarhandle/index.html new file mode 100755 index 00000000..5d023a41 --- /dev/null +++ b/0.5/xmldocs/llama.native.safellamagrammarhandle/index.html @@ -0,0 +1,2431 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.safellamagrammarhandle - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

SafeLLamaGrammarHandle

+

Namespace: LLama.Native

+

A safe reference to a llama_grammar

+
public class SafeLLamaGrammarHandle : SafeLLamaHandleBase, System.IDisposable
+
+

Inheritance ObjectCriticalFinalizerObjectSafeHandleSafeLLamaHandleBaseSafeLLamaGrammarHandle
+Implements IDisposable

+

Properties

+

IsInvalid

+
public bool IsInvalid { get; }
+
+

Property Value

+

Boolean

+

IsClosed

+
public bool IsClosed { get; }
+
+

Property Value

+

Boolean

+

Methods

+

ReleaseHandle()

+
protected bool ReleaseHandle()
+
+

Returns

+

Boolean

+

Create(IReadOnlyList<GrammarRule>, UInt64)

+

Create a new llama_grammar

+
public static SafeLLamaGrammarHandle Create(IReadOnlyList<GrammarRule> rules, ulong start_rule_index)
+
+

Parameters

+

rules IReadOnlyList<GrammarRule>
+A list of list of elements, each inner list makes up one grammar rule

+

start_rule_index UInt64
+The index (in the outer list) of the start rule

+

Returns

+

SafeLLamaGrammarHandle

+

Exceptions

+

RuntimeError

+

Create(LLamaGrammarElement, UInt64, UInt64)**

+

Create a new llama_grammar

+
public static SafeLLamaGrammarHandle Create(LLamaGrammarElement** rules, ulong nrules, ulong start_rule_index)
+
+

Parameters

+

rules LLamaGrammarElement**
+rules list, each rule is a list of rule elements (terminated by a LLamaGrammarElementType.END element)

+

nrules UInt64
+total number of rules

+

start_rule_index UInt64
+index of the start rule of the grammar

+

Returns

+

SafeLLamaGrammarHandle

+

Exceptions

+

RuntimeError

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.safellamahandlebase/index.html b/0.5/xmldocs/llama.native.safellamahandlebase/index.html new file mode 100755 index 00000000..161736f1 --- /dev/null +++ b/0.5/xmldocs/llama.native.safellamahandlebase/index.html @@ -0,0 +1,2267 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.safellamahandlebase - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

SafeLLamaHandleBase

+

Namespace: LLama.Native

+

Base class for all llama handles to native resources

+
public abstract class SafeLLamaHandleBase : System.Runtime.InteropServices.SafeHandle, System.IDisposable
+
+

Inheritance ObjectCriticalFinalizerObjectSafeHandleSafeLLamaHandleBase
+Implements IDisposable

+

Properties

+

IsInvalid

+
public bool IsInvalid { get; }
+
+

Property Value

+

Boolean

+

IsClosed

+
public bool IsClosed { get; }
+
+

Property Value

+

Boolean

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.safellamamodelhandle/index.html b/0.5/xmldocs/llama.native.safellamamodelhandle/index.html new file mode 100755 index 00000000..bbdc0dfa --- /dev/null +++ b/0.5/xmldocs/llama.native.safellamamodelhandle/index.html @@ -0,0 +1,2856 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.safellamamodelhandle - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

SafeLlamaModelHandle

+

Namespace: LLama.Native

+

A reference to a set of llama model weights

+
public sealed class SafeLlamaModelHandle : SafeLLamaHandleBase, System.IDisposable
+
+

Inheritance ObjectCriticalFinalizerObjectSafeHandleSafeLLamaHandleBaseSafeLlamaModelHandle
+Implements IDisposable

+

Properties

+

VocabCount

+

Total number of tokens in vocabulary of this model

+
public int VocabCount { get; }
+
+

Property Value

+

Int32

+

ContextSize

+

Total number of tokens in the context

+
public int ContextSize { get; }
+
+

Property Value

+

Int32

+

EmbeddingSize

+

Dimension of embedding vectors

+
public int EmbeddingSize { get; }
+
+

Property Value

+

Int32

+

IsInvalid

+
public bool IsInvalid { get; }
+
+

Property Value

+

Boolean

+

IsClosed

+
public bool IsClosed { get; }
+
+

Property Value

+

Boolean

+

Methods

+

ReleaseHandle()

+
protected bool ReleaseHandle()
+
+

Returns

+

Boolean

+

LoadFromFile(String, LLamaContextParams)

+

Load a model from the given file path into memory

+
public static SafeLlamaModelHandle LoadFromFile(string modelPath, LLamaContextParams lparams)
+
+

Parameters

+

modelPath String

+

lparams LLamaContextParams

+

Returns

+

SafeLlamaModelHandle

+

Exceptions

+

RuntimeError

+

ApplyLoraFromFile(String, String, Int32)

+

Apply a LoRA adapter to a loaded model

+
public void ApplyLoraFromFile(string lora, string modelBase, int threads)
+
+

Parameters

+

lora String

+

modelBase String
+A path to a higher quality model to use as a base for the layers modified by the + adapter. Can be NULL to use the current loaded model.

+

threads Int32

+

Exceptions

+

RuntimeError

+

TokenToSpan(Int32, Span<Byte>)

+

Convert a single llama token into bytes

+
public int TokenToSpan(int llama_token, Span<byte> dest)
+
+

Parameters

+

llama_token Int32
+Token to decode

+

dest Span<Byte>
+A span to attempt to write into. If this is too small nothing will be written

+

Returns

+

Int32
+The size of this token. nothing will be written if this is larger than dest

+

TokenToString(Int32, Encoding)

+

Convert a single llama token into a string

+
public string TokenToString(int llama_token, Encoding encoding)
+
+

Parameters

+

llama_token Int32

+

encoding Encoding
+Encoding to use to decode the bytes into a string

+

Returns

+

String

+

TokenToString(Int32, Encoding, StringBuilder)

+

Append a single llama token to a string builder

+
public void TokenToString(int llama_token, Encoding encoding, StringBuilder dest)
+
+

Parameters

+

llama_token Int32
+Token to decode

+

encoding Encoding

+

dest StringBuilder
+string builder to append the result to

+

Tokenize(String, Boolean, Encoding)

+

Convert a string of text into tokens

+
public Int32[] Tokenize(string text, bool add_bos, Encoding encoding)
+
+

Parameters

+

text String

+

add_bos Boolean

+

encoding Encoding

+

Returns

+

Int32[]

+

CreateContext(LLamaContextParams)

+

Create a new context for this model

+
public SafeLLamaContextHandle CreateContext(LLamaContextParams params)
+
+

Parameters

+

params LLamaContextParams

+

Returns

+

SafeLLamaContextHandle

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.native.samplingapi/index.html b/0.5/xmldocs/llama.native.samplingapi/index.html new file mode 100755 index 00000000..75ef781e --- /dev/null +++ b/0.5/xmldocs/llama.native.samplingapi/index.html @@ -0,0 +1,3001 @@ + + + + + + + + + + + + + + + + + + + + + + llama.native.samplingapi - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

SamplingApi

+

Namespace: LLama.Native

+

Direct translation of the llama.cpp sampling API

+
public class SamplingApi
+
+

Inheritance ObjectSamplingApi

+

Constructors

+

SamplingApi()

+
public SamplingApi()
+
+

Methods

+

llama_sample_grammar(SafeLLamaContextHandle, LLamaTokenDataArray, SafeLLamaGrammarHandle)

+

Apply grammar rules to candidate tokens

+
public static void llama_sample_grammar(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, SafeLLamaGrammarHandle grammar)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray

+

grammar SafeLLamaGrammarHandle

+

llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, UInt64, Single)

+

Caution

+

last_tokens_size parameter is no longer needed

+
+

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

+
public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, ulong last_tokens_size, float penalty)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

last_tokens Memory<Int32>

+

last_tokens_size UInt64

+

penalty Single

+

llama_sample_repetition_penalty(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, Single)

+

Repetition penalty described in CTRL academic paper https://arxiv.org/abs/1909.05858, with negative logit fix.

+
public static void llama_sample_repetition_penalty(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, float penalty)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

last_tokens Memory<Int32>

+

penalty Single

+

llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, UInt64, Single, Single)

+

Caution

+

last_tokens_size parameter is no longer needed

+
+

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

+
public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, ulong last_tokens_size, float alpha_frequency, float alpha_presence)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

last_tokens Memory<Int32>

+

last_tokens_size UInt64

+

alpha_frequency Single

+

alpha_presence Single

+

llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle, LLamaTokenDataArray, Memory<Int32>, Single, Single)

+

Frequency and presence penalties described in OpenAI API https://platform.openai.com/docs/api-reference/parameter-details.

+
public static void llama_sample_frequency_and_presence_penalties(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, Memory<int> last_tokens, float alpha_frequency, float alpha_presence)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

last_tokens Memory<Int32>

+

alpha_frequency Single

+

alpha_presence Single

+

llama_sample_softmax(SafeLLamaContextHandle, LLamaTokenDataArray)

+

Sorts candidate tokens by their logits in descending order and calculate probabilities based on logits.

+
public static void llama_sample_softmax(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

llama_sample_top_k(SafeLLamaContextHandle, LLamaTokenDataArray, Int32, UInt64)

+

Top-K sampling described in academic paper "The Curious Case of Neural Text Degeneration" https://arxiv.org/abs/1904.09751

+
public static void llama_sample_top_k(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, int k, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

k Int32

+

min_keep UInt64

+

llama_sample_top_p(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)

+

Nucleus sampling described in academic paper "The Curious Case of Neural Text Degeneration" https://arxiv.org/abs/1904.09751

+
public static void llama_sample_top_p(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float p, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

p Single

+

min_keep UInt64

+

llama_sample_tail_free(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)

+

Tail Free Sampling described in https://www.trentonbricken.com/Tail-Free-Sampling/.

+
public static void llama_sample_tail_free(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float z, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

z Single

+

min_keep UInt64

+

llama_sample_typical(SafeLLamaContextHandle, LLamaTokenDataArray, Single, UInt64)

+

Locally Typical Sampling implementation described in the paper https://arxiv.org/abs/2202.00666.

+
public static void llama_sample_typical(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float p, ulong min_keep)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

p Single

+

min_keep UInt64

+

llama_sample_temperature(SafeLLamaContextHandle, LLamaTokenDataArray, Single)

+

Sample with temperature. + As temperature increases, the prediction becomes diverse but also vulnerable to hallucinations -- generating tokens that are sensible but not factual

+
public static void llama_sample_temperature(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float temp)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray

+

temp Single

+

llama_sample_token_mirostat(SafeLLamaContextHandle, LLamaTokenDataArray, Single, Single, Int32, Single&)

+

Mirostat 1.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

+
public static int llama_sample_token_mirostat(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float tau, float eta, int m, Single& mu)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+A vector of LLamaTokenData containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

+

tau Single
+The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

+

eta Single
+The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

+

m Int32
+The number of tokens considered in the estimation of s_hat. This is an arbitrary value that is used to calculate s_hat, which in turn helps to calculate the value of k. In the paper, they use m = 100, but you can experiment with different values to see how it affects the performance of the algorithm.

+

mu Single&
+Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

+

Returns

+

Int32

+

llama_sample_token_mirostat_v2(SafeLLamaContextHandle, LLamaTokenDataArray, Single, Single, Single&)

+

Mirostat 2.0 algorithm described in the paper https://arxiv.org/abs/2007.14966. Uses tokens instead of words.

+
public static int llama_sample_token_mirostat_v2(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates, float tau, float eta, Single& mu)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+A vector of LLamaTokenData containing the candidate tokens, their probabilities (p), and log-odds (logit) for the current position in the generated text.

+

tau Single
+The target cross-entropy (or surprise) value you want to achieve for the generated text. A higher value corresponds to more surprising or less predictable text, while a lower value corresponds to less surprising or more predictable text.

+

eta Single
+The learning rate used to update mu based on the error between the target and observed surprisal of the sampled word. A larger learning rate will cause mu to be updated more quickly, while a smaller learning rate will result in slower updates.

+

mu Single&
+Maximum cross-entropy. This value is initialized to be twice the target cross-entropy (2 * tau) and is updated in the algorithm based on the error between the target and observed surprisal.

+

Returns

+

Int32

+

llama_sample_token_greedy(SafeLLamaContextHandle, LLamaTokenDataArray)

+

Selects the token with the highest probability.

+
public static int llama_sample_token_greedy(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

Returns

+

Int32

+

llama_sample_token(SafeLLamaContextHandle, LLamaTokenDataArray)

+

Randomly selects a token from the candidates based on their probabilities.

+
public static int llama_sample_token(SafeLLamaContextHandle ctx, LLamaTokenDataArray candidates)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

candidates LLamaTokenDataArray
+Pointer to LLamaTokenDataArray

+

Returns

+

Int32

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletion/index.html b/0.5/xmldocs/llama.oldversion.chatcompletion/index.html new file mode 100755 index 00000000..31684c59 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletion/index.html @@ -0,0 +1,2863 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletion - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletion

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletion : System.IEquatable`1[[LLama.OldVersion.ChatCompletion, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletion
+Implements IEquatable<ChatCompletion>

+

Properties

+

Id

+
public string Id { get; set; }
+
+

Property Value

+

String

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Created

+
public int Created { get; set; }
+
+

Property Value

+

Int32

+

Model

+
public string Model { get; set; }
+
+

Property Value

+

String

+

Choices

+
public ChatCompletionChoice[] Choices { get; set; }
+
+

Property Value

+

ChatCompletionChoice[]

+

Usage

+
public CompletionUsage Usage { get; set; }
+
+

Property Value

+

CompletionUsage

+

Constructors

+

ChatCompletion(String, String, Int32, String, ChatCompletionChoice[], CompletionUsage)

+
public ChatCompletion(string Id, string Object, int Created, string Model, ChatCompletionChoice[] Choices, CompletionUsage Usage)
+
+

Parameters

+

Id String

+

Object String

+

Created Int32

+

Model String

+

Choices ChatCompletionChoice[]

+

Usage CompletionUsage

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletion)

+
public bool Equals(ChatCompletion other)
+
+

Parameters

+

other ChatCompletion

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletion <Clone>$()
+
+

Returns

+

ChatCompletion

+

Deconstruct(String&, String&, Int32&, String&, ChatCompletionChoice[]&, CompletionUsage&)

+
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, ChatCompletionChoice[]& Choices, CompletionUsage& Usage)
+
+

Parameters

+

Id String&

+

Object String&

+

Created Int32&

+

Model String&

+

Choices ChatCompletionChoice[]&

+

Usage CompletionUsage&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletionchoice/index.html b/0.5/xmldocs/llama.oldversion.chatcompletionchoice/index.html new file mode 100755 index 00000000..af88833b --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletionchoice/index.html @@ -0,0 +1,2722 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletionchoice - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletionChoice

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletionChoice : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletionChoice
+Implements IEquatable<ChatCompletionChoice>

+

Properties

+

Index

+
public int Index { get; set; }
+
+

Property Value

+

Int32

+

Message

+
public ChatCompletionMessage Message { get; set; }
+
+

Property Value

+

ChatCompletionMessage

+

FinishReason

+
public string FinishReason { get; set; }
+
+

Property Value

+

String

+

Constructors

+

ChatCompletionChoice(Int32, ChatCompletionMessage, String)

+
public ChatCompletionChoice(int Index, ChatCompletionMessage Message, string FinishReason)
+
+

Parameters

+

Index Int32

+

Message ChatCompletionMessage

+

FinishReason String

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletionChoice)

+
public bool Equals(ChatCompletionChoice other)
+
+

Parameters

+

other ChatCompletionChoice

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletionChoice <Clone>$()
+
+

Returns

+

ChatCompletionChoice

+

Deconstruct(Int32&, ChatCompletionMessage&, String&)

+
public void Deconstruct(Int32& Index, ChatCompletionMessage& Message, String& FinishReason)
+
+

Parameters

+

Index Int32&

+

Message ChatCompletionMessage&

+

FinishReason String&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletionchunk/index.html b/0.5/xmldocs/llama.oldversion.chatcompletionchunk/index.html new file mode 100755 index 00000000..f0bc3953 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletionchunk/index.html @@ -0,0 +1,2816 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletionchunk - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletionChunk

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletionChunk : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunk, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletionChunk
+Implements IEquatable<ChatCompletionChunk>

+

Properties

+

Id

+
public string Id { get; set; }
+
+

Property Value

+

String

+

Model

+
public string Model { get; set; }
+
+

Property Value

+

String

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Created

+
public int Created { get; set; }
+
+

Property Value

+

Int32

+

Choices

+
public ChatCompletionChunkChoice[] Choices { get; set; }
+
+

Property Value

+

ChatCompletionChunkChoice[]

+

Constructors

+

ChatCompletionChunk(String, String, String, Int32, ChatCompletionChunkChoice[])

+
public ChatCompletionChunk(string Id, string Model, string Object, int Created, ChatCompletionChunkChoice[] Choices)
+
+

Parameters

+

Id String

+

Model String

+

Object String

+

Created Int32

+

Choices ChatCompletionChunkChoice[]

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletionChunk)

+
public bool Equals(ChatCompletionChunk other)
+
+

Parameters

+

other ChatCompletionChunk

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletionChunk <Clone>$()
+
+

Returns

+

ChatCompletionChunk

+

Deconstruct(String&, String&, String&, Int32&, ChatCompletionChunkChoice[]&)

+
public void Deconstruct(String& Id, String& Model, String& Object, Int32& Created, ChatCompletionChunkChoice[]& Choices)
+
+

Parameters

+

Id String&

+

Model String&

+

Object String&

+

Created Int32&

+

Choices ChatCompletionChunkChoice[]&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html b/0.5/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html new file mode 100755 index 00000000..baed8a54 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html @@ -0,0 +1,2722 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletionchunkchoice - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletionChunkChoice

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletionChunkChoice : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunkChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletionChunkChoice
+Implements IEquatable<ChatCompletionChunkChoice>

+

Properties

+

Index

+
public int Index { get; set; }
+
+

Property Value

+

Int32

+

Delta

+
public ChatCompletionChunkDelta Delta { get; set; }
+
+

Property Value

+

ChatCompletionChunkDelta

+

FinishReason

+
public string FinishReason { get; set; }
+
+

Property Value

+

String

+

Constructors

+

ChatCompletionChunkChoice(Int32, ChatCompletionChunkDelta, String)

+
public ChatCompletionChunkChoice(int Index, ChatCompletionChunkDelta Delta, string FinishReason)
+
+

Parameters

+

Index Int32

+

Delta ChatCompletionChunkDelta

+

FinishReason String

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletionChunkChoice)

+
public bool Equals(ChatCompletionChunkChoice other)
+
+

Parameters

+

other ChatCompletionChunkChoice

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletionChunkChoice <Clone>$()
+
+

Returns

+

ChatCompletionChunkChoice

+

Deconstruct(Int32&, ChatCompletionChunkDelta&, String&)

+
public void Deconstruct(Int32& Index, ChatCompletionChunkDelta& Delta, String& FinishReason)
+
+

Parameters

+

Index Int32&

+

Delta ChatCompletionChunkDelta&

+

FinishReason String&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html b/0.5/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html new file mode 100755 index 00000000..93a643b6 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html @@ -0,0 +1,2675 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletionchunkdelta - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletionChunkDelta

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletionChunkDelta : System.IEquatable`1[[LLama.OldVersion.ChatCompletionChunkDelta, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletionChunkDelta
+Implements IEquatable<ChatCompletionChunkDelta>

+

Properties

+

Role

+
public string Role { get; set; }
+
+

Property Value

+

String

+

Content

+
public string Content { get; set; }
+
+

Property Value

+

String

+

Constructors

+

ChatCompletionChunkDelta(String, String)

+
public ChatCompletionChunkDelta(string Role, string Content)
+
+

Parameters

+

Role String

+

Content String

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletionChunkDelta)

+
public bool Equals(ChatCompletionChunkDelta other)
+
+

Parameters

+

other ChatCompletionChunkDelta

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletionChunkDelta <Clone>$()
+
+

Returns

+

ChatCompletionChunkDelta

+

Deconstruct(String&, String&)

+
public void Deconstruct(String& Role, String& Content)
+
+

Parameters

+

Role String&

+

Content String&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatcompletionmessage/index.html b/0.5/xmldocs/llama.oldversion.chatcompletionmessage/index.html new file mode 100755 index 00000000..3bcff2d1 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatcompletionmessage/index.html @@ -0,0 +1,2722 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatcompletionmessage - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatCompletionMessage

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatCompletionMessage : System.IEquatable`1[[LLama.OldVersion.ChatCompletionMessage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatCompletionMessage
+Implements IEquatable<ChatCompletionMessage>

+

Properties

+

Role

+
public ChatRole Role { get; set; }
+
+

Property Value

+

ChatRole

+

Content

+
public string Content { get; set; }
+
+

Property Value

+

String

+

Name

+
public string Name { get; set; }
+
+

Property Value

+

String

+

Constructors

+

ChatCompletionMessage(ChatRole, String, String)

+
public ChatCompletionMessage(ChatRole Role, string Content, string Name)
+
+

Parameters

+

Role ChatRole

+

Content String

+

Name String

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatCompletionMessage)

+
public bool Equals(ChatCompletionMessage other)
+
+

Parameters

+

other ChatCompletionMessage

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatCompletionMessage <Clone>$()
+
+

Returns

+

ChatCompletionMessage

+

Deconstruct(ChatRole&, String&, String&)

+
public void Deconstruct(ChatRole& Role, String& Content, String& Name)
+
+

Parameters

+

Role ChatRole&

+

Content String&

+

Name String&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatmessagerecord/index.html b/0.5/xmldocs/llama.oldversion.chatmessagerecord/index.html new file mode 100755 index 00000000..61c7ea1c --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatmessagerecord/index.html @@ -0,0 +1,2675 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatmessagerecord - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatMessageRecord

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatMessageRecord : System.IEquatable`1[[LLama.OldVersion.ChatMessageRecord, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectChatMessageRecord
+Implements IEquatable<ChatMessageRecord>

+

Properties

+

Message

+
public ChatCompletionMessage Message { get; set; }
+
+

Property Value

+

ChatCompletionMessage

+

Time

+
public DateTime Time { get; set; }
+
+

Property Value

+

DateTime

+

Constructors

+

ChatMessageRecord(ChatCompletionMessage, DateTime)

+
public ChatMessageRecord(ChatCompletionMessage Message, DateTime Time)
+
+

Parameters

+

Message ChatCompletionMessage

+

Time DateTime

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(ChatMessageRecord)

+
public bool Equals(ChatMessageRecord other)
+
+

Parameters

+

other ChatMessageRecord

+

Returns

+

Boolean

+

<Clone>$()

+
public ChatMessageRecord <Clone>$()
+
+

Returns

+

ChatMessageRecord

+

Deconstruct(ChatCompletionMessage&, DateTime&)

+
public void Deconstruct(ChatCompletionMessage& Message, DateTime& Time)
+
+

Parameters

+

Message ChatCompletionMessage&

+

Time DateTime&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatrole/index.html b/0.5/xmldocs/llama.oldversion.chatrole/index.html new file mode 100755 index 00000000..a2b18fe8 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatrole/index.html @@ -0,0 +1,2108 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatrole - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+ +
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.chatsession-1/index.html b/0.5/xmldocs/llama.oldversion.chatsession-1/index.html new file mode 100755 index 00000000..5a559a02 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.chatsession-1/index.html @@ -0,0 +1,2457 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.chatsession-1 - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

ChatSession<T>

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class ChatSession<T>
+
+

Type Parameters

+

T

+

Inheritance ObjectChatSession<T>

+

Constructors

+

ChatSession(T)

+
public ChatSession(T model)
+
+

Parameters

+

model T

+

Methods

+

Chat(String, String, String)

+
public IEnumerable<string> Chat(string text, string prompt, string encoding)
+
+

Parameters

+

text String

+

prompt String

+

encoding String

+

Returns

+

IEnumerable<String>

+

WithPrompt(String, String)

+
public ChatSession<T> WithPrompt(string prompt, string encoding)
+
+

Parameters

+

prompt String

+

encoding String

+

Returns

+

ChatSession<T>

+

WithPromptFile(String, String)

+
public ChatSession<T> WithPromptFile(string promptFilename, string encoding)
+
+

Parameters

+

promptFilename String

+

encoding String

+

Returns

+

ChatSession<T>

+

WithAntiprompt(String[])

+

Set the keywords to split the return value of chat AI.

+
public ChatSession<T> WithAntiprompt(String[] antiprompt)
+
+

Parameters

+

antiprompt String[]

+

Returns

+

ChatSession<T>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.completion/index.html b/0.5/xmldocs/llama.oldversion.completion/index.html new file mode 100755 index 00000000..e76db76a --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.completion/index.html @@ -0,0 +1,2863 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.completion - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Completion

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class Completion : System.IEquatable`1[[LLama.OldVersion.Completion, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectCompletion
+Implements IEquatable<Completion>

+

Properties

+

Id

+
public string Id { get; set; }
+
+

Property Value

+

String

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Created

+
public int Created { get; set; }
+
+

Property Value

+

Int32

+

Model

+
public string Model { get; set; }
+
+

Property Value

+

String

+

Choices

+
public CompletionChoice[] Choices { get; set; }
+
+

Property Value

+

CompletionChoice[]

+

Usage

+
public CompletionUsage Usage { get; set; }
+
+

Property Value

+

CompletionUsage

+

Constructors

+

Completion(String, String, Int32, String, CompletionChoice[], CompletionUsage)

+
public Completion(string Id, string Object, int Created, string Model, CompletionChoice[] Choices, CompletionUsage Usage)
+
+

Parameters

+

Id String

+

Object String

+

Created Int32

+

Model String

+

Choices CompletionChoice[]

+

Usage CompletionUsage

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(Completion)

+
public bool Equals(Completion other)
+
+

Parameters

+

other Completion

+

Returns

+

Boolean

+

<Clone>$()

+
public Completion <Clone>$()
+
+

Returns

+

Completion

+

Deconstruct(String&, String&, Int32&, String&, CompletionChoice[]&, CompletionUsage&)

+
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, CompletionChoice[]& Choices, CompletionUsage& Usage)
+
+

Parameters

+

Id String&

+

Object String&

+

Created Int32&

+

Model String&

+

Choices CompletionChoice[]&

+

Usage CompletionUsage&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.completionchoice/index.html b/0.5/xmldocs/llama.oldversion.completionchoice/index.html new file mode 100755 index 00000000..eb75eb70 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.completionchoice/index.html @@ -0,0 +1,2769 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.completionchoice - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

CompletionChoice

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class CompletionChoice : System.IEquatable`1[[LLama.OldVersion.CompletionChoice, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectCompletionChoice
+Implements IEquatable<CompletionChoice>

+

Properties

+

Text

+
public string Text { get; set; }
+
+

Property Value

+

String

+

Index

+
public int Index { get; set; }
+
+

Property Value

+

Int32

+

Logprobs

+
public CompletionLogprobs Logprobs { get; set; }
+
+

Property Value

+

CompletionLogprobs

+

FinishReason

+
public string FinishReason { get; set; }
+
+

Property Value

+

String

+

Constructors

+

CompletionChoice(String, Int32, CompletionLogprobs, String)

+
public CompletionChoice(string Text, int Index, CompletionLogprobs Logprobs, string FinishReason)
+
+

Parameters

+

Text String

+

Index Int32

+

Logprobs CompletionLogprobs

+

FinishReason String

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(CompletionChoice)

+
public bool Equals(CompletionChoice other)
+
+

Parameters

+

other CompletionChoice

+

Returns

+

Boolean

+

<Clone>$()

+
public CompletionChoice <Clone>$()
+
+

Returns

+

CompletionChoice

+

Deconstruct(String&, Int32&, CompletionLogprobs&, String&)

+
public void Deconstruct(String& Text, Int32& Index, CompletionLogprobs& Logprobs, String& FinishReason)
+
+

Parameters

+

Text String&

+

Index Int32&

+

Logprobs CompletionLogprobs&

+

FinishReason String&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.completionchunk/index.html b/0.5/xmldocs/llama.oldversion.completionchunk/index.html new file mode 100755 index 00000000..d89bccdf --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.completionchunk/index.html @@ -0,0 +1,2816 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.completionchunk - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

CompletionChunk

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class CompletionChunk : System.IEquatable`1[[LLama.OldVersion.CompletionChunk, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectCompletionChunk
+Implements IEquatable<CompletionChunk>

+

Properties

+

Id

+
public string Id { get; set; }
+
+

Property Value

+

String

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Created

+
public int Created { get; set; }
+
+

Property Value

+

Int32

+

Model

+
public string Model { get; set; }
+
+

Property Value

+

String

+

Choices

+
public CompletionChoice[] Choices { get; set; }
+
+

Property Value

+

CompletionChoice[]

+

Constructors

+

CompletionChunk(String, String, Int32, String, CompletionChoice[])

+
public CompletionChunk(string Id, string Object, int Created, string Model, CompletionChoice[] Choices)
+
+

Parameters

+

Id String

+

Object String

+

Created Int32

+

Model String

+

Choices CompletionChoice[]

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(CompletionChunk)

+
public bool Equals(CompletionChunk other)
+
+

Parameters

+

other CompletionChunk

+

Returns

+

Boolean

+

<Clone>$()

+
public CompletionChunk <Clone>$()
+
+

Returns

+

CompletionChunk

+

Deconstruct(String&, String&, Int32&, String&, CompletionChoice[]&)

+
public void Deconstruct(String& Id, String& Object, Int32& Created, String& Model, CompletionChoice[]& Choices)
+
+

Parameters

+

Id String&

+

Object String&

+

Created Int32&

+

Model String&

+

Choices CompletionChoice[]&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.completionlogprobs/index.html b/0.5/xmldocs/llama.oldversion.completionlogprobs/index.html new file mode 100755 index 00000000..ffce5309 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.completionlogprobs/index.html @@ -0,0 +1,2769 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.completionlogprobs - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

CompletionLogprobs

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class CompletionLogprobs : System.IEquatable`1[[LLama.OldVersion.CompletionLogprobs, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectCompletionLogprobs
+Implements IEquatable<CompletionLogprobs>

+

Properties

+

TextOffset

+
public Int32[] TextOffset { get; set; }
+
+

Property Value

+

Int32[]

+

TokenLogProbs

+
public Single[] TokenLogProbs { get; set; }
+
+

Property Value

+

Single[]

+

Tokens

+
public String[] Tokens { get; set; }
+
+

Property Value

+

String[]

+

TopLogprobs

+
public Dictionary`2[] TopLogprobs { get; set; }
+
+

Property Value

+

Dictionary`2[]

+

Constructors

+

CompletionLogprobs(Int32[], Single[], String[], Dictionary`2[])

+
public CompletionLogprobs(Int32[] TextOffset, Single[] TokenLogProbs, String[] Tokens, Dictionary`2[] TopLogprobs)
+
+

Parameters

+

TextOffset Int32[]

+

TokenLogProbs Single[]

+

Tokens String[]

+

TopLogprobs Dictionary`2[]

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(CompletionLogprobs)

+
public bool Equals(CompletionLogprobs other)
+
+

Parameters

+

other CompletionLogprobs

+

Returns

+

Boolean

+

<Clone>$()

+
public CompletionLogprobs <Clone>$()
+
+

Returns

+

CompletionLogprobs

+

Deconstruct(Int32[]&, Single[]&, String[]&, Dictionary`2[]&)

+
public void Deconstruct(Int32[]& TextOffset, Single[]& TokenLogProbs, String[]& Tokens, Dictionary`2[]& TopLogprobs)
+
+

Parameters

+

TextOffset Int32[]&

+

TokenLogProbs Single[]&

+

Tokens String[]&

+

TopLogprobs Dictionary`2[]&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.completionusage/index.html b/0.5/xmldocs/llama.oldversion.completionusage/index.html new file mode 100755 index 00000000..856218f9 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.completionusage/index.html @@ -0,0 +1,2722 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.completionusage - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

CompletionUsage

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class CompletionUsage : System.IEquatable`1[[LLama.OldVersion.CompletionUsage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectCompletionUsage
+Implements IEquatable<CompletionUsage>

+

Properties

+

PromptTokens

+
public int PromptTokens { get; set; }
+
+

Property Value

+

Int32

+

CompletionTokens

+
public int CompletionTokens { get; set; }
+
+

Property Value

+

Int32

+

TotalTokens

+
public int TotalTokens { get; set; }
+
+

Property Value

+

Int32

+

Constructors

+

CompletionUsage(Int32, Int32, Int32)

+
public CompletionUsage(int PromptTokens, int CompletionTokens, int TotalTokens)
+
+

Parameters

+

PromptTokens Int32

+

CompletionTokens Int32

+

TotalTokens Int32

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(CompletionUsage)

+
public bool Equals(CompletionUsage other)
+
+

Parameters

+

other CompletionUsage

+

Returns

+

Boolean

+

<Clone>$()

+
public CompletionUsage <Clone>$()
+
+

Returns

+

CompletionUsage

+

Deconstruct(Int32&, Int32&, Int32&)

+
public void Deconstruct(Int32& PromptTokens, Int32& CompletionTokens, Int32& TotalTokens)
+
+

Parameters

+

PromptTokens Int32&

+

CompletionTokens Int32&

+

TotalTokens Int32&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.embedding/index.html b/0.5/xmldocs/llama.oldversion.embedding/index.html new file mode 100755 index 00000000..6aa01c15 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.embedding/index.html @@ -0,0 +1,2769 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.embedding - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

Embedding

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class Embedding : System.IEquatable`1[[LLama.OldVersion.Embedding, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectEmbedding
+Implements IEquatable<Embedding>

+

Properties

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Model

+
public string Model { get; set; }
+
+

Property Value

+

String

+

Data

+
public EmbeddingData[] Data { get; set; }
+
+

Property Value

+

EmbeddingData[]

+

Usage

+
public EmbeddingUsage Usage { get; set; }
+
+

Property Value

+

EmbeddingUsage

+

Constructors

+

Embedding(String, String, EmbeddingData[], EmbeddingUsage)

+
public Embedding(string Object, string Model, EmbeddingData[] Data, EmbeddingUsage Usage)
+
+

Parameters

+

Object String

+

Model String

+

Data EmbeddingData[]

+

Usage EmbeddingUsage

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(Embedding)

+
public bool Equals(Embedding other)
+
+

Parameters

+

other Embedding

+

Returns

+

Boolean

+

<Clone>$()

+
public Embedding <Clone>$()
+
+

Returns

+

Embedding

+

Deconstruct(String&, String&, EmbeddingData[]&, EmbeddingUsage&)

+
public void Deconstruct(String& Object, String& Model, EmbeddingData[]& Data, EmbeddingUsage& Usage)
+
+

Parameters

+

Object String&

+

Model String&

+

Data EmbeddingData[]&

+

Usage EmbeddingUsage&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.embeddingdata/index.html b/0.5/xmldocs/llama.oldversion.embeddingdata/index.html new file mode 100755 index 00000000..ef72d2ef --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.embeddingdata/index.html @@ -0,0 +1,2722 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.embeddingdata - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

EmbeddingData

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class EmbeddingData : System.IEquatable`1[[LLama.OldVersion.EmbeddingData, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectEmbeddingData
+Implements IEquatable<EmbeddingData>

+

Properties

+

Index

+
public int Index { get; set; }
+
+

Property Value

+

Int32

+

Object

+
public string Object { get; set; }
+
+

Property Value

+

String

+

Embedding

+
public Single[] Embedding { get; set; }
+
+

Property Value

+

Single[]

+

Constructors

+

EmbeddingData(Int32, String, Single[])

+
public EmbeddingData(int Index, string Object, Single[] Embedding)
+
+

Parameters

+

Index Int32

+

Object String

+

Embedding Single[]

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(EmbeddingData)

+
public bool Equals(EmbeddingData other)
+
+

Parameters

+

other EmbeddingData

+

Returns

+

Boolean

+

<Clone>$()

+
public EmbeddingData <Clone>$()
+
+

Returns

+

EmbeddingData

+

Deconstruct(Int32&, String&, Single[]&)

+
public void Deconstruct(Int32& Index, String& Object, Single[]& Embedding)
+
+

Parameters

+

Index Int32&

+

Object String&

+

Embedding Single[]&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.embeddingusage/index.html b/0.5/xmldocs/llama.oldversion.embeddingusage/index.html new file mode 100755 index 00000000..8b7ebea7 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.embeddingusage/index.html @@ -0,0 +1,2675 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.embeddingusage - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

EmbeddingUsage

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class EmbeddingUsage : System.IEquatable`1[[LLama.OldVersion.EmbeddingUsage, LLamaSharp, Version=0.5.0.0, Culture=neutral, PublicKeyToken=null]]
+
+

Inheritance ObjectEmbeddingUsage
+Implements IEquatable<EmbeddingUsage>

+

Properties

+

PromptTokens

+
public int PromptTokens { get; set; }
+
+

Property Value

+

Int32

+

TotalTokens

+
public int TotalTokens { get; set; }
+
+

Property Value

+

Int32

+

Constructors

+

EmbeddingUsage(Int32, Int32)

+
public EmbeddingUsage(int PromptTokens, int TotalTokens)
+
+

Parameters

+

PromptTokens Int32

+

TotalTokens Int32

+

Methods

+

ToString()

+
public string ToString()
+
+

Returns

+

String

+

PrintMembers(StringBuilder)

+
protected bool PrintMembers(StringBuilder builder)
+
+

Parameters

+

builder StringBuilder

+

Returns

+

Boolean

+

GetHashCode()

+
public int GetHashCode()
+
+

Returns

+

Int32

+

Equals(Object)

+
public bool Equals(object obj)
+
+

Parameters

+

obj Object

+

Returns

+

Boolean

+

Equals(EmbeddingUsage)

+
public bool Equals(EmbeddingUsage other)
+
+

Parameters

+

other EmbeddingUsage

+

Returns

+

Boolean

+

<Clone>$()

+
public EmbeddingUsage <Clone>$()
+
+

Returns

+

EmbeddingUsage

+

Deconstruct(Int32&, Int32&)

+
public void Deconstruct(Int32& PromptTokens, Int32& TotalTokens)
+
+

Parameters

+

PromptTokens Int32&

+

TotalTokens Int32&

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.ichatmodel/index.html b/0.5/xmldocs/llama.oldversion.ichatmodel/index.html new file mode 100755 index 00000000..b2757082 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.ichatmodel/index.html @@ -0,0 +1,2346 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.ichatmodel - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

IChatModel

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public interface IChatModel
+
+

Properties

+

Name

+
public abstract string Name { get; }
+
+

Property Value

+

String

+

Methods

+

Chat(String, String, String)

+
IEnumerable<string> Chat(string text, string prompt, string encoding)
+
+

Parameters

+

text String

+

prompt String

+

encoding String

+

Returns

+

IEnumerable<String>

+

InitChatPrompt(String, String)

+

Init a prompt for chat and automatically produce the next prompt during the chat.

+
void InitChatPrompt(string prompt, string encoding)
+
+

Parameters

+

prompt String

+

encoding String

+

InitChatAntiprompt(String[])

+
void InitChatAntiprompt(String[] antiprompt)
+
+

Parameters

+

antiprompt String[]

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.llamaembedder/index.html b/0.5/xmldocs/llama.oldversion.llamaembedder/index.html new file mode 100755 index 00000000..cba22d9b --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.llamaembedder/index.html @@ -0,0 +1,2274 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.llamaembedder - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaEmbedder

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class LLamaEmbedder : System.IDisposable
+
+

Inheritance ObjectLLamaEmbedder
+Implements IDisposable

+

Constructors

+

LLamaEmbedder(LLamaParams)

+
public LLamaEmbedder(LLamaParams params)
+
+

Parameters

+

params LLamaParams

+

Methods

+

GetEmbeddings(String, Int32, Boolean, String)

+
public Single[] GetEmbeddings(string text, int n_thread, bool add_bos, string encoding)
+
+

Parameters

+

text String

+

n_thread Int32

+

add_bos Boolean

+

encoding String

+

Returns

+

Single[]

+

Dispose()

+
public void Dispose()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.llamamodel/index.html b/0.5/xmldocs/llama.oldversion.llamamodel/index.html new file mode 100755 index 00000000..46a31749 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.llamamodel/index.html @@ -0,0 +1,3137 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.llamamodel - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaModel

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public class LLamaModel : IChatModel, System.IDisposable
+
+

Inheritance ObjectLLamaModel
+Implements IChatModel, IDisposable

+

Properties

+

Name

+
public string Name { get; set; }
+
+

Property Value

+

String

+

Verbose

+
public bool Verbose { get; set; }
+
+

Property Value

+

Boolean

+

NativeHandle

+
public SafeLLamaContextHandle NativeHandle { get; }
+
+

Property Value

+

SafeLLamaContextHandle

+

Constructors

+

LLamaModel(String, String, Boolean, Int32, Int32, Int32, Int32, Int32, Int32, Int32, Dictionary<Int32, Single>, Int32, Single, Single, Single, Single, Single, Int32, Single, Single, Int32, Single, Single, String, String, String, String, List<String>, String, String, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, String)

+

Please refer LLamaParams to find the meanings of each arg. Be sure to have set the n_gpu_layers, otherwise it will + load 20 layers to gpu by default.

+
public LLamaModel(string model_path, string model_name, bool verbose, int seed, int n_threads, int n_predict, int n_ctx, int n_batch, int n_keep, int n_gpu_layers, Dictionary<int, float> logit_bias, int top_k, float top_p, float tfs_z, float typical_p, float temp, float repeat_penalty, int repeat_last_n, float frequency_penalty, float presence_penalty, int mirostat, float mirostat_tau, float mirostat_eta, string prompt, string path_session, string input_prefix, string input_suffix, List<string> antiprompt, string lora_adapter, string lora_base, bool memory_f16, bool random_prompt, bool use_color, bool interactive, bool embedding, bool interactive_first, bool prompt_cache_all, bool instruct, bool penalize_nl, bool perplexity, bool use_mmap, bool use_mlock, bool mem_test, bool verbose_prompt, string encoding)
+
+

Parameters

+

model_path String
+The model file path.

+

model_name String
+The model name.

+

verbose Boolean
+Whether to print details when running the model.

+

seed Int32

+

n_threads Int32

+

n_predict Int32

+

n_ctx Int32

+

n_batch Int32

+

n_keep Int32

+

n_gpu_layers Int32

+

logit_bias Dictionary<Int32, Single>

+

top_k Int32

+

top_p Single

+

tfs_z Single

+

typical_p Single

+

temp Single

+

repeat_penalty Single

+

repeat_last_n Int32

+

frequency_penalty Single

+

presence_penalty Single

+

mirostat Int32

+

mirostat_tau Single

+

mirostat_eta Single

+

prompt String

+

path_session String

+

input_prefix String

+

input_suffix String

+

antiprompt List<String>

+

lora_adapter String

+

lora_base String

+

memory_f16 Boolean

+

random_prompt Boolean

+

use_color Boolean

+

interactive Boolean

+

embedding Boolean

+

interactive_first Boolean

+

prompt_cache_all Boolean

+

instruct Boolean

+

penalize_nl Boolean

+

perplexity Boolean

+

use_mmap Boolean

+

use_mlock Boolean

+

mem_test Boolean

+

verbose_prompt Boolean

+

encoding String

+

LLamaModel(LLamaParams, String, Boolean, String)

+

Please refer LLamaParams to find the meanings of each arg. Be sure to have set the n_gpu_layers, otherwise it will + load 20 layers to gpu by default.

+
public LLamaModel(LLamaParams params, string name, bool verbose, string encoding)
+
+

Parameters

+

params LLamaParams
+The LLamaModel params

+

name String
+Model name

+

verbose Boolean
+Whether to output the detailed info.

+

encoding String

+

Exceptions

+

RuntimeError

+

Methods

+

WithPrompt(String, String)

+

Apply a prompt to the model.

+
public LLamaModel WithPrompt(string prompt, string encoding)
+
+

Parameters

+

prompt String

+

encoding String

+

Returns

+

LLamaModel

+

Exceptions

+

ArgumentException

+

WithPromptFile(String)

+

Apply the prompt file to the model.

+
public LLamaModel WithPromptFile(string promptFileName)
+
+

Parameters

+

promptFileName String

+

Returns

+

LLamaModel

+

InitChatPrompt(String, String)

+
public void InitChatPrompt(string prompt, string encoding)
+
+

Parameters

+

prompt String

+

encoding String

+

InitChatAntiprompt(String[])

+
public void InitChatAntiprompt(String[] antiprompt)
+
+

Parameters

+

antiprompt String[]

+

Chat(String, String, String)

+

Chat with the LLaMa model under interactive mode.

+
public IEnumerable<string> Chat(string text, string prompt, string encoding)
+
+

Parameters

+

text String

+

prompt String

+

encoding String

+

Returns

+

IEnumerable<String>

+

Exceptions

+

ArgumentException

+

SaveState(String)

+

Save the state to specified path.

+
public void SaveState(string filename)
+
+

Parameters

+

filename String

+

LoadState(String, Boolean)

+

Load the state from specified path.

+
public void LoadState(string filename, bool clearPreviousEmbed)
+
+

Parameters

+

filename String

+

clearPreviousEmbed Boolean
+Whether to clear previous footprints of this model.

+

Exceptions

+

RuntimeError

+

Tokenize(String, String)

+

Tokenize a string.

+
public List<int> Tokenize(string text, string encoding)
+
+

Parameters

+

text String
+The utf-8 encoded string to tokenize.

+

encoding String

+

Returns

+

List<Int32>
+A list of tokens.

+

Exceptions

+

RuntimeError
+If the tokenization failed.

+

DeTokenize(IEnumerable<Int32>)

+

Detokenize a list of tokens.

+
public string DeTokenize(IEnumerable<int> tokens)
+
+

Parameters

+

tokens IEnumerable<Int32>
+The list of tokens to detokenize.

+

Returns

+

String
+The detokenized string.

+

Call(String, String)

+

Call the model to run inference.

+
public IEnumerable<string> Call(string text, string encoding)
+
+

Parameters

+

text String

+

encoding String

+

Returns

+

IEnumerable<String>

+

Exceptions

+

RuntimeError

+

Dispose()

+
public void Dispose()
+
+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.oldversion.llamaparams/index.html b/0.5/xmldocs/llama.oldversion.llamaparams/index.html new file mode 100755 index 00000000..258c1785 --- /dev/null +++ b/0.5/xmldocs/llama.oldversion.llamaparams/index.html @@ -0,0 +1,2947 @@ + + + + + + + + + + + + + + + + + + + + + + llama.oldversion.llamaparams - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

LLamaParams

+

Namespace: LLama.OldVersion

+

Caution

+

The entire LLama.OldVersion namespace will be removed

+
+
public struct LLamaParams
+
+

Inheritance ObjectValueTypeLLamaParams

+

Fields

+

seed

+
public int seed;
+
+

n_threads

+
public int n_threads;
+
+

n_predict

+
public int n_predict;
+
+

n_ctx

+
public int n_ctx;
+
+

n_batch

+
public int n_batch;
+
+

n_keep

+
public int n_keep;
+
+

n_gpu_layers

+
public int n_gpu_layers;
+
+

logit_bias

+
public Dictionary<int, float> logit_bias;
+
+

top_k

+
public int top_k;
+
+

top_p

+
public float top_p;
+
+

tfs_z

+
public float tfs_z;
+
+

typical_p

+
public float typical_p;
+
+

temp

+
public float temp;
+
+

repeat_penalty

+
public float repeat_penalty;
+
+

repeat_last_n

+
public int repeat_last_n;
+
+

frequency_penalty

+
public float frequency_penalty;
+
+

presence_penalty

+
public float presence_penalty;
+
+

mirostat

+
public int mirostat;
+
+

mirostat_tau

+
public float mirostat_tau;
+
+

mirostat_eta

+
public float mirostat_eta;
+
+

model

+
public string model;
+
+

prompt

+
public string prompt;
+
+

path_session

+
public string path_session;
+
+

input_prefix

+
public string input_prefix;
+
+

input_suffix

+
public string input_suffix;
+
+

antiprompt

+
public List<string> antiprompt;
+
+

lora_adapter

+
public string lora_adapter;
+
+

lora_base

+
public string lora_base;
+
+

memory_f16

+
public bool memory_f16;
+
+

random_prompt

+
public bool random_prompt;
+
+

use_color

+
public bool use_color;
+
+

interactive

+
public bool interactive;
+
+

prompt_cache_all

+
public bool prompt_cache_all;
+
+

embedding

+
public bool embedding;
+
+

interactive_first

+
public bool interactive_first;
+
+

instruct

+
public bool instruct;
+
+

penalize_nl

+
public bool penalize_nl;
+
+

perplexity

+
public bool perplexity;
+
+

use_mmap

+
public bool use_mmap;
+
+

use_mlock

+
public bool use_mlock;
+
+

mem_test

+
public bool mem_test;
+
+

verbose_prompt

+
public bool verbose_prompt;
+
+

Constructors

+

LLamaParams(Int32, Int32, Int32, Int32, Int32, Int32, Int32, Dictionary<Int32, Single>, Int32, Single, Single, Single, Single, Single, Int32, Single, Single, Int32, Single, Single, String, String, String, String, String, List<String>, String, String, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean, Boolean)

+
LLamaParams(int seed, int n_threads, int n_predict, int n_ctx, int n_batch, int n_keep, int n_gpu_layers, Dictionary<int, float> logit_bias, int top_k, float top_p, float tfs_z, float typical_p, float temp, float repeat_penalty, int repeat_last_n, float frequency_penalty, float presence_penalty, int mirostat, float mirostat_tau, float mirostat_eta, string model, string prompt, string path_session, string input_prefix, string input_suffix, List<string> antiprompt, string lora_adapter, string lora_base, bool memory_f16, bool random_prompt, bool use_color, bool interactive, bool prompt_cache_all, bool embedding, bool interactive_first, bool instruct, bool penalize_nl, bool perplexity, bool use_mmap, bool use_mlock, bool mem_test, bool verbose_prompt)
+
+

Parameters

+

seed Int32

+

n_threads Int32

+

n_predict Int32

+

n_ctx Int32

+

n_batch Int32

+

n_keep Int32

+

n_gpu_layers Int32

+

logit_bias Dictionary<Int32, Single>

+

top_k Int32

+

top_p Single

+

tfs_z Single

+

typical_p Single

+

temp Single

+

repeat_penalty Single

+

repeat_last_n Int32

+

frequency_penalty Single

+

presence_penalty Single

+

mirostat Int32

+

mirostat_tau Single

+

mirostat_eta Single

+

model String

+

prompt String

+

path_session String

+

input_prefix String

+

input_suffix String

+

antiprompt List<String>

+

lora_adapter String

+

lora_base String

+

memory_f16 Boolean

+

random_prompt Boolean

+

use_color Boolean

+

interactive Boolean

+

prompt_cache_all Boolean

+

embedding Boolean

+

interactive_first Boolean

+

instruct Boolean

+

penalize_nl Boolean

+

perplexity Boolean

+

use_mmap Boolean

+

use_mlock Boolean

+

mem_test Boolean

+

verbose_prompt Boolean

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.statefulexecutorbase/index.html b/0.5/xmldocs/llama.statefulexecutorbase/index.html new file mode 100755 index 00000000..bcd146f4 --- /dev/null +++ b/0.5/xmldocs/llama.statefulexecutorbase/index.html @@ -0,0 +1,2899 @@ + + + + + + + + + + + + + + + + + + + + + + llama.statefulexecutorbase - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + + + + + +
+
+ + + + +

StatefulExecutorBase

+

Namespace: LLama

+

The base class for stateful LLama executors.

+
public abstract class StatefulExecutorBase : LLama.Abstractions.ILLamaExecutor
+
+

Inheritance ObjectStatefulExecutorBase
+Implements ILLamaExecutor

+

Properties

+

Context

+

The context used by the executor.

+
public LLamaContext Context { get; }
+
+

Property Value

+

LLamaContext

+

Methods

+

WithSessionFile(String)

+

This API is currently not verified.

+
public StatefulExecutorBase WithSessionFile(string filename)
+
+

Parameters

+

filename String

+

Returns

+

StatefulExecutorBase

+

Exceptions

+

ArgumentNullException

+

RuntimeError

+

SaveSessionFile(String)

+

This API has not been verified currently.

+
public void SaveSessionFile(string filename)
+
+

Parameters

+

filename String

+

HandleRunOutOfContext(Int32)

+

After running out of the context, take some tokens from the original prompt and recompute the logits in batches.

+
protected void HandleRunOutOfContext(int tokensToKeep)
+
+

Parameters

+

tokensToKeep Int32

+

TryReuseMathingPrefix()

+

Try to reuse the matching prefix from the session file.

+
protected void TryReuseMathingPrefix()
+
+

GetLoopCondition(InferStateArgs)

+

Decide whether to continue the loop.

+
protected abstract bool GetLoopCondition(InferStateArgs args)
+
+

Parameters

+

args InferStateArgs

+

Returns

+

Boolean

+

PreprocessInputs(String, InferStateArgs)

+

Preprocess the inputs before the inference.

+
protected abstract void PreprocessInputs(string text, InferStateArgs args)
+
+

Parameters

+

text String

+

args InferStateArgs

+

PostProcess(IInferenceParams, InferStateArgs, IEnumerable`1&)

+

Do some post processing after the inference.

+
protected abstract bool PostProcess(IInferenceParams inferenceParams, InferStateArgs args, IEnumerable`1& extraOutputs)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+

extraOutputs IEnumerable`1&

+

Returns

+

Boolean

+

InferInternal(IInferenceParams, InferStateArgs)

+

The core inference logic.

+
protected abstract void InferInternal(IInferenceParams inferenceParams, InferStateArgs args)
+
+

Parameters

+

inferenceParams IInferenceParams

+

args InferStateArgs

+

SaveState(String)

+

Save the current state to a file.

+
public abstract void SaveState(string filename)
+
+

Parameters

+

filename String

+

GetStateData()

+

Get the current state data.

+
public abstract ExecutorBaseState GetStateData()
+
+

Returns

+

ExecutorBaseState

+

LoadState(ExecutorBaseState)

+

Load the state from data.

+
public abstract void LoadState(ExecutorBaseState data)
+
+

Parameters

+

data ExecutorBaseState

+

LoadState(String)

+

Load the state from a file.

+
public abstract void LoadState(string filename)
+
+

Parameters

+

filename String

+

Infer(String, IInferenceParams, CancellationToken)

+

Execute the inference.

+
public IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

text String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IEnumerable<String>

+

InferAsync(String, IInferenceParams, CancellationToken)

+

Execute the inference asynchronously.

+
public IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

text String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IAsyncEnumerable<String>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.statelessexecutor/index.html b/0.5/xmldocs/llama.statelessexecutor/index.html new file mode 100755 index 00000000..580f8a45 --- /dev/null +++ b/0.5/xmldocs/llama.statelessexecutor/index.html @@ -0,0 +1,2442 @@ + + + + + + + + + + + + + + + + + + + + + + llama.statelessexecutor - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + +
+
+
+ + + +
+
+
+ + + +
+
+ + + + +

StatelessExecutor

+

Namespace: LLama

+

This executor infer the input as one-time job. Previous inputs won't impact on the + response to current input.

+
public class StatelessExecutor : LLama.Abstractions.ILLamaExecutor
+
+

Inheritance ObjectStatelessExecutor
+Implements ILLamaExecutor

+

Properties

+

Context

+

The context used by the executor when running the inference.

+
public LLamaContext Context { get; private set; }
+
+

Property Value

+

LLamaContext

+

Constructors

+

StatelessExecutor(LLamaWeights, IModelParams)

+

Create a new stateless executor which will use the given model

+
public StatelessExecutor(LLamaWeights weights, IModelParams params)
+
+

Parameters

+

weights LLamaWeights

+

params IModelParams

+

StatelessExecutor(LLamaContext)

+

Caution

+

Use the constructor which automatically creates contexts using the LLamaWeights

+
+

Create a new stateless executor which will use the model used to create the given context

+
public StatelessExecutor(LLamaContext context)
+
+

Parameters

+

context LLamaContext

+

Methods

+

Infer(String, IInferenceParams, CancellationToken)

+
public IEnumerable<string> Infer(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

text String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IEnumerable<String>

+

InferAsync(String, IInferenceParams, CancellationToken)

+
public IAsyncEnumerable<string> InferAsync(string text, IInferenceParams inferenceParams, CancellationToken cancellationToken)
+
+

Parameters

+

text String

+

inferenceParams IInferenceParams

+

cancellationToken CancellationToken

+

Returns

+

IAsyncEnumerable<String>

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/0.5/xmldocs/llama.utils/index.html b/0.5/xmldocs/llama.utils/index.html new file mode 100755 index 00000000..6159d852 --- /dev/null +++ b/0.5/xmldocs/llama.utils/index.html @@ -0,0 +1,2582 @@ + + + + + + + + + + + + + + + + + + + + llama.utils - LLamaSharp Documentation + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + + +
+ +
+ + + + + + + + +
+ + +
+ +
+ + + + + + +
+
+ + + +
+
+
+ + + + +
+
+
+ + + + + + + +
+
+ + + + +

Utils

+

Namespace: LLama

+

Assorted llama utilities

+
public static class Utils
+
+

Inheritance ObjectUtils

+

Methods

+

InitLLamaContextFromModelParams(IModelParams)

+

Caution

+

Use LLamaWeights.LoadFromFile and LLamaWeights.CreateContext instead

+
+
public static SafeLLamaContextHandle InitLLamaContextFromModelParams(IModelParams params)
+
+

Parameters

+

params IModelParams

+

Returns

+

SafeLLamaContextHandle

+

Tokenize(SafeLLamaContextHandle, String, Boolean, Encoding)

+

Caution

+

Use SafeLLamaContextHandle Tokenize method instead

+
+
public static IEnumerable<int> Tokenize(SafeLLamaContextHandle ctx, string text, bool add_bos, Encoding encoding)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

text String

+

add_bos Boolean

+

encoding Encoding

+

Returns

+

IEnumerable<Int32>

+

GetLogits(SafeLLamaContextHandle, Int32)

+

Caution

+

Use SafeLLamaContextHandle GetLogits method instead

+
+
public static Span<float> GetLogits(SafeLLamaContextHandle ctx, int length)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

length Int32

+

Returns

+

Span<Single>

+

Eval(SafeLLamaContextHandle, Int32[], Int32, Int32, Int32, Int32)

+

Caution

+

Use SafeLLamaContextHandle Eval method instead

+
+
public static int Eval(SafeLLamaContextHandle ctx, Int32[] tokens, int startIndex, int n_tokens, int n_past, int n_threads)
+
+

Parameters

+

ctx SafeLLamaContextHandle

+

tokens Int32[]

+

startIndex Int32

+

n_tokens Int32

+

n_past Int32

+

n_threads Int32

+

Returns

+

Int32

+

TokenToString(Int32, SafeLLamaContextHandle, Encoding)

+

Caution

+

Use SafeLLamaContextHandle TokenToString method instead

+
+
public static string TokenToString(int token, SafeLLamaContextHandle ctx, Encoding encoding)
+
+

Parameters

+

token Int32

+

ctx SafeLLamaContextHandle

+

encoding Encoding

+

Returns

+

String

+

PtrToString(IntPtr, Encoding)

+

Caution

+

No longer used internally by LlamaSharp

+
+
public static string PtrToString(IntPtr ptr, Encoding encoding)
+
+

Parameters

+

ptr IntPtr

+

encoding Encoding

+

Returns

+

String

+ + + + + + +
+
+ + +
+ +
+ + + +
+
+
+
+ + + + + + + + + \ No newline at end of file diff --git a/latest/404.html b/latest/404.html index e2b60413..6dfa0d53 100644 --- a/latest/404.html +++ b/latest/404.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../0.4/404.html... + Redirecting to ../0.5/404.html... \ No newline at end of file diff --git a/latest/Architecture/index.html b/latest/Architecture/index.html index 803acd82..bfdcbd89 100644 --- a/latest/Architecture/index.html +++ b/latest/Architecture/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../0.4/Architecture/... + Redirecting to ../../0.5/Architecture/... \ No newline at end of file diff --git a/latest/ChatSession/basic-usages/index.html b/latest/ChatSession/basic-usages/index.html index dd84d615..d5bb823a 100644 --- a/latest/ChatSession/basic-usages/index.html +++ b/latest/ChatSession/basic-usages/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/ChatSession/basic-usages/... + Redirecting to ../../../0.5/ChatSession/basic-usages/... \ No newline at end of file diff --git a/latest/ChatSession/save-load-session/index.html b/latest/ChatSession/save-load-session/index.html index 2b362438..8ec617e7 100644 --- a/latest/ChatSession/save-load-session/index.html +++ b/latest/ChatSession/save-load-session/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/ChatSession/save-load-session/... + Redirecting to ../../../0.5/ChatSession/save-load-session/... \ No newline at end of file diff --git a/latest/ChatSession/transforms/index.html b/latest/ChatSession/transforms/index.html index 4699105e..2bc92b45 100644 --- a/latest/ChatSession/transforms/index.html +++ b/latest/ChatSession/transforms/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/ChatSession/transforms/... + Redirecting to ../../../0.5/ChatSession/transforms/... \ No newline at end of file diff --git a/latest/ContributingGuide/index.html b/latest/ContributingGuide/index.html index 1ffa3652..4c538668 100644 --- a/latest/ContributingGuide/index.html +++ b/latest/ContributingGuide/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../0.4/ContributingGuide/... + Redirecting to ../../0.5/ContributingGuide/... \ No newline at end of file diff --git a/latest/Examples/ChatSessionStripRoleName/index.html b/latest/Examples/ChatSessionStripRoleName/index.html index 47a12b88..dbfac7e4 100644 --- a/latest/Examples/ChatSessionStripRoleName/index.html +++ b/latest/Examples/ChatSessionStripRoleName/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/ChatSessionStripRoleName/... + Redirecting to ../../../0.5/Examples/ChatSessionStripRoleName/... \ No newline at end of file diff --git a/latest/Examples/ChatSessionWithRoleName/index.html b/latest/Examples/ChatSessionWithRoleName/index.html index adedfd2c..46a48a71 100644 --- a/latest/Examples/ChatSessionWithRoleName/index.html +++ b/latest/Examples/ChatSessionWithRoleName/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/ChatSessionWithRoleName/... + Redirecting to ../../../0.5/Examples/ChatSessionWithRoleName/... \ No newline at end of file diff --git a/latest/Examples/GetEmbeddings/index.html b/latest/Examples/GetEmbeddings/index.html index 0c29c365..8721fde1 100644 --- a/latest/Examples/GetEmbeddings/index.html +++ b/latest/Examples/GetEmbeddings/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/GetEmbeddings/... + Redirecting to ../../../0.5/Examples/GetEmbeddings/... \ No newline at end of file diff --git a/latest/Examples/InstructModeExecute/index.html b/latest/Examples/InstructModeExecute/index.html index c50f48d6..d4cf4c35 100644 --- a/latest/Examples/InstructModeExecute/index.html +++ b/latest/Examples/InstructModeExecute/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/InstructModeExecute/... + Redirecting to ../../../0.5/Examples/InstructModeExecute/... \ No newline at end of file diff --git a/latest/Examples/InteractiveModeExecute/index.html b/latest/Examples/InteractiveModeExecute/index.html index c24f0b5a..a7d73348 100644 --- a/latest/Examples/InteractiveModeExecute/index.html +++ b/latest/Examples/InteractiveModeExecute/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/InteractiveModeExecute/... + Redirecting to ../../../0.5/Examples/InteractiveModeExecute/... \ No newline at end of file diff --git a/latest/Examples/LoadAndSaveSession/index.html b/latest/Examples/LoadAndSaveSession/index.html index 16728338..cad732bd 100644 --- a/latest/Examples/LoadAndSaveSession/index.html +++ b/latest/Examples/LoadAndSaveSession/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/LoadAndSaveSession/... + Redirecting to ../../../0.5/Examples/LoadAndSaveSession/... \ No newline at end of file diff --git a/latest/Examples/LoadAndSaveState/index.html b/latest/Examples/LoadAndSaveState/index.html index fa92765b..ee1c4390 100644 --- a/latest/Examples/LoadAndSaveState/index.html +++ b/latest/Examples/LoadAndSaveState/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/LoadAndSaveState/... + Redirecting to ../../../0.5/Examples/LoadAndSaveState/... \ No newline at end of file diff --git a/latest/Examples/QuantizeModel/index.html b/latest/Examples/QuantizeModel/index.html index 83aa5cf6..2bed73a0 100644 --- a/latest/Examples/QuantizeModel/index.html +++ b/latest/Examples/QuantizeModel/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/QuantizeModel/... + Redirecting to ../../../0.5/Examples/QuantizeModel/... \ No newline at end of file diff --git a/latest/Examples/StatelessModeExecute/index.html b/latest/Examples/StatelessModeExecute/index.html index f6501675..9f5632f1 100644 --- a/latest/Examples/StatelessModeExecute/index.html +++ b/latest/Examples/StatelessModeExecute/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/Examples/StatelessModeExecute/... + Redirecting to ../../../0.5/Examples/StatelessModeExecute/... \ No newline at end of file diff --git a/latest/GetStarted/index.html b/latest/GetStarted/index.html index 3489fc44..2d536da5 100644 --- a/latest/GetStarted/index.html +++ b/latest/GetStarted/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../0.4/GetStarted/... + Redirecting to ../../0.5/GetStarted/... \ No newline at end of file diff --git a/latest/HighLevelApps/bot-sharp/index.html b/latest/HighLevelApps/bot-sharp/index.html index c2b27ea7..527e0ec0 100644 --- a/latest/HighLevelApps/bot-sharp/index.html +++ b/latest/HighLevelApps/bot-sharp/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/HighLevelApps/bot-sharp/... + Redirecting to ../../../0.5/HighLevelApps/bot-sharp/... \ No newline at end of file diff --git a/latest/HighLevelApps/semantic-kernel/index.html b/latest/HighLevelApps/semantic-kernel/index.html new file mode 100644 index 00000000..3026e6ff --- /dev/null +++ b/latest/HighLevelApps/semantic-kernel/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/HighLevelApps/semantic-kernel/... + + \ No newline at end of file diff --git a/latest/LLamaExecutors/differences/index.html b/latest/LLamaExecutors/differences/index.html index fd70331d..0596cdfb 100644 --- a/latest/LLamaExecutors/differences/index.html +++ b/latest/LLamaExecutors/differences/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaExecutors/differences/... + Redirecting to ../../../0.5/LLamaExecutors/differences/... \ No newline at end of file diff --git a/latest/LLamaExecutors/parameters/index.html b/latest/LLamaExecutors/parameters/index.html index 488f2c3d..b1ca21d0 100644 --- a/latest/LLamaExecutors/parameters/index.html +++ b/latest/LLamaExecutors/parameters/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaExecutors/parameters/... + Redirecting to ../../../0.5/LLamaExecutors/parameters/... \ No newline at end of file diff --git a/latest/LLamaExecutors/save-load-state/index.html b/latest/LLamaExecutors/save-load-state/index.html index e664cedd..97b6d21e 100644 --- a/latest/LLamaExecutors/save-load-state/index.html +++ b/latest/LLamaExecutors/save-load-state/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaExecutors/save-load-state/... + Redirecting to ../../../0.5/LLamaExecutors/save-load-state/... \ No newline at end of file diff --git a/latest/LLamaExecutors/text-to-text-apis/index.html b/latest/LLamaExecutors/text-to-text-apis/index.html index 04652ebe..e9eb487e 100644 --- a/latest/LLamaExecutors/text-to-text-apis/index.html +++ b/latest/LLamaExecutors/text-to-text-apis/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaExecutors/text-to-text-apis/... + Redirecting to ../../../0.5/LLamaExecutors/text-to-text-apis/... \ No newline at end of file diff --git a/latest/LLamaModel/embeddings/index.html b/latest/LLamaModel/embeddings/index.html index 6b411ec4..4cc9eed6 100644 --- a/latest/LLamaModel/embeddings/index.html +++ b/latest/LLamaModel/embeddings/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaModel/embeddings/... + Redirecting to ../../../0.5/LLamaModel/embeddings/... \ No newline at end of file diff --git a/latest/LLamaModel/parameters/index.html b/latest/LLamaModel/parameters/index.html index f2bf323b..0aad0213 100644 --- a/latest/LLamaModel/parameters/index.html +++ b/latest/LLamaModel/parameters/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaModel/parameters/... + Redirecting to ../../../0.5/LLamaModel/parameters/... \ No newline at end of file diff --git a/latest/LLamaModel/quantization/index.html b/latest/LLamaModel/quantization/index.html index 7bc98708..3d6e243b 100644 --- a/latest/LLamaModel/quantization/index.html +++ b/latest/LLamaModel/quantization/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaModel/quantization/... + Redirecting to ../../../0.5/LLamaModel/quantization/... \ No newline at end of file diff --git a/latest/LLamaModel/save-load-state/index.html b/latest/LLamaModel/save-load-state/index.html index 19bec36f..34f1070b 100644 --- a/latest/LLamaModel/save-load-state/index.html +++ b/latest/LLamaModel/save-load-state/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaModel/save-load-state/... + Redirecting to ../../../0.5/LLamaModel/save-load-state/... \ No newline at end of file diff --git a/latest/LLamaModel/tokenization/index.html b/latest/LLamaModel/tokenization/index.html index 83a40bc2..f57940a4 100644 --- a/latest/LLamaModel/tokenization/index.html +++ b/latest/LLamaModel/tokenization/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/LLamaModel/tokenization/... + Redirecting to ../../../0.5/LLamaModel/tokenization/... \ No newline at end of file diff --git a/latest/More/log/index.html b/latest/More/log/index.html index 9243d647..bec86e09 100644 --- a/latest/More/log/index.html +++ b/latest/More/log/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/More/log/... + Redirecting to ../../../0.5/More/log/... \ No newline at end of file diff --git a/latest/NonEnglishUsage/Chinese/index.html b/latest/NonEnglishUsage/Chinese/index.html index 81232dfd..0693c54e 100644 --- a/latest/NonEnglishUsage/Chinese/index.html +++ b/latest/NonEnglishUsage/Chinese/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/NonEnglishUsage/Chinese/... + Redirecting to ../../../0.5/NonEnglishUsage/Chinese/... \ No newline at end of file diff --git a/latest/Tricks/index.html b/latest/Tricks/index.html index 3a58e26a..ae575e7f 100644 --- a/latest/Tricks/index.html +++ b/latest/Tricks/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../0.4/Tricks/... + Redirecting to ../../0.5/Tricks/... \ No newline at end of file diff --git a/latest/index.html b/latest/index.html index cd6dfdb2..c153150b 100644 --- a/latest/index.html +++ b/latest/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../0.4/... + Redirecting to ../0.5/... \ No newline at end of file diff --git a/latest/xmldocs/index.html b/latest/xmldocs/index.html index e8fa666c..91dd8f1e 100644 --- a/latest/xmldocs/index.html +++ b/latest/xmldocs/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../0.4/xmldocs/... + Redirecting to ../../0.5/xmldocs/... \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.ihistorytransform/index.html b/latest/xmldocs/llama.abstractions.ihistorytransform/index.html index 829c5d23..b7e993b9 100644 --- a/latest/xmldocs/llama.abstractions.ihistorytransform/index.html +++ b/latest/xmldocs/llama.abstractions.ihistorytransform/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.abstractions.ihistorytransform/... + Redirecting to ../../../0.5/xmldocs/llama.abstractions.ihistorytransform/... \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.iinferenceparams/index.html b/latest/xmldocs/llama.abstractions.iinferenceparams/index.html new file mode 100644 index 00000000..b49d7ed7 --- /dev/null +++ b/latest/xmldocs/llama.abstractions.iinferenceparams/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.abstractions.iinferenceparams/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.illamaexecutor/index.html b/latest/xmldocs/llama.abstractions.illamaexecutor/index.html index a645804e..711bd0a2 100644 --- a/latest/xmldocs/llama.abstractions.illamaexecutor/index.html +++ b/latest/xmldocs/llama.abstractions.illamaexecutor/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.abstractions.illamaexecutor/... + Redirecting to ../../../0.5/xmldocs/llama.abstractions.illamaexecutor/... \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.imodelparams/index.html b/latest/xmldocs/llama.abstractions.imodelparams/index.html new file mode 100644 index 00000000..a9f07489 --- /dev/null +++ b/latest/xmldocs/llama.abstractions.imodelparams/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.abstractions.imodelparams/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.itextstreamtransform/index.html b/latest/xmldocs/llama.abstractions.itextstreamtransform/index.html index 21ecf4d4..8e0c017d 100644 --- a/latest/xmldocs/llama.abstractions.itextstreamtransform/index.html +++ b/latest/xmldocs/llama.abstractions.itextstreamtransform/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.abstractions.itextstreamtransform/... + Redirecting to ../../../0.5/xmldocs/llama.abstractions.itextstreamtransform/... \ No newline at end of file diff --git a/latest/xmldocs/llama.abstractions.itexttransform/index.html b/latest/xmldocs/llama.abstractions.itexttransform/index.html index c3a54684..73f583ff 100644 --- a/latest/xmldocs/llama.abstractions.itexttransform/index.html +++ b/latest/xmldocs/llama.abstractions.itexttransform/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.abstractions.itexttransform/... + Redirecting to ../../../0.5/xmldocs/llama.abstractions.itexttransform/... \ No newline at end of file diff --git a/latest/xmldocs/llama.chatsession/index.html b/latest/xmldocs/llama.chatsession/index.html index 1966a4b6..5ac2d03a 100644 --- a/latest/xmldocs/llama.chatsession/index.html +++ b/latest/xmldocs/llama.chatsession/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.chatsession/... + Redirecting to ../../../0.5/xmldocs/llama.chatsession/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.authorrole/index.html b/latest/xmldocs/llama.common.authorrole/index.html index 008ca669..042f8da3 100644 --- a/latest/xmldocs/llama.common.authorrole/index.html +++ b/latest/xmldocs/llama.common.authorrole/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.authorrole/... + Redirecting to ../../../0.5/xmldocs/llama.common.authorrole/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.chathistory/index.html b/latest/xmldocs/llama.common.chathistory/index.html index d1bbd858..ed216b24 100644 --- a/latest/xmldocs/llama.common.chathistory/index.html +++ b/latest/xmldocs/llama.common.chathistory/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.chathistory/... + Redirecting to ../../../0.5/xmldocs/llama.common.chathistory/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.fixedsizequeue-1/index.html b/latest/xmldocs/llama.common.fixedsizequeue-1/index.html index 3d2f4e94..0299edaf 100644 --- a/latest/xmldocs/llama.common.fixedsizequeue-1/index.html +++ b/latest/xmldocs/llama.common.fixedsizequeue-1/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.fixedsizequeue-1/... + Redirecting to ../../../0.5/xmldocs/llama.common.fixedsizequeue-1/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.illamalogger/index.html b/latest/xmldocs/llama.common.illamalogger/index.html index 98627e8f..7181fc47 100644 --- a/latest/xmldocs/llama.common.illamalogger/index.html +++ b/latest/xmldocs/llama.common.illamalogger/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.illamalogger/... + Redirecting to ../../../0.5/xmldocs/llama.common.illamalogger/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.inferenceparams/index.html b/latest/xmldocs/llama.common.inferenceparams/index.html index c5d25da9..90aa3e04 100644 --- a/latest/xmldocs/llama.common.inferenceparams/index.html +++ b/latest/xmldocs/llama.common.inferenceparams/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.inferenceparams/... + Redirecting to ../../../0.5/xmldocs/llama.common.inferenceparams/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.llamadefaultlogger/index.html b/latest/xmldocs/llama.common.llamadefaultlogger/index.html index e7956022..abc2f828 100644 --- a/latest/xmldocs/llama.common.llamadefaultlogger/index.html +++ b/latest/xmldocs/llama.common.llamadefaultlogger/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.llamadefaultlogger/... + Redirecting to ../../../0.5/xmldocs/llama.common.llamadefaultlogger/... \ No newline at end of file diff --git a/latest/xmldocs/llama.common.mirostatetype/index.html b/latest/xmldocs/llama.common.mirostatetype/index.html deleted file mode 100644 index 6fea7b44..00000000 --- a/latest/xmldocs/llama.common.mirostatetype/index.html +++ /dev/null @@ -1,16 +0,0 @@ - - - - - Redirecting - - - - - Redirecting to ../../../0.4/xmldocs/llama.common.mirostatetype/... - - \ No newline at end of file diff --git a/latest/xmldocs/llama.common.mirostattype/index.html b/latest/xmldocs/llama.common.mirostattype/index.html new file mode 100644 index 00000000..810fcc6d --- /dev/null +++ b/latest/xmldocs/llama.common.mirostattype/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.common.mirostattype/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.common.modelparams/index.html b/latest/xmldocs/llama.common.modelparams/index.html index 106b8408..982fde5c 100644 --- a/latest/xmldocs/llama.common.modelparams/index.html +++ b/latest/xmldocs/llama.common.modelparams/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.common.modelparams/... + Redirecting to ../../../0.5/xmldocs/llama.common.modelparams/... \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarexpectedname/index.html b/latest/xmldocs/llama.exceptions.grammarexpectedname/index.html new file mode 100644 index 00000000..fe5a735f --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarexpectedname/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarexpectedname/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarexpectednext/index.html b/latest/xmldocs/llama.exceptions.grammarexpectednext/index.html new file mode 100644 index 00000000..cd7a8a66 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarexpectednext/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarexpectednext/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarexpectedprevious/index.html b/latest/xmldocs/llama.exceptions.grammarexpectedprevious/index.html new file mode 100644 index 00000000..4399d2ae --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarexpectedprevious/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarexpectedprevious/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarformatexception/index.html b/latest/xmldocs/llama.exceptions.grammarformatexception/index.html new file mode 100644 index 00000000..c119d736 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarformatexception/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarformatexception/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html b/latest/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html new file mode 100644 index 00000000..19deae50 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunexpectedcharaltelement/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html b/latest/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html new file mode 100644 index 00000000..a3f3fbb7 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunexpectedcharrngelement/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html b/latest/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html new file mode 100644 index 00000000..faf804fa --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunexpectedendelement/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunexpectedendelement/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html b/latest/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html new file mode 100644 index 00000000..ed6ebd3b --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunexpectedendofinput/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunexpectedendofinput/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html b/latest/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html new file mode 100644 index 00000000..ad3e96a5 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunexpectedhexcharscount/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html b/latest/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html new file mode 100644 index 00000000..9c8da207 --- /dev/null +++ b/latest/xmldocs/llama.exceptions.grammarunknownescapecharacter/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.exceptions.grammarunknownescapecharacter/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.exceptions.runtimeerror/index.html b/latest/xmldocs/llama.exceptions.runtimeerror/index.html index ff1becfa..bd1ba399 100644 --- a/latest/xmldocs/llama.exceptions.runtimeerror/index.html +++ b/latest/xmldocs/llama.exceptions.runtimeerror/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.exceptions.runtimeerror/... + Redirecting to ../../../0.5/xmldocs/llama.exceptions.runtimeerror/... \ No newline at end of file diff --git a/latest/xmldocs/llama.extensions.dictionaryextension/index.html b/latest/xmldocs/llama.extensions.dictionaryextension/index.html deleted file mode 100644 index 2da66de8..00000000 --- a/latest/xmldocs/llama.extensions.dictionaryextension/index.html +++ /dev/null @@ -1,16 +0,0 @@ - - - - - Redirecting - - - - - Redirecting to ../../../0.4/xmldocs/llama.extensions.dictionaryextension/... - - \ No newline at end of file diff --git a/latest/xmldocs/llama.extensions.imodelparamsextensions/index.html b/latest/xmldocs/llama.extensions.imodelparamsextensions/index.html new file mode 100644 index 00000000..e5d8aa03 --- /dev/null +++ b/latest/xmldocs/llama.extensions.imodelparamsextensions/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.extensions.imodelparamsextensions/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.extensions.keyvaluepairextensions/index.html b/latest/xmldocs/llama.extensions.keyvaluepairextensions/index.html new file mode 100644 index 00000000..7ae9f95e --- /dev/null +++ b/latest/xmldocs/llama.extensions.keyvaluepairextensions/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.extensions.keyvaluepairextensions/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.grammars.grammar/index.html b/latest/xmldocs/llama.grammars.grammar/index.html new file mode 100644 index 00000000..8651a9b1 --- /dev/null +++ b/latest/xmldocs/llama.grammars.grammar/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.grammars.grammar/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.grammars.grammarrule/index.html b/latest/xmldocs/llama.grammars.grammarrule/index.html new file mode 100644 index 00000000..fece5f52 --- /dev/null +++ b/latest/xmldocs/llama.grammars.grammarrule/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.grammars.grammarrule/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.instructexecutor/index.html b/latest/xmldocs/llama.instructexecutor/index.html index 6751fafa..1eacf6bc 100644 --- a/latest/xmldocs/llama.instructexecutor/index.html +++ b/latest/xmldocs/llama.instructexecutor/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.instructexecutor/... + Redirecting to ../../../0.5/xmldocs/llama.instructexecutor/... \ No newline at end of file diff --git a/latest/xmldocs/llama.interactiveexecutor/index.html b/latest/xmldocs/llama.interactiveexecutor/index.html index d9e400bb..2eeb347b 100644 --- a/latest/xmldocs/llama.interactiveexecutor/index.html +++ b/latest/xmldocs/llama.interactiveexecutor/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.interactiveexecutor/... + Redirecting to ../../../0.5/xmldocs/llama.interactiveexecutor/... \ No newline at end of file diff --git a/latest/xmldocs/llama.llamacontext/index.html b/latest/xmldocs/llama.llamacontext/index.html new file mode 100644 index 00000000..f74021c2 --- /dev/null +++ b/latest/xmldocs/llama.llamacontext/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.llamacontext/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.llamaembedder/index.html b/latest/xmldocs/llama.llamaembedder/index.html index f618c119..d143b82f 100644 --- a/latest/xmldocs/llama.llamaembedder/index.html +++ b/latest/xmldocs/llama.llamaembedder/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.llamaembedder/... + Redirecting to ../../../0.5/xmldocs/llama.llamaembedder/... \ No newline at end of file diff --git a/latest/xmldocs/llama.llamamodel/index.html b/latest/xmldocs/llama.llamamodel/index.html deleted file mode 100644 index 844268b6..00000000 --- a/latest/xmldocs/llama.llamamodel/index.html +++ /dev/null @@ -1,16 +0,0 @@ - - - - - Redirecting - - - - - Redirecting to ../../../0.4/xmldocs/llama.llamamodel/... - - \ No newline at end of file diff --git a/latest/xmldocs/llama.llamaquantizer/index.html b/latest/xmldocs/llama.llamaquantizer/index.html index 392c3b5b..a891bdf5 100644 --- a/latest/xmldocs/llama.llamaquantizer/index.html +++ b/latest/xmldocs/llama.llamaquantizer/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.llamaquantizer/... + Redirecting to ../../../0.5/xmldocs/llama.llamaquantizer/... \ No newline at end of file diff --git a/latest/xmldocs/llama.llamatransforms/index.html b/latest/xmldocs/llama.llamatransforms/index.html index 5aa6a051..93613cd5 100644 --- a/latest/xmldocs/llama.llamatransforms/index.html +++ b/latest/xmldocs/llama.llamatransforms/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.llamatransforms/... + Redirecting to ../../../0.5/xmldocs/llama.llamatransforms/... \ No newline at end of file diff --git a/latest/xmldocs/llama.llamaweights/index.html b/latest/xmldocs/llama.llamaweights/index.html new file mode 100644 index 00000000..7fdf9f55 --- /dev/null +++ b/latest/xmldocs/llama.llamaweights/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.llamaweights/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamacontextparams/index.html b/latest/xmldocs/llama.native.llamacontextparams/index.html index 3032c2fc..9fa67e3d 100644 --- a/latest/xmldocs/llama.native.llamacontextparams/index.html +++ b/latest/xmldocs/llama.native.llamacontextparams/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.llamacontextparams/... + Redirecting to ../../../0.5/xmldocs/llama.native.llamacontextparams/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamaftype/index.html b/latest/xmldocs/llama.native.llamaftype/index.html index b484613d..57487e25 100644 --- a/latest/xmldocs/llama.native.llamaftype/index.html +++ b/latest/xmldocs/llama.native.llamaftype/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.llamaftype/... + Redirecting to ../../../0.5/xmldocs/llama.native.llamaftype/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamagrammarelement/index.html b/latest/xmldocs/llama.native.llamagrammarelement/index.html new file mode 100644 index 00000000..4332c2db --- /dev/null +++ b/latest/xmldocs/llama.native.llamagrammarelement/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.llamagrammarelement/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamagrammarelementtype/index.html b/latest/xmldocs/llama.native.llamagrammarelementtype/index.html new file mode 100644 index 00000000..00a92732 --- /dev/null +++ b/latest/xmldocs/llama.native.llamagrammarelementtype/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.llamagrammarelementtype/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamamodelquantizeparams/index.html b/latest/xmldocs/llama.native.llamamodelquantizeparams/index.html new file mode 100644 index 00000000..83bbf339 --- /dev/null +++ b/latest/xmldocs/llama.native.llamamodelquantizeparams/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.llamamodelquantizeparams/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamatokendata/index.html b/latest/xmldocs/llama.native.llamatokendata/index.html index e66e3d8d..02fc8a93 100644 --- a/latest/xmldocs/llama.native.llamatokendata/index.html +++ b/latest/xmldocs/llama.native.llamatokendata/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.llamatokendata/... + Redirecting to ../../../0.5/xmldocs/llama.native.llamatokendata/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamatokendataarray/index.html b/latest/xmldocs/llama.native.llamatokendataarray/index.html index 7c131a2c..acbe488f 100644 --- a/latest/xmldocs/llama.native.llamatokendataarray/index.html +++ b/latest/xmldocs/llama.native.llamatokendataarray/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.llamatokendataarray/... + Redirecting to ../../../0.5/xmldocs/llama.native.llamatokendataarray/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.llamatokendataarraynative/index.html b/latest/xmldocs/llama.native.llamatokendataarraynative/index.html index dd41c244..26b15304 100644 --- a/latest/xmldocs/llama.native.llamatokendataarraynative/index.html +++ b/latest/xmldocs/llama.native.llamatokendataarraynative/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.llamatokendataarraynative/... + Redirecting to ../../../0.5/xmldocs/llama.native.llamatokendataarraynative/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.nativeapi/index.html b/latest/xmldocs/llama.native.nativeapi/index.html index 35ef5b27..2c2453c6 100644 --- a/latest/xmldocs/llama.native.nativeapi/index.html +++ b/latest/xmldocs/llama.native.nativeapi/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.nativeapi/... + Redirecting to ../../../0.5/xmldocs/llama.native.nativeapi/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.safellamacontexthandle/index.html b/latest/xmldocs/llama.native.safellamacontexthandle/index.html index 75a04ae7..12c2b155 100644 --- a/latest/xmldocs/llama.native.safellamacontexthandle/index.html +++ b/latest/xmldocs/llama.native.safellamacontexthandle/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.safellamacontexthandle/... + Redirecting to ../../../0.5/xmldocs/llama.native.safellamacontexthandle/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.safellamagrammarhandle/index.html b/latest/xmldocs/llama.native.safellamagrammarhandle/index.html new file mode 100644 index 00000000..586a7b51 --- /dev/null +++ b/latest/xmldocs/llama.native.safellamagrammarhandle/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.safellamagrammarhandle/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.safellamahandlebase/index.html b/latest/xmldocs/llama.native.safellamahandlebase/index.html index 83fbe5e1..4081d867 100644 --- a/latest/xmldocs/llama.native.safellamahandlebase/index.html +++ b/latest/xmldocs/llama.native.safellamahandlebase/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.native.safellamahandlebase/... + Redirecting to ../../../0.5/xmldocs/llama.native.safellamahandlebase/... \ No newline at end of file diff --git a/latest/xmldocs/llama.native.safellamamodelhandle/index.html b/latest/xmldocs/llama.native.safellamamodelhandle/index.html new file mode 100644 index 00000000..9cd07a49 --- /dev/null +++ b/latest/xmldocs/llama.native.safellamamodelhandle/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.safellamamodelhandle/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.native.samplingapi/index.html b/latest/xmldocs/llama.native.samplingapi/index.html new file mode 100644 index 00000000..56046aa3 --- /dev/null +++ b/latest/xmldocs/llama.native.samplingapi/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.native.samplingapi/... + + \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletion/index.html b/latest/xmldocs/llama.oldversion.chatcompletion/index.html index 12dfbd98..22b823c0 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletion/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletion/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletion/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletion/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletionchoice/index.html b/latest/xmldocs/llama.oldversion.chatcompletionchoice/index.html index 77879fee..1dc40340 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletionchoice/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletionchoice/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletionchoice/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletionchoice/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletionchunk/index.html b/latest/xmldocs/llama.oldversion.chatcompletionchunk/index.html index a5b3d818..ca86d4fd 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletionchunk/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletionchunk/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletionchunk/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletionchunk/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html b/latest/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html index d9becf4e..c901d862 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletionchunkchoice/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletionchunkchoice/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletionchunkchoice/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html b/latest/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html index 20f57850..83b7039a 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletionchunkdelta/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletionchunkdelta/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletionchunkdelta/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatcompletionmessage/index.html b/latest/xmldocs/llama.oldversion.chatcompletionmessage/index.html index 5f9e6954..8519a944 100644 --- a/latest/xmldocs/llama.oldversion.chatcompletionmessage/index.html +++ b/latest/xmldocs/llama.oldversion.chatcompletionmessage/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatcompletionmessage/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatcompletionmessage/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatmessagerecord/index.html b/latest/xmldocs/llama.oldversion.chatmessagerecord/index.html index b8fa3c97..bdb8302d 100644 --- a/latest/xmldocs/llama.oldversion.chatmessagerecord/index.html +++ b/latest/xmldocs/llama.oldversion.chatmessagerecord/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatmessagerecord/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatmessagerecord/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatrole/index.html b/latest/xmldocs/llama.oldversion.chatrole/index.html index 9f665af8..9bc8c7eb 100644 --- a/latest/xmldocs/llama.oldversion.chatrole/index.html +++ b/latest/xmldocs/llama.oldversion.chatrole/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatrole/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatrole/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.chatsession-1/index.html b/latest/xmldocs/llama.oldversion.chatsession-1/index.html index 8db5a660..3f093c4f 100644 --- a/latest/xmldocs/llama.oldversion.chatsession-1/index.html +++ b/latest/xmldocs/llama.oldversion.chatsession-1/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.chatsession-1/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.chatsession-1/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.completion/index.html b/latest/xmldocs/llama.oldversion.completion/index.html index 00a084ea..51e619be 100644 --- a/latest/xmldocs/llama.oldversion.completion/index.html +++ b/latest/xmldocs/llama.oldversion.completion/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.completion/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.completion/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.completionchoice/index.html b/latest/xmldocs/llama.oldversion.completionchoice/index.html index 83c0f9b4..082d3473 100644 --- a/latest/xmldocs/llama.oldversion.completionchoice/index.html +++ b/latest/xmldocs/llama.oldversion.completionchoice/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.completionchoice/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.completionchoice/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.completionchunk/index.html b/latest/xmldocs/llama.oldversion.completionchunk/index.html index fb1c4fdc..6bf3b815 100644 --- a/latest/xmldocs/llama.oldversion.completionchunk/index.html +++ b/latest/xmldocs/llama.oldversion.completionchunk/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.completionchunk/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.completionchunk/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.completionlogprobs/index.html b/latest/xmldocs/llama.oldversion.completionlogprobs/index.html index bb438fcb..e498410d 100644 --- a/latest/xmldocs/llama.oldversion.completionlogprobs/index.html +++ b/latest/xmldocs/llama.oldversion.completionlogprobs/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.completionlogprobs/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.completionlogprobs/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.completionusage/index.html b/latest/xmldocs/llama.oldversion.completionusage/index.html index ea1712ad..f63885dc 100644 --- a/latest/xmldocs/llama.oldversion.completionusage/index.html +++ b/latest/xmldocs/llama.oldversion.completionusage/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.completionusage/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.completionusage/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.embedding/index.html b/latest/xmldocs/llama.oldversion.embedding/index.html index 9f30f7ea..7114e9fb 100644 --- a/latest/xmldocs/llama.oldversion.embedding/index.html +++ b/latest/xmldocs/llama.oldversion.embedding/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.embedding/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.embedding/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.embeddingdata/index.html b/latest/xmldocs/llama.oldversion.embeddingdata/index.html index 35234485..799ec4c3 100644 --- a/latest/xmldocs/llama.oldversion.embeddingdata/index.html +++ b/latest/xmldocs/llama.oldversion.embeddingdata/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.embeddingdata/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.embeddingdata/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.embeddingusage/index.html b/latest/xmldocs/llama.oldversion.embeddingusage/index.html index b9f7e277..068174a0 100644 --- a/latest/xmldocs/llama.oldversion.embeddingusage/index.html +++ b/latest/xmldocs/llama.oldversion.embeddingusage/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.embeddingusage/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.embeddingusage/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.ichatmodel/index.html b/latest/xmldocs/llama.oldversion.ichatmodel/index.html index 49050774..d902db61 100644 --- a/latest/xmldocs/llama.oldversion.ichatmodel/index.html +++ b/latest/xmldocs/llama.oldversion.ichatmodel/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.ichatmodel/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.ichatmodel/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.llamaembedder/index.html b/latest/xmldocs/llama.oldversion.llamaembedder/index.html index 4b7a7d09..d15c1796 100644 --- a/latest/xmldocs/llama.oldversion.llamaembedder/index.html +++ b/latest/xmldocs/llama.oldversion.llamaembedder/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.llamaembedder/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.llamaembedder/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.llamamodel/index.html b/latest/xmldocs/llama.oldversion.llamamodel/index.html index c452d200..551752a1 100644 --- a/latest/xmldocs/llama.oldversion.llamamodel/index.html +++ b/latest/xmldocs/llama.oldversion.llamamodel/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.llamamodel/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.llamamodel/... \ No newline at end of file diff --git a/latest/xmldocs/llama.oldversion.llamaparams/index.html b/latest/xmldocs/llama.oldversion.llamaparams/index.html index fd3cd081..09ea966c 100644 --- a/latest/xmldocs/llama.oldversion.llamaparams/index.html +++ b/latest/xmldocs/llama.oldversion.llamaparams/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.oldversion.llamaparams/... + Redirecting to ../../../0.5/xmldocs/llama.oldversion.llamaparams/... \ No newline at end of file diff --git a/latest/xmldocs/llama.resettablellamamodel/index.html b/latest/xmldocs/llama.resettablellamamodel/index.html deleted file mode 100644 index b6dc6db1..00000000 --- a/latest/xmldocs/llama.resettablellamamodel/index.html +++ /dev/null @@ -1,16 +0,0 @@ - - - - - Redirecting - - - - - Redirecting to ../../../0.4/xmldocs/llama.resettablellamamodel/... - - \ No newline at end of file diff --git a/latest/xmldocs/llama.statefulexecutorbase/index.html b/latest/xmldocs/llama.statefulexecutorbase/index.html index 9f27b417..bcf978d4 100644 --- a/latest/xmldocs/llama.statefulexecutorbase/index.html +++ b/latest/xmldocs/llama.statefulexecutorbase/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.statefulexecutorbase/... + Redirecting to ../../../0.5/xmldocs/llama.statefulexecutorbase/... \ No newline at end of file diff --git a/latest/xmldocs/llama.statelessexecutor/index.html b/latest/xmldocs/llama.statelessexecutor/index.html index 7b33656b..52f52b89 100644 --- a/latest/xmldocs/llama.statelessexecutor/index.html +++ b/latest/xmldocs/llama.statelessexecutor/index.html @@ -4,13 +4,13 @@ Redirecting - Redirecting to ../../../0.4/xmldocs/llama.statelessexecutor/... + Redirecting to ../../../0.5/xmldocs/llama.statelessexecutor/... \ No newline at end of file diff --git a/latest/xmldocs/llama.utils/index.html b/latest/xmldocs/llama.utils/index.html new file mode 100644 index 00000000..a42a22ed --- /dev/null +++ b/latest/xmldocs/llama.utils/index.html @@ -0,0 +1,16 @@ + + + + + Redirecting + + + + + Redirecting to ../../../0.5/xmldocs/llama.utils/... + + \ No newline at end of file diff --git a/versions.json b/versions.json index dbe6fd2c..9bae761b 100644 --- a/versions.json +++ b/versions.json @@ -1 +1 @@ -[{"version": "0.4", "title": "0.4", "aliases": ["latest"]}] \ No newline at end of file +[{"version": "0.5", "title": "0.5", "aliases": ["latest"]}, {"version": "0.4", "title": "0.4", "aliases": []}] \ No newline at end of file