You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

LLamaVocabType.cs 555 B

April 2024 Binary Update (#662) * Updated binaries, using [this build](https://github.com/SciSharp/LLamaSharp/actions/runs/8654672719/job/23733195669) for llama.cpp commit `f7001ccc5aa359fcf41bba19d1c99c3d25c9bcc7`. - Added all new functions. - Moved some functions (e.g. `SafeLlamaModelHandle` specific functions) into `SafeLlamaModelHandle.cs` - Exposed tokens on `SafeLlamaModelHandle` and `LLamaWeights` through a `Tokens` property. As new special tokens are added in the future they can be added here. - Changed all token properties to return nullable tokens, to handle some models not having some tokens. - Fixed `DefaultSamplingPipeline` to handle no newline token in some models. * Moved native methods to more specific locations. - Context specific things have been moved into `SafeLLamaContextHandle.cs` and made private - they're exposed through C# properties and methods already. - Checking that GPU layer count is zero if GPU offload is not supported. - Moved methods for creating default structs (`llama_model_quantize_default_params` and `llama_context_default_params`) into relevant structs. * Removed exception if `GpuLayerCount > 0` when GPU is not supported. * - Added low level wrapper methods for new per-sequence state load/save in `SafeLLamaContextHandle` - Added high level wrapper methods (save/load with `State` object or memory mapped file) in `LLamaContext` - Moved native methods for per-sequence state load/save into `SafeLLamaContextHandle` * Added update and defrag methods for KV cache in `SafeLLamaContextHandle` * Updated submodule to `f7001ccc5aa359fcf41bba19d1c99c3d25c9bcc7` * Passing the sequence ID when saving a single sequence state
1 year ago
12345678910111213141516171819202122232425262728
  1. namespace LLama.Native;
  2. /// <summary>
  3. ///
  4. /// </summary>
  5. /// <remarks>llama_vocab_type</remarks>
  6. public enum LLamaVocabType
  7. {
  8. /// <summary>
  9. /// For models without vocab
  10. /// </summary>
  11. None = 0,
  12. /// <summary>
  13. /// LLaMA tokenizer based on byte-level BPE with byte fallback
  14. /// </summary>
  15. SentencePiece = 1,
  16. /// <summary>
  17. /// GPT-2 tokenizer based on byte-level BPE
  18. /// </summary>
  19. BytePairEncoding = 2,
  20. /// <summary>
  21. /// BERT tokenizer based on WordPiece
  22. /// </summary>
  23. WordPiece = 3,
  24. }