You can not select more than 25 topics Topics must start with a chinese character,a letter or number, can include dashes ('-') and can be up to 35 characters long.

Constants.cs 1.2 kB

April 2024 Binary Update (#662) * Updated binaries, using [this build](https://github.com/SciSharp/LLamaSharp/actions/runs/8654672719/job/23733195669) for llama.cpp commit `f7001ccc5aa359fcf41bba19d1c99c3d25c9bcc7`. - Added all new functions. - Moved some functions (e.g. `SafeLlamaModelHandle` specific functions) into `SafeLlamaModelHandle.cs` - Exposed tokens on `SafeLlamaModelHandle` and `LLamaWeights` through a `Tokens` property. As new special tokens are added in the future they can be added here. - Changed all token properties to return nullable tokens, to handle some models not having some tokens. - Fixed `DefaultSamplingPipeline` to handle no newline token in some models. * Moved native methods to more specific locations. - Context specific things have been moved into `SafeLLamaContextHandle.cs` and made private - they're exposed through C# properties and methods already. - Checking that GPU layer count is zero if GPU offload is not supported. - Moved methods for creating default structs (`llama_model_quantize_default_params` and `llama_context_default_params`) into relevant structs. * Removed exception if `GpuLayerCount > 0` when GPU is not supported. * - Added low level wrapper methods for new per-sequence state load/save in `SafeLLamaContextHandle` - Added high level wrapper methods (save/load with `State` object or memory mapped file) in `LLamaContext` - Moved native methods for per-sequence state load/save into `SafeLLamaContextHandle` * Added update and defrag methods for KV cache in `SafeLLamaContextHandle` * Updated submodule to `f7001ccc5aa359fcf41bba19d1c99c3d25c9bcc7` * Passing the sequence ID when saving a single sequence state
1 year ago
12345678910111213141516171819202122232425262728293031323334
  1. using System.Runtime.InteropServices;
  2. namespace LLama.Unittest
  3. {
  4. internal static class Constants
  5. {
  6. public static readonly string GenerativeModelPath = "Models/llama-2-7b-chat.Q3_K_S.gguf";
  7. public static readonly string EmbeddingModelPath = "Models/all-MiniLM-L12-v2.Q8_0.gguf";
  8. public static readonly string LLavaModelPath = "Models/llava-v1.6-mistral-7b.Q3_K_XS.gguf";
  9. public static readonly string LLavaMmpPath = "Models/mmproj-model-f16.gguf";
  10. public static readonly string LLavaImage = "Models/extreme-ironing-taxi-610x427.jpg";
  11. /// <summary>
  12. /// Calculate GpuLayer Count to use in UnitTest
  13. /// </summary>
  14. /// <returns> Defaults to 20 in all the cases, except MacOS/OSX release (to disable METAL on github CI)</returns>
  15. public static int CIGpuLayerCount
  16. {
  17. get
  18. {
  19. if (RuntimeInformation.IsOSPlatform(OSPlatform.OSX))
  20. {
  21. #if DEBUG
  22. return 20;
  23. #else
  24. return 0;
  25. #endif
  26. }
  27. else return 20;
  28. }
  29. }
  30. }
  31. }