| @@ -27,13 +27,19 @@ namespace LLama.Native | |||||
| /// <summary> | /// <summary> | ||||
| /// how to split layers across multiple GPUs (size: <see cref="NativeApi.llama_max_devices"/>) | /// how to split layers across multiple GPUs (size: <see cref="NativeApi.llama_max_devices"/>) | ||||
| /// </summary> | /// </summary> | ||||
| public float* tensor_split; | |||||
| public float* tensor_split; | |||||
| /// <summary> | /// <summary> | ||||
| /// called with a progress value between 0 and 1, pass NULL to disable. If the provided progress_callback | /// called with a progress value between 0 and 1, pass NULL to disable. If the provided progress_callback | ||||
| /// returns true, model loading continues. If it returns false, model loading is immediately aborted. | /// returns true, model loading continues. If it returns false, model loading is immediately aborted. | ||||
| /// </summary> | /// </summary> | ||||
| #if NETSTANDARD2_0 | |||||
| // this code is intended to be used when running LlamaSharp on NET Framework 4.8 (NET Standard 2.0) | |||||
| // as NET Framework 4.8 does not play nice with the LlamaProgressCallback type | |||||
| public IntPtr progress_callback; | |||||
| #else | |||||
| public LlamaProgressCallback progress_callback; | public LlamaProgressCallback progress_callback; | ||||
| #endif | |||||
| /// <summary> | /// <summary> | ||||
| /// context pointer passed to the progress callback | /// context pointer passed to the progress callback | ||||