Browse Source

- Added `LoadFromFileAsync` method for `LLavaWeights`

- Fixed checking for invalid handles in `clip_model_load`
pull/703/head
Martin Evans 1 year ago
parent
commit
377ebf3664
3 changed files with 23 additions and 7 deletions
  1. +2
    -2
      LLama.Examples/Examples/LlavaInteractiveModeExecute.cs
  2. +16
    -3
      LLama/LLavaWeights.cs
  3. +5
    -2
      LLama/Native/SafeLlavaModelHandle.cs

+ 2
- 2
LLama.Examples/Examples/LlavaInteractiveModeExecute.cs View File

@@ -24,9 +24,9 @@ namespace LLama.Examples.Examples
using var context = model.CreateContext(parameters); using var context = model.CreateContext(parameters);
// Llava Init // Llava Init
using var clipModel = LLavaWeights.LoadFromFile(multiModalProj);
using var clipModel = await LLavaWeights.LoadFromFileAsync(multiModalProj);
var ex = new InteractiveExecutor(context, clipModel );
var ex = new InteractiveExecutor(context, clipModel);


Console.ForegroundColor = ConsoleColor.Yellow; Console.ForegroundColor = ConsoleColor.Yellow;
Console.WriteLine("The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to {0} and the context size is {1}.", maxTokens, parameters.ContextSize ); Console.WriteLine("The executor has been enabled. In this example, the prompt is printed, the maximum tokens is set to {0} and the context size is {1}.", maxTokens, parameters.ContextSize );


+ 16
- 3
LLama/LLavaWeights.cs View File

@@ -1,5 +1,7 @@


using System; using System;
using System.Threading;
using System.Threading.Tasks;
using LLama.Native; using LLama.Native;


namespace LLama; namespace LLama;
@@ -13,9 +15,9 @@ public sealed class LLavaWeights : IDisposable
/// The native handle, which is used in the native APIs /// The native handle, which is used in the native APIs
/// </summary> /// </summary>
/// <remarks>Be careful how you use this!</remarks> /// <remarks>Be careful how you use this!</remarks>
public SafeLlavaModelHandle NativeHandle { get; }
internal LLavaWeights(SafeLlavaModelHandle weights)
public SafeLlavaModelHandle NativeHandle { get; }
private LLavaWeights(SafeLlavaModelHandle weights)
{ {
NativeHandle = weights; NativeHandle = weights;
} }
@@ -31,6 +33,17 @@ public sealed class LLavaWeights : IDisposable
return new LLavaWeights(weights); return new LLavaWeights(weights);
} }


/// <summary>
/// Load weights into memory
/// </summary>
/// <param name="mmProject">path to the "mmproj" model file</param>
/// <param name="token"></param>
/// <returns></returns>
public static Task<LLavaWeights> LoadFromFileAsync(string mmProject, CancellationToken token = default)
{
return Task.Run(() => LoadFromFile(mmProject), token);
}

/// <summary> /// <summary>
/// Create the Image Embeddings from the bytes of an image. /// Create the Image Embeddings from the bytes of an image.
/// </summary> /// </summary>


+ 5
- 2
LLama/Native/SafeLlavaModelHandle.cs View File

@@ -39,8 +39,11 @@ namespace LLama.Native
if (!fs.CanRead) if (!fs.CanRead)
throw new InvalidOperationException($"Llava MMP Model file '{modelPath}' is not readable"); throw new InvalidOperationException($"Llava MMP Model file '{modelPath}' is not readable");
return clip_model_load(modelPath, verbosity)
?? throw new LoadWeightsFailedException(modelPath);
var handle = clip_model_load(modelPath, verbosity);
if (handle.IsInvalid)
throw new LoadWeightsFailedException(modelPath);

return handle;
} }


/// <summary> /// <summary>


Loading…
Cancel
Save