diff --git a/README.md b/README.md index 3ab1fa0c..78fc55e9 100644 --- a/README.md +++ b/README.md @@ -48,9 +48,12 @@ Here's the mapping of them and corresponding model samples provided by `LLamaSha | - | v0.2.1 | [WizardLM](https://huggingface.co/TheBloke/wizardLM-7B-GGML/tree/previous_llama), [Vicuna (filenames with "old")](https://huggingface.co/eachadea/ggml-vicuna-13b-1.1/tree/main) | - | | v0.2.2 | v0.2.2, v0.2.3 | [WizardLM](https://huggingface.co/TheBloke/wizardLM-7B-GGML/tree/previous_llama_ggmlv2), [Vicuna (filenames without "old")](https://huggingface.co/eachadea/ggml-vicuna-13b-1.1/tree/main) | 63d2046 | | v0.3.0, v0.3.1 | v0.3.0, v0.4.0 | [LLamaSharpSamples v0.3.0](https://huggingface.co/AsakusaRinne/LLamaSharpSamples/tree/v0.3.0), [WizardLM](https://huggingface.co/TheBloke/wizardLM-7B-GGML/tree/main) | 7e4ea5b | +| v0.4.1 | v0.4.1 | [Open llama 3b](https://huggingface.co/SlyEcho/open_llama_3b_ggml), [Open Buddy](https://huggingface.co/OpenBuddy/openbuddy-llama-ggml)| aacdbd4 | + +Many hands make light work. If you have found any other model resource that could work for a version, we'll appreciate it for opening an PR about it! 😊 We publish the backend with cpu, cuda11 and cuda12 because they are the most popular ones. If none of them matches, please compile the [llama.cpp](https://github.com/ggerganov/llama.cpp) -from source and put the `libllama` under your project's output path. When building from source, please add `-DBUILD_SHARED_LIBS=ON` to enable the library generation. +from source and put the `libllama` under your project's output path ([guide](https://scisharp.github.io/LLamaSharp/0.4/ContributingGuide/)). ## FAQ @@ -148,6 +151,10 @@ Since we are in short of hands, if you're familiar with ASP.NET core, we'll appr ⚠️ Semantic-kernel Integration +🔳 Fine-tune + +🔳 Local document search + 🔳 MAUI Integration 🔳 Follow up llama.cpp and improve performance