Replies: 2 comments 1 reply
-
Great idea! Hoping to contribute where I can. |
Beta Was this translation helpful? Give feedback.
1 reply
-
The GPT4All desktop app provides a local OpenAI-compatible server with LocalDocs support, and there is also a Python binding for our backend although it hasn't been updated in a while. We are unfortunately stuck on an old version of llama.cpp because we rely on many patches for the Kompute backend, which we have not had time to rebase on the new backend API. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Proposing a living doc about all the frameworks that work with (or should work with)
llama.cpp
, at any level.The list is long so let's keep it roughly sorted by decreasing community contributions or stars or something ✌️ (direct edits from contributors / suggestions of edits in comments highly welcome, I've probably made a gazillion mistakes and omissions already!)
Part of the goal is to identify which projects would benefit from a documentation update or small patches for direct support. For instance, a few Python projects only document the (amazing)
llama-cpp-python
bindings and could use instructions on how to also use llama-server (our canonical OpenAI-compatible server)Projects with some integration (non exhaustive list!)
llama.cpp
contributors: @ggerganov, @slaren, @JohannesGaessler, @ngxson & too many to count / large overlap 🤗🤗🤗🤗🤗🤗🤗libllama
?llama.cpp
contributors: @cebtenzzre (PRs) 🤗llama-server
(see Hugging Face Inference Endpoints now supports GGUF out of the box! #9669, revshare goes to ggml.ai)llama.cpp
contributors: @ngxson (PRs) 🤗llama.cpp
/ modified serverllama.cpp
contributors: @jart (PRs) 🤗libllama
+ includes llama.cpp's JSON schema conversionllama.cpp
contributors: @abetlen (PRs) 🤗libllama
?llama.cpp
contributors: @mudler (PRs) 🤗llama-server
can only partially use Ollama models (custom incompatible chat template format)ggml
libllama
(?)llama-server
(OpenAI
+base_url
override: example)llama-server
(OpenAI like integration w/api_base
): needs docsLlamaCPP
)llama-server
(using OpenAI +base_uri
override): needs docsllama-server
(OpenAI-compatible endpoint doc): needs docsllama-server
: not yet / needs help ❌llama-server
: not yet / needs help ❌base_url
override?llama-server
: not yet / needs help ❌llama-cpp-python
(labelledllama.cpp
)llama-server
(OpenAI Chat Model Node + overrideBase URL
/ API key): needs docsllama-server
: not yet / needs help ❌base_url
paramllama-server
orOllama
(use OpenAI + overridebase_url
&api_key
): needs docsllama-server
: not yet / needs help ❌base_uri
overridellama-server
(use OpenAI + overridebase_uri
): needs docsllama-server
: not yet / needs help ❌base_url
overridellama-server
✅llama-cpp-python
llama-server
: not yet / needs help ❌llama-server
powered)llama-cpp
(using LLMEndpointConfig +llm_base_url
)Projects w/o integration
Beta Was this translation helpful? Give feedback.
All reactions