-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Embedding unreachable, but llama is running #4112
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
updated to 0.27.1, tried different models thanks to more ram, the local embedding is still marked as 'unreachable', same errors |
workaround: use http for embeddings, not I literally copied the
I think the |
hi @LucaFulchir, Did you utilize the llama.cpp that came with Tabby, or was it installed manually as a separate component? |
tabby is configured to use nixos llama.cpp, built to use vulkan. Now I notice that when I run llama manually instead it uses realase |
The The for more detail, you could check: https://tabby.tabbyml.com/docs/references/models-http-api/llama.cpp/ |
Uh oh!
There was an error while loading. Please reload this page.
Hello, I'm trying to run/test tabby, but I have problems with the embedding instance
Using version 0.27, NixOS unstable server.
Ai completion and Ai chat seem to work, but I can not add a
git
context provider of a public repo, it seems to clone successfully but can't parse a single file.config.toml:
running with:
testing on
AMD Ryzen 7 8845HS w/ Radeon 780M Graphics
on the tabby web interface, on the systems page I see "Unreachable" only under "Enbedding", with error "error decoding response body"
The llama instance seems to be UP and by dumping the local traffic I see the following req/responses:
Additional tabby logging even when running with
RUST_LOG=debug
are all like:what can I try/what I am doing wrong?
The text was updated successfully, but these errors were encountered: