Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GPU support #128

Closed
draco1544 opened this issue Apr 3, 2023 · 7 comments · May be fixed by #944
Closed

GPU support #128

draco1544 opened this issue Apr 3, 2023 · 7 comments · May be fixed by #944

Comments

@draco1544
Copy link

It’s possible to make the webui and the container work with a version of llama that use gpu ?

@voarsh2
Copy link

voarsh2 commented Apr 4, 2023

Here's hoping. Consumers would only be able to use 8/12 GB's of VRAM though.

@gaby
Copy link
Member

gaby commented Apr 12, 2023

This would be something to ask the llama.cpp team, until they don't add support for GPU's we can't add it to Serge.

@gaby gaby closed this as completed Apr 12, 2023
@edricus
Copy link

edricus commented Jun 13, 2023

This would be something to ask the llama.cpp team, until they don't add support for GPU's we can't add it to Serge.

It seems it's now officially supported ggml-org/llama.cpp#1827

@gaby
Copy link
Member

gaby commented Jun 13, 2023

This would be something to ask the llama.cpp team, until they don't add support for GPU's we can't add it to Serge.

It seems it's now officially supported ggerganov/llama.cpp#1827

That's a draft, it's not merged/released yet

@roperscrossroads
Copy link

@gaby

I'm not familiar with this project but I saw this issue mentioned in the llama.cpp pull request. llama.cpp already has cuBLAS and CLBlast support in the master branch. The PR mentioned above improves performance for GPUs with CUDA.

Check out the docs where they mention building with LLAMA_OPENBLAS and LLAMA_CUBLAS enabled for more information.

@rakuri255
Copy link

Seems this issue should be open

@gaby
Copy link
Member

gaby commented Dec 4, 2023

Duplicated of #43

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging a pull request may close this issue.

6 participants