Help regarding loading custom trained models #81
Replies: 1 comment
-
For example: # make sure that you have git lfs installed
git lfs install
git clone https://huggingface.co/BAAI/bge-base-en-v1.5
# mount the model directory to /model inside the container and pass it to the `--model-id` CLI arg
docker run --gpus all -p 8080:80 -v $PWD/bge-base-en-v1.5:/model --pull always ghcr.io/huggingface/text-embeddings-inference:0.4.0 --model-id /model |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I need documentation/examples for loading models using the path of the local directory. It is looking for a URL when passed through --model-id, is there any additional tag to pass as an argument?
Beta Was this translation helpful? Give feedback.
All reactions