You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+15-14
Original file line number
Diff line number
Diff line change
@@ -24,9 +24,10 @@ Do checkout [TabbyAPI](https://github.com/theroyallab/tabbyAPI) if you want a re
24
24
25
25
# NEW - Vision Model
26
26
27
-
From `gallama` version 0.0.7, there is experimental support for Vision model.
27
+
As of v0.0.8post1, Qwen 2 VL (Image only, no Video) and Pixtral are supported via Exllama (>=0.2.4).
28
28
29
-
Currently, as of v0.0.8, Pixtral is supported via Exllama (>=0.2.4) and Qwen 2 VL series of model is supported via transformers.
29
+
For Pixtral, please install Exllama V2 `v0.2.4` onwards
30
+
For Exllama V2, please install `dev` branch of Exllama V2 as the code is not yet merged to `v0.2.4`.
30
31
31
32
After Exllama roll out support for Qwen 2 VL, running model via transformers will be depreciated.
32
33
Currently, both exllamaV2 and llama.cpp do not support Vision model yet. Hence, this is achieved by running `transformers` with the use of awq for quantization.
@@ -49,16 +50,16 @@ This is already be handled in the requirements.txt, however, getting transformer
49
50
After installation you can download by following commands (choose a version that fit your VRAM):
0 commit comments