You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: README.md
+2-3
Original file line number
Diff line number
Diff line change
@@ -44,10 +44,9 @@ gallama run -id "model_id=pixtral max_seq_len=32768"
44
44
```
45
45
46
46
2. Qwen 2 VL:
47
-
As of this release, the transformers build in pip is not yet updated with bugfix for Qwen 2 VL, hence you will need to install the latest code from github.
48
-
This is already be handled in the requirements.txt, however, getting transformers dependency working can be tricky.
47
+
For Exllama V2, please install `dev` branch of Exllama V2 as the code is not yet merged to `v0.2.4`.
49
48
50
-
After installation you can download by following commands (choose a version that fit your VRAM):
49
+
After installation you can download by following commands (choose a version that fit your VRAM), there are other quantization available from 3 to 8bpw.
0 commit comments