You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In this test, I ran Llama 3.1 8B and Llama 3.3 70B Instruct Q40 on the same machine with 2 x NVIDIA GeForce RTX 3090 GPUs. I performed two tests: one using a single GPU, and another using both GPUs with full tensor parallelism on a single machine.
Because Distributed Llama still has some issues with Vulkan warmup, I took five measurements from within the inference process to calculate the inference speed.
1 x GeForce RTX 3090 24 GB
2 x GeForce RTX 3090 24 GB
Llama 3.1 8B Q40 - prediction
17.8 tok / s
25.5 tok / s
Llama 3.3 70B Instruct Q40 - prediction
not enough memory
2.6 tok / s
Distributed Llama was built with Vulkan support. How I runned on 2 GPUs?
🔶 Pred 56 ms Sync 0 ms | Sent 0 kB Recv 0 kB | The
🔶 Pred 56 ms Sync 0 ms | Sent 0 kB Recv 0 kB | tensor
🔶 Pred 56 ms Sync 0 ms | Sent 0 kB Recv 0 kB | parallel
🔶 Pred 56 ms Sync 0 ms | Sent 0 kB Recv 0 kB | ism
🔶 Pred 56 ms Sync 0 ms | Sent 0 kB Recv 0 kB | is
2 GPUs
Llama 3.1 8B Instruct Q40
🔶 Pred 38 ms Sync 4 ms | Sent 288 kB Recv 522 kB | However
🔶 Pred 40 ms Sync 3 ms | Sent 288 kB Recv 522 kB | ,
🔶 Pred 39 ms Sync 3 ms | Sent 288 kB Recv 522 kB | if
🔶 Pred 39 ms Sync 4 ms | Sent 288 kB Recv 522 kB | the
🔶 Pred 40 ms Sync 4 ms | Sent 288 kB Recv 522 kB | processor
Llama 3.3 70B Instruct Q40
🔶 Pred 385 ms Sync 29 ms | Sent 1392 kB Recv 1610 kB | Model
🔶 Pred 385 ms Sync 16 ms | Sent 1392 kB Recv 1610 kB | parallel
🔶 Pred 383 ms Sync 11 ms | Sent 1392 kB Recv 1610 kB | ism
🔶 Pred 389 ms Sync 8 ms | Sent 1392 kB Recv 1610 kB | refers
🔶 Pred 388 ms Sync 9 ms | Sent 1392 kB Recv 1610 kB | to
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Note
This result was improved a lot in the higher version.
In this test, I ran Llama 3.1 8B and Llama 3.3 70B Instruct Q40 on the same machine with 2 x NVIDIA GeForce RTX 3090 GPUs. I performed two tests: one using a single GPU, and another using both GPUs with full tensor parallelism on a single machine.
Because Distributed Llama still has some issues with Vulkan warmup, I took five measurements from within the inference process to calculate the inference speed.
Distributed Llama was built with Vulkan support. How I runned on 2 GPUs?
1 GPU
Llama 3.1 8B Instruct Q40
2 GPUs
Llama 3.1 8B Instruct Q40
Llama 3.3 70B Instruct Q40
Spec
Beta Was this translation helpful? Give feedback.
All reactions