You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
My graphics card was RTX3060 12G, the model used was Qwen2.5-7B-instruct-Q4_k_M, normally the model should only take up 4~5G VRam, so I thought the VRam of my GPU was sufficient to handle the quantization model, but I found that my Ram was occupied all the time. The amount of Ram used by each application in windows Task Manager is inconsistent with the actual total Ram usage in Windows Task Manager, and it will not be released until I finish the python script, is the Ram usage necessary or is it just a BUG?
The text was updated successfully, but these errors were encountered:
My graphics card was RTX3060 12G, the model used was Qwen2.5-7B-instruct-Q4_k_M, normally the model should only take up 4~5G VRam, so I thought the VRam of my GPU was sufficient to handle the quantization model, but I found that my Ram was occupied all the time. The amount of Ram used by each application in windows Task Manager is inconsistent with the actual total Ram usage in Windows Task Manager, and it will not be released until I finish the python script, is the Ram usage necessary or is it just a BUG?
The text was updated successfully, but these errors were encountered: