Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

data_process中,使用load_4bit加载liuhaotian/llava-v1.6-34b模型报错 #408

Closed
sunfan1997 opened this issue May 17, 2024 · 2 comments
Labels

Comments

@sunfan1997
Copy link

显存只有16G,想处理一批自己的数据,如何量化加载liuhaotian/llava-v1.6-34b模型,报错ValueError: Calling cuda() is not supported for 4-bit or 8-bit quantized models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct dtype.有什么办法减少显存占用吗,或者说最低需要多少显存才能跑data process部分。

Copy link

This issue is stale because it has been open for 7 days with no activity.

@github-actions github-actions bot added the stale label May 25, 2024
Copy link

github-actions bot commented Jun 1, 2024

This issue was closed because it has been inactive for 7 days since being marked as stale.

@github-actions github-actions bot closed this as not planned Won't fix, can't repro, duplicate, stale Jun 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant