Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does AirLLM Support Running Quantized Models (e.g., unsloth/Qwen2-72B-bnb-4bit)? #213

Open
NEWbie0709 opened this issue Feb 24, 2025 · 1 comment

Comments

@NEWbie0709
Copy link

NEWbie0709 commented Feb 24, 2025

Does AirLLM currently support running 4-bit quantized models like unsloth/Qwen2-72B-bnb-4bit? I’m trying to load and run this model using AirLLM, but I’m encountering the following error during generation:

RuntimeError: Attempted to call variable.set_data(tensor), but variable and tensor have incompatible tensor type.

Other than that, I also tried using the smaller version of Qwen, for example: Qwen/Qwen2.5-0.5B, but I encountered this error.

AssertionError: model.safetensors.index.json should exist

@NEWbie0709
Copy link
Author

i tried running with Qwen-72B-instruct and this is the error i got

Image

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant