You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi
As I tried with 13b version in TGI, it works fine with bitsandbytes quantization.
While trying with AWQ quantization in TGI, it shows error as "Cannot load 'awq' weight, make sure the model is already quantized"
I am wondering if AWQ is too new to this model while deploying by TGI
Or there is any suggestion or comment?
Thanks
The text was updated successfully, but these errors were encountered:
nigue3025
changed the title
Support of AWQ quantization in TGI
Support for AWQ quantization in TGI
May 14, 2024
Hi
As I tried with 13b version in TGI, it works fine with bitsandbytes quantization.
While trying with AWQ quantization in TGI, it shows error as "Cannot load 'awq' weight, make sure the model is already quantized"
I am wondering if AWQ is too new to this model while deploying by TGI
Or there is any suggestion or comment?
Thanks
The text was updated successfully, but these errors were encountered: