You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hello, I'm trying to use AIMET_TORCH to quantize a LLM model, e.g.: llama v2。 where can I find a jupyter NB example which shows quantization simulation for a LLM model?
The text was updated successfully, but these errors were encountered:
I tried to quantize llama1b, but the quantized model output random charactors...
I kind of wondering the different between aimet pro and aimet public version. I see that the llama3b is quantized pretty good with aimet pro by the qualcomm team. But there is no example released of how they do that.
Hello, I'm trying to use AIMET_TORCH to quantize a LLM model, e.g.: llama v2。 where can I find a jupyter NB example which shows quantization simulation for a LLM model?
The text was updated successfully, but these errors were encountered: