Open
Description
Describe the issue
I am trying to enable AWQ support with IPEX repo in CPU.
IPEX 2.5.0 release states that it has the support for AWQ Quantization.
But we could see only the GPTQ support added in the official repo.
In the below script file,
https://github.com/intel/intel-extension-for-pytorch/blob/release/xpu/2.5.10/examples/cpu/llm/inference/utils/run_gptq.py stated that it is deprecated and recommended to use INC.
What is the correct approach that we need to use to enable the support for AWQ with IPEX repo?
Config used:
- Python - 3.9
- IPEX - 2.5.0
- Build type: release
- Torch - 2.5.0
- Transformers - 4.43.2