Hello. Is it possible to quantize to AWQ so that this model is compatible with lower GPUs on vLLM?
· Sign up or log in to comment