bitorch_engine.utils.quant_operators.gptq_style_unpacking
- bitorch_engine.utils.quant_operators.gptq_style_unpacking(qweight) Tuple[Tensor, Tensor] [source]
Reconstructs the fp16 weight tensor from the input quantized weight parameter in GPTQ style.
- Parameters:
qweight – The quantized weight parameter object containing all necessary quantization information.
- Returns:
The reconstructed weight tensor in fp16 format.
- Return type:
torch.Tensor