bitorch_engine.utils.quant_operators.gptq_style_unpacking

bitorch_engine.utils.quant_operators.gptq_style_unpacking(qweight) Tuple[Tensor, Tensor][source]

Reconstructs the fp16 weight tensor from the input quantized weight parameter in GPTQ style.

Parameters:

qweight – The quantized weight parameter object containing all necessary quantization information.

Returns:

The reconstructed weight tensor in fp16 format.

Return type:

torch.Tensor