Skip to main content

Module quantize

Module quantize 

Source
Expand description

QuantizedBrick Implementation (PMAT-013)

Implements quantized weight support for ComputeBricks per cbtop spec S17.

§Supported Formats

FormatBits/WeightMemoryPerplexity Delta
Q4_04.025%~0.5%
Q4_K4.528%~0.3%
Q5_K5.534%~0.1%
Q8_08.050%~0.01%

§Citations

  • [Dettmers et al. 2022] “LLM.int8(): 8-bit Matrix Multiplication” NeurIPS
  • [Frantar et al. 2023] “GPTQ: Accurate Post-Training Quantization” ICLR
  • [Lin et al. 2023] “AWQ: Activation-aware Weight Quantization” MLSys

Structs§

GgufHeader
GGUF file header (simplified parsing).
GgufLoader
GGUF file loader (basic implementation).
GgufTensorInfo
GGUF tensor info.
LayerQuantStats
Per-layer quantization statistics.
QuantStats
Quantization statistics for a model or layer.
QuantizedBrick
QuantizedBrick wraps compute operations with quantized weights.
QuantizedWeights
Quantized weight storage for a single layer.

Enums§

DequantStrategy
Dequantization strategy.
GgufError
GGUF parsing errors.
GgufValue
GGUF metadata value types.
QuantFormat
Supported quantization formats for ComputeBricks.

Functions§

ggml_type_to_format
GGML tensor type to QuantFormat mapping.

Type Aliases§

GgufResult
Result type for GGUF operations.