Skip to content

Commit 8b9f795

Browse files
committed
change default
Signed-off-by: Kyle Sayers <[email protected]>
1 parent a9c7ddc commit 8b9f795

File tree

1 file changed

+3
-2
lines changed
  • src/llmcompressor/modifiers/quantization/gptq

1 file changed

+3
-2
lines changed

src/llmcompressor/modifiers/quantization/gptq/base.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -76,8 +76,9 @@ class GPTQModifier(Modifier, QuantizationMixin):
7676
:param block_size: Used to determine number of columns to compress in one pass
7777
:param dampening_frac: Amount of dampening to apply to H, as a fraction of the
7878
diagonal norm
79-
:param actorder: order in which weight columns are quantized. For more information,
80-
on actorder options, see https://github.com/vllm-project/vllm/pull/8135
79+
:param actorder: order in which weight columns are quantized. Defaults to "static"
80+
activation ordering, which achieves best accuracy recovery with no runtime cost.
81+
For more information, see https://github.com/vllm-project/vllm/pull/8135
8182
:param offload_hessians: Set to True for decreased memory usage but increased
8283
runtime.
8384

0 commit comments

Comments
 (0)