From 2dbad9ae5fd135a5d1b9b70865ca0847eb549444 Mon Sep 17 00:00:00 2001 From: Nick Fraser Date: Fri, 1 Nov 2024 11:39:18 +0000 Subject: [PATCH] Fix (examples/generative): Fixed argument order for `quantize_model` --- .../common/generative/quantize.py | 42 +++++++++---------- 1 file changed, 21 insertions(+), 21 deletions(-) diff --git a/src/brevitas_examples/common/generative/quantize.py b/src/brevitas_examples/common/generative/quantize.py index eaabf4d81..f3a1c4add 100644 --- a/src/brevitas_examples/common/generative/quantize.py +++ b/src/brevitas_examples/common/generative/quantize.py @@ -478,27 +478,27 @@ def quantize_model( input_kwargs=None): linear_input_quant, weight_quant, input_quant, q_scaled_quant, k_transposed_quant, v_quant, attn_output_weights_quant = generate_quantizers( - dtype, - weight_bit_width, - weight_param_method, - weight_scale_precision, - weight_quant_type, - weight_quant_granularity, - weight_group_size, - quantize_weight_zero_point, - weight_quant_format, - input_bit_width, - input_quant_format, - input_scale_precision, - input_scale_type, - input_param_method, - input_quant_type, - input_quant_granularity, - input_group_size, - quantize_input_zero_point, - device, - weight_kwargs, - input_kwargs) + dtype=dtype, + weight_bit_width=weight_bit_width, + weight_param_method=weight_param_method, + weight_scale_precision=weight_scale_precision, + weight_quant_type=weight_quant_type, + weight_quant_granularity=weight_quant_granularity, + weight_group_size=weight_group_size, + quantize_weight_zero_point=quantize_weight_zero_point, + weight_quant_format=weight_quant_format, + input_bit_width=input_bit_width, + input_quant_format=input_quant_format, + input_scale_precision=input_scale_precision, + input_scale_type=input_scale_type, + input_param_method=input_param_method, + input_quant_type=input_quant_type, + input_quant_granularity=input_quant_granularity, + input_group_size=input_group_size, + quantize_input_zero_point=quantize_input_zero_point, + device=device, + weight_kwargs=weight_kwargs, + input_kwargs=input_kwargs) layer_map = generate_quant_maps( linear_input_quant, weight_quant,