Skip to content

Commit c90ab60

Browse files
authored
[ACT] Rename 'Quantization' to 'QuantAware'. (#1536)
* [ACT] Rename 'Quantization' to 'QuantAware'. * Fix examples. * Fix tests.
1 parent dfe5bb3 commit c90ab60

File tree

43 files changed

+54
-53
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

43 files changed

+54
-53
lines changed

example/auto_compression/README.md

+1-1
Original file line numberDiff line numberDiff line change
@@ -178,7 +178,7 @@ ac = AutoCompression(
178178
params_filename="inference.pdiparams",
179179
save_dir="MobileNetV1_quant",
180180
config={"QuantPost": {}, "HyperParameterOptimization": {'ptq_algo': ['avg'], 'max_quant_count': 3}},
181-
### config={"Quantization": {}, "Distillation": {}}, ### 如果您的系统为Windows系统, 请使用当前这一行配置
181+
### config={"QuantAware": {}, "Distillation": {}}, ### 如果您的系统为Windows系统, 请使用当前这一行配置
182182
train_dataloader=train_loader,
183183
eval_dataloader=train_loader)
184184
ac.compress()

example/auto_compression/detection/configs/picodet_s_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ Distillation:
1818
- conv2d_154.tmp_1
1919
- tmp_8
2020

21-
Quantization:
21+
QuantAware:
2222
use_pact: true
2323
activation_quantize_type: 'moving_average_abs_max'
2424
weight_bits: 8

example/auto_compression/detection/configs/ppyoloe_l_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
alpha: 1.0
1212
loss: soft_label
1313

14-
Quantization:
14+
QuantAware:
1515
onnx_format: true
1616
use_pact: true
1717
activation_quantize_type: 'moving_average_abs_max'

example/auto_compression/detection/configs/ppyoloe_s_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
alpha: 1.0
1212
loss: soft_label
1313

14-
Quantization:
14+
QuantAware:
1515
onnx_format: true
1616
use_pact: true
1717
activation_quantize_type: 'moving_average_abs_max'

example/auto_compression/detection/configs/ssd_mbv1_voc_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ Distillation:
1313
- concat_2.tmp_0
1414
- concat_1.tmp_0
1515

16-
Quantization:
16+
QuantAware:
1717
use_pact: True
1818
weight_quantize_type: 'channel_wise_abs_max'
1919
activation_quantize_type: 'moving_average_abs_max'

example/auto_compression/detection/configs/tinypose_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
node:
1313
- conv2d_441.tmp_0
1414

15-
Quantization:
15+
QuantAware:
1616
use_pact: true
1717
activation_quantize_type: 'moving_average_abs_max'
1818
weight_quantize_type: 'channel_wise_abs_max' # 'abs_max' is layer wise quant

example/auto_compression/detection/configs/yolov3_mbv1_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ Distillation:
1313
- conv2d_85.tmp_0
1414
- conv2d_86.tmp_0
1515

16-
Quantization:
16+
QuantAware:
1717
activation_quantize_type: 'range_abs_max'
1818
quantize_op_types:
1919
- conv2d

example/auto_compression/hyperparameter_tutorial.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -3,11 +3,11 @@
33

44
## 1.1 各压缩方法超参解析
55

6-
### 1.1.1 量化训练(quantization
6+
### 1.1.1 量化训练(quant aware
77

88
量化参数主要设置量化比特数和量化op类型,其中量化op包含卷积层(conv2d, depthwise_conv2d)和全连接层(mul, matmul_v2)。以下为只量化卷积层的示例:
99
```yaml
10-
Quantization:
10+
QuantAware:
1111
use_pact: false # 量化训练是否使用PACT方法
1212
weight_quantize_type: 'channel_wise_abs_max' # 权重量化方式
1313
quantize_op_types: [conv2d, depthwise_conv2d] # 量化OP列表

example/auto_compression/image_classification/configs/EfficientNetB0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- softmax_1.tmp_0
1313

14-
Quantization:
14+
QuantAware:
1515
use_pact: true
1616
activation_bits: 8
1717
is_full_quantize: false

example/auto_compression/image_classification/configs/GhostNet_x1_0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_0.tmp_0
13-
Quantization:
13+
QuantAware:
1414
use_pact: true
1515
activation_bits: 8
1616
is_full_quantize: false

example/auto_compression/image_classification/configs/InceptionV3/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
loss: l2
1313
node:
1414
- softmax_1.tmp_0
15-
Quantization:
15+
QuantAware:
1616
is_full_quantize: false
1717
activation_quantize_type: moving_average_abs_max
1818
weight_quantize_type: channel_wise_abs_max

example/auto_compression/image_classification/configs/MobileNetV1/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_0.tmp_0
13-
Quantization:
13+
QuantAware:
1414
use_pact: true
1515
activation_bits: 8
1616
is_full_quantize: false

example/auto_compression/image_classification/configs/MobileNetV3_large_x1_0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Distillation:
99
alpha: 1.0
1010
loss: soft_label
1111

12-
Quantization:
12+
QuantAware:
1313
use_pact: true
1414
activation_bits: 8
1515
is_full_quantize: false

example/auto_compression/image_classification/configs/PPHGNet_tiny/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- softmax_1.tmp_0
1313

14-
Quantization:
14+
QuantAware:
1515
use_pact: true
1616
activation_bits: 8
1717
is_full_quantize: false

example/auto_compression/image_classification/configs/PPLCNetV2_base/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- softmax_1.tmp_0
1313

14-
Quantization:
14+
QuantAware:
1515
use_pact: true
1616
activation_bits: 8
1717
is_full_quantize: false

example/auto_compression/image_classification/configs/PPLCNet_x1_0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_1.tmp_0
13-
Quantization:
13+
QuantAware:
1414
use_pact: true
1515
activation_bits: 8
1616
is_full_quantize: false

example/auto_compression/image_classification/configs/ResNet50_vd/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- softmax_0.tmp_0
1313

14-
Quantization:
14+
QuantAware:
1515
use_pact: true
1616
activation_bits: 8
1717
is_full_quantize: false

example/auto_compression/image_classification/configs/ShuffleNetV2_x1_0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_0.tmp_0
13-
Quantization:
13+
QuantAware:
1414
use_pact: true
1515
activation_bits: 8
1616
is_full_quantize: false

example/auto_compression/image_classification/configs/SqueezeNet1_0/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_0.tmp_0
13-
Quantization:
13+
QuantAware:
1414
activation_bits: 8
1515
is_full_quantize: false
1616
activation_quantize_type: moving_average_abs_max

example/auto_compression/image_classification/configs/SwinTransformer_base_patch4_window7_224/qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
loss: l2
1111
node:
1212
- softmax_48.tmp_0
13-
Quantization:
13+
QuantAware:
1414
use_pact: true
1515
activation_bits: 8
1616
is_full_quantize: false

example/auto_compression/ocr/configs/ppocrv3_det_qat_dist.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Distillation:
99
alpha: 1.0
1010
loss: l2
1111

12-
Quantization:
12+
QuantAware:
1313
use_pact: true
1414
activation_bits: 8
1515
is_full_quantize: false

example/auto_compression/pytorch_yolo_series/configs/yolov5s_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
alpha: 1.0
1313
loss: soft_label
1414

15-
Quantization:
15+
QuantAware:
1616
onnx_format: true
1717
use_pact: true
1818
activation_quantize_type: 'moving_average_abs_max'

example/auto_compression/pytorch_yolo_series/configs/yolov6s_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
alpha: 1.0
1313
loss: soft_label
1414

15-
Quantization:
15+
QuantAware:
1616
onnx_format: true
1717
activation_quantize_type: 'moving_average_abs_max'
1818
quantize_op_types:

example/auto_compression/pytorch_yolo_series/configs/yolov6s_v2_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ Distillation:
1313
alpha: 1.0
1414
loss: soft_label
1515

16-
Quantization:
16+
QuantAware:
1717
onnx_format: true
1818
activation_quantize_type: 'moving_average_abs_max'
1919
quantize_op_types:

example/auto_compression/pytorch_yolo_series/configs/yolov7_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
alpha: 1.0
1313
loss: soft_label
1414

15-
Quantization:
15+
QuantAware:
1616
onnx_format: true
1717
activation_quantize_type: 'moving_average_abs_max'
1818
quantize_op_types:

example/auto_compression/pytorch_yolo_series/configs/yolov7_tiny_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ Distillation:
1212
alpha: 1.0
1313
loss: soft_label
1414

15-
Quantization:
15+
QuantAware:
1616
onnx_format: true
1717
activation_quantize_type: 'moving_average_abs_max'
1818
quantize_op_types:

example/auto_compression/semantic_segmentation/configs/BiSeNetV2/BiSeNetV2_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- conv2d_103.tmp_1
1313

14-
Quantization:
14+
QuantAware:
1515
onnx_format: True
1616
quantize_op_types:
1717
- conv2d

example/auto_compression/semantic_segmentation/configs/deeplabv3/deeplabv3_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Distillation:
1111
node:
1212
- conv2d_123.tmp_1
1313

14-
Quantization:
14+
QuantAware:
1515
onnx_format: True
1616
quantize_op_types:
1717
- conv2d

example/auto_compression/semantic_segmentation/configs/hrnet/hrnet_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
node:
1111
- conv2d_613.tmp_1
1212

13-
Quantization:
13+
QuantAware:
1414
onnx_format: True
1515
quantize_op_types:
1616
- conv2d

example/auto_compression/semantic_segmentation/configs/pp_humanseg/pp_humanseg_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
node:
1111
- batch_norm_47.tmp_2
1212

13-
Quantization:
13+
QuantAware:
1414
onnx_format: True
1515
quantize_op_types:
1616
- conv2d

example/auto_compression/semantic_segmentation/configs/pp_liteseg/pp_liteseg_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
node:
1111
- conv2d_95.tmp_0
1212

13-
Quantization:
13+
QuantAware:
1414
onnx_format: True
1515
quantize_op_types:
1616
- conv2d

example/auto_compression/semantic_segmentation/configs/unet/unet_qat.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ Distillation:
1010
node:
1111
- conv2d_37.tmp_1
1212

13-
Quantization:
13+
QuantAware:
1414
onnx_format: True
1515
quantize_op_types:
1616
- conv2d

example/auto_compression/tensorflow_mobilenet/configs/mbv1_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ Distillation:
3838
- batch_norm_26.tmp_3
3939
- conv2d_42.tmp_1
4040

41-
Quantization:
41+
QuantAware:
4242
use_pact: true
4343
activation_bits: 8
4444
is_full_quantize: false

example/full_quantization/image_classification/configs/mobilenetv3_large_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ Global:
99
Distillation:
1010
alpha: 1.0
1111
loss: soft_label
12-
Quantization:
12+
QuantAware:
1313
use_pact: true
1414
activation_bits: 8
1515
activation_quantize_type: moving_average_abs_max

example/full_quantization/picodet/configs/picodet_npu.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ PTQ: # Post Training Quantization
1515
batch_size: 32
1616
batch_nums: 10
1717

18-
Quantization: # Auto Compression
18+
QuantAware: # Auto Compression
1919
use_pact: true
2020
activation_quantize_type: 'moving_average_abs_max'
2121
weight_bits: 8

example/full_quantization/picodet/configs/picodet_npu_with_postprocess.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ PTQ: # Post Training Quantization
1515
batch_size: 32
1616
batch_nums: 10
1717

18-
Quantization: # Auto Compression
18+
QuantAware: # Auto Compression
1919
use_pact: true
2020
activation_quantize_type: 'moving_average_abs_max'
2121
weight_bits: 8

example/full_quantization/ppyoloe/configs/ppyoloe_s_416_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ Distillation:
2020
alpha: 1.0
2121
loss: soft_label
2222

23-
Quantization: # Auto Compression
23+
QuantAware: # Auto Compression
2424
onnx_format: true
2525
use_pact: true
2626
activation_quantize_type: 'moving_average_abs_max'

example/full_quantization/ppyoloe/configs/ppyoloe_s_qat_dis.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -16,7 +16,7 @@ PTQ: # Post Training Quantization
1616
batch_size: 10
1717
batch_nums: 10
1818

19-
Quantization: # Auto Compression
19+
QuantAware: # Auto Compression
2020
onnx_format: true
2121
use_pact: true
2222
activation_quantize_type: 'moving_average_abs_max'

paddleslim/auto_compression/__init__.py

+2-1
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,8 @@
2020

2121
__all__ = [
2222
"AutoCompression",
23-
"Quantization",
23+
"QuantAware",
24+
"QuantPost",
2425
"Distillation",
2526
"MultiTeacherDistillation",
2627
"HyperParameterOptimization",

paddleslim/auto_compression/auto_strategy.py

+6-6
Original file line numberDiff line numberDiff line change
@@ -125,17 +125,17 @@ def create_strategy_config(strategy_str, model_type):
125125
### only platform is linux can use smac to do hyperparameter optimization
126126
### choose quant_aware to do quantization in other platform
127127
if platform.system().lower() == 'linux':
128-
quant_config = Quantization(**default_quant_config)
128+
quant_config = QuantAware(**default_quant_config)
129129
hpo_config = HyperParameterOptimization(**hpo_config_tester)
130130
configs.append({
131131
'QuantPost': quant_config,
132132
'HyperParameterOptimization': hpo_config
133133
})
134134
else:
135-
quant_config = Quantization(**default_quant_config)
135+
quant_config = QuantAware(**default_quant_config)
136136
dis_config = Distillation()
137137
configs.append({
138-
'Quantization': quant_config,
138+
'QuantAware': quant_config,
139139
'Distillation': dis_config
140140
})
141141

@@ -248,7 +248,7 @@ def get_final_quant_config(ptq_loss, model_type=None):
248248
return None
249249
### if emd loss less than MAGIC_MAX_EMD_DISTANCE, select quant_post & hpo.
250250
elif ptq_loss < MAGIC_MAX_EMD_DISTANCE:
251-
quant_config = Quantization(**default_quant_config)
251+
quant_config = QuantAware(**default_quant_config)
252252
hpo_config = HyperParameterOptimization(**default_hpo_config)
253253
configs = [{
254254
'QuantPost': quant_config,
@@ -257,9 +257,9 @@ def get_final_quant_config(ptq_loss, model_type=None):
257257

258258
### if emd loss greater than MAGIC_MAX_EMD_DISTANCE, select qat & dist.
259259
else:
260-
quant_config = Quantization(**default_quant_config)
260+
quant_config = QuantAware(**default_quant_config)
261261
dis_config = Distillation()
262-
configs = [{'Quantization': quant_config, 'Distillation': dis_config}]
262+
configs = [{'QuantAware': quant_config, 'Distillation': dis_config}]
263263
_logger.info("Start Quantization and Distillation Training.")
264264

265265
return configs

0 commit comments

Comments
 (0)