Files
FastDeploy/tools/common_tools/auto_compression/configs
yunyaoXYY 2d3d941372 [Quantization] Add new PaddleSeg models quantization support. (#953)
* Fix links in readme

* Fix links in readme

* Update PPOCRv2/v3 examples

* Update auto compression configs

* Add neww quantization  support for paddleclas model

* Update quantized Yolov6s model download link

* Improve PPOCR comments

* Add English doc for quantization

* Fix PPOCR rec model bug

* Add  new paddleseg quantization support

* Add  new paddleseg quantization support

* Add  new paddleseg quantization support

* Add  new paddleseg quantization support
2022-12-25 16:33:57 +08:00
..

Quantization Config File on FastDeploy

The FastDeploy quantization configuration file contains global configuration, quantization distillation training configuration, post-training quantization configuration and training configuration. In addition to using the configuration files provided by FastDeploy directly in this directory, users can modify the relevant configuration files according to their needs

Demo

# Global config
Global:
  model_dir: ./ppyoloe_plus_crn_s_80e_coco    #Path to input model
  format: paddle                              #Input model format, please select 'paddle' for paddle model
  model_filename: model.pdmodel               #Quantized model name in Paddle format
  params_filename: model.pdiparams            #Parameter name for quantized paddle model
  qat_image_path: ./COCO_train_320            #Data set paths for quantization distillation training
  ptq_image_path: ./COCO_val_320              #Data set paths for PTQ
  input_list: ['image','scale_factor']        #Input name of the model to be quanzitzed
  qat_preprocess: ppyoloe_plus_withNMS_image_preprocess # The preprocessing function for Quantization distillation training
  ptq_preprocess: ppyoloe_plus_withNMS_image_preprocess # The preprocessing function for PTQ
  qat_batch_size: 4                           #Batch size


# Quantization distillation training configuration
Distillation:
  alpha: 1.0                                  #Distillation loss weight
  loss: soft_label                            #Distillation loss algorithm

QuantAware:
  onnx_format: true                           #Whether to use ONNX quantization standard format or not, must be true to deploy on FastDeploy
  use_pact: true                              #Whether to use the PACT method for training
  activation_quantize_type: 'moving_average_abs_max'     #Activations quantization methods
  quantize_op_types:                          #OPs that need to be quantized
  - conv2d
  - depthwise_conv2d

# Post-Training Quantization
PTQ:
  calibration_method: 'avg'                   #Activations calibration algorithm of post-training quantization , Options: avg, abs_max, hist, KL, mse, emd
  skip_tensor_list: None                      #Developers can skip some conv layers quantization

# Training Config
TrainConfig:
  train_iter: 3000
  learning_rate: 0.00001
  optimizer_builder:
    optimizer:
      type: SGD
    weight_decay: 4.0e-05
  target_metric: 0.365

More details

FastDeploy one-click quantization tool is powered by PaddeSlim, please refer to Auto Compression Hyperparameter Tutorial for more details.