This tutorial briefly introduces how to export an OpenMMlab model to a specific backend using MMDeploy tools. Notes:
- Supported backends are ONNXRuntime, TensorRT, ncnn, PPLNN, OpenVINO.
- Supported codebases are MMPretrain, MMDetection, MMSegmentation, MMOCR, MMagic.
- Install and build your target backend. You could refer to ONNXRuntime-install, TensorRT-install, ncnn-install, PPLNN-install, OpenVINO-install for more information.
- Install and build your target codebase. You could refer to MMPretrain-install, MMDetection-install, MMSegmentation-install, MMOCR-install, MMagic-install.
python ./tools/deploy.py \
${DEPLOY_CFG_PATH} \
${MODEL_CFG_PATH} \
${MODEL_CHECKPOINT_PATH} \
${INPUT_IMG} \
--test-img ${TEST_IMG} \
--work-dir ${WORK_DIR} \
--calib-dataset-cfg ${CALIB_DATA_CFG} \
--device ${DEVICE} \
--log-level INFO \
--show \
--dump-infodeploy_cfg: The deployment configuration of mmdeploy for the model, including the type of inference framework, whether quantize, whether the input shape is dynamic, etc. There may be a reference relationship between configuration files,mmdeploy/mmpretrain/classification_ncnn_static.pyis an example.model_cfg: Model configuration for algorithm library, e.g.mmpretrain/configs/vision_transformer/vit-base-p32_ft-64xb64_in1k-384.py, regardless of the path to mmdeploy.checkpoint: torch model path. It can start with http/https, see the implementation ofmmcv.FileClientfor details.img: The path to the image or point cloud file used for testing during the model conversion.--test-img: The path of the image file that is used to test the model. If not specified, it will be set toNone.--work-dir: The path of the work directory that is used to save logs and models.--calib-dataset-cfg: Only valid in int8 mode. The config used for calibration. If not specified, it will be set toNoneand use the "val" dataset in the model config for calibration.--device: The device used for model conversion. If not specified, it will be set tocpu. For trt, usecuda:0format.--log-level: To set log level which in'CRITICAL', 'FATAL', 'ERROR', 'WARN', 'WARNING', 'INFO', 'DEBUG', 'NOTSET'. If not specified, it will be set toINFO.--show: Whether to show detection outputs.--dump-info: Whether to output information for SDK.
- Find the model's codebase folder in
configs/. For converting a yolov3 model, you need to checkconfigs/mmdetfolder. - Find the model's task folder in
configs/codebase_folder/. For a yolov3 model, you need to checkconfigs/mmdet/detectionfolder. - Find the deployment config file in
configs/codebase_folder/task_folder/. For deploying a yolov3 model to the onnx backend, you could useconfigs/mmdet/detection/detection_onnxruntime_dynamic.py.
python ./tools/deploy.py \
configs/mmdet/detection/detection_tensorrt_dynamic-320x320-1344x1344.py \
$PATH_TO_MMDET/configs/yolo/yolov3_d53_8xb8-ms-608-273e_coco.py \
$PATH_TO_MMDET/checkpoints/yolo/yolov3_d53_mstrain-608_273e_coco_20210518_115020-a2c3acb8.pth \
$PATH_TO_MMDET/demo/demo.jpg \
--work-dir work_dir \
--show \
--device cuda:0You can try to evaluate model, referring to how_to_evaluate_a_model.
Refer to Support model list