Skip to content

Can we borrow information from aimet-onnx when doing Mix-QAT #4051

@iamweiweishi

Description

@iamweiweishi

I used aimet-onnx to do mix-precision quantization, and found that 15% percentage of int16 activation could meet my requirements.

Then, I tried to do QAT to improve the accuracy. But I found that the results from aimet-onnx cannot be used when doing QAT.

Is there any way to achieve this? Or, is there any way to config the activation of specific layers to be int16-activation, like in config_file ?

Thank you.

Metadata

Metadata

Assignees

No one assigned

    Labels

    aimet-onnxNew feature or bug fix for AIMET ONNX

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions