Skip to content
Discussion options

You must be logged in to vote

Greetings, @stktyagi!

Thank you for the idea, it looks interesting!
Gradient-based methods, unfortunately, can be quite costly in terms of both time and memory. We are focusing on training-free methods now.

We have other potential directions.

  1. For example, unstructured sparsity https://github.com/openvinotoolkit/nncf/blob/develop/examples/pruning/torch/resnet18/main.py#L61
    Currently, it’s only magnitude based, but there’s no 2:4 or m:n sparsity, which is opportunity for contribution.
  2. If you still want something with training, we have efficient distillation with lora adapters https://github.com/openvinotoolkit/nncf/tree/develop/examples/llm_compression/torch/distillation_qat_with_lora
    It w…

Replies: 2 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@stktyagi
Comment options

Answer selected by stktyagi
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Ideas
Labels
None yet
3 participants