Skip to content

Conversation

@gkisalapl
Copy link
Contributor

Dependency of the PR

Commits to be reviewed in this PR

{commit-1}

{commit message}

Self evaluation:

  1. Build test: [X]Passed [ ]Failed [ ]Skipped
  2. Run test: [X]Passed [ ]Failed [ ]Skipped

Signed-off-by: {your_name} <{your_email}>

{commit-2}

{commit message}

Self evaluation:

  1. Build test: [X]Passed [ ]Failed [ ]Skipped
  2. Run test: [X]Passed [ ]Failed [ ]Skipped

Signed-off-by: {your_name} <{your_email}>

Summary

  • {Summary of PR 1}
  • {Summary of PR 2}

Signed-off-by: {your_name} <{your_email}>

gkisalapl and others added 11 commits October 6, 2025 12:45
Fix problem of group size incosistency

Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
**Self evaluation:**
1. Build test:     [X]Passed [ ]Failed [ ]Skipped
2. Run test:     [X]Passed [ ]Failed [ ]Skipped

Signed-off-by: Maciej Nalewaj <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
Signed-off-by: Grzegorz Kisala/Neural Computing (AIS) /SRPOL/Senior Professional/Samsung Electronics <[email protected]>
std::vector<uint8_t> quantized_weights_int4;
std::vector<uint16_t> quantized_scales_int4;
Int4Utils::quantizeAndRepack(dequantized_weights_q4.data(), N, K,
Int4Utils::quantizeAndRepack((float *)(file_view + start_from), N, K,
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

FP32 data should be transposed before it is quantized. Otherwise, it would produce an incorrect output.


std::cout << "New size: " << size << std::endl;
} else {
output.write(file_view + start_from, size);
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

In the case of tie word embedding, you need to filter out the LM head case and not save it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants