Skip to content

Conversation

@Xiaoming-AMD
Copy link
Collaborator

Add configuration files for Meta's LLaMA2 7B/70B models,

init_method_std: 0.02

# multi_latent_attention does not support apply_rope_fusion
apply_rope_fusion: false
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个可以改成True,llama里面目前没有开启mla,可以用fusion

Copy link
Contributor

@wenxie-amd wenxie-amd left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@wenxie-amd wenxie-amd merged commit f6aaa58 into main Apr 9, 2025
1 check passed
@Xiaoming-AMD Xiaoming-AMD changed the title [Feat] Add LLaMA2 7B & 70B model configuration files feature(llama): Add LLaMA2 7B & 70B model configuration files Jun 4, 2025
@Xiaoming-AMD Xiaoming-AMD changed the title feature(llama): Add LLaMA2 7B & 70B model configuration files feat(llama): Add LLaMA2 7B & 70B model configuration files Jun 25, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants