Skip to content
Discussion options

You must be logged in to vote

Hello, you can indeed change RoPE by updating/override the model config. Say you're using this model https://huggingface.co/allenai/OLMo-2-0325-32B-SFT/blob/main/config.json :

You can set the rope_scaling similar to the Olmo3 32B model https://huggingface.co/allenai/Olmo-3-32B-Think/blob/main/config.json#L79-L94

overrides_of_model_config:
  rope_scaling:
    type: ${ROPE_SCALING_TYPE}
    factor: ${ROPE_SCALING_FACTOR}
  rope_theta: ${ROPE_THETA}

sequence_len: ${NEW_SEQ}

Make sure that: the model itself Olmo2 in transformers supports rope config too.

Curious question: why not start from Olmo3 which already has their context expanded?

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@cancan4
Comment options

@NanoCode012
Comment options

Answer selected by cancan4
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants