Skip to content

RoPE vs RPB #268

@Bezdarnost

Description

@Bezdarnost

Hello,

First and foremost, thank you for your excellent work on this project.

I've noticed that support for Relative Position Bias (RPB) was removed starting with version 0.20.0. I would like to kindly request if you would consider re-enabling it.

The reason for this request is that models using RPB demonstrate significantly better performance when trained on a single resolution and then used for inference across a wide variety of different resolutions. With RPB, the relative positions within the attention window are always consistent, meaning these relationships are always well-trained.

In contrast, when changing resolutions at inference time, a model trained with RoPE tends to show a degradation in performance if it hasn't been exposed to those resolutions during training.

Thank you for your time and consideration.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions