Skip to content

Fix flaky penalty test#319

Closed
rdraskicTT wants to merge 1 commit intodevfrom
rdraskic/fix-flaky-penalty-test
Closed

Fix flaky penalty test#319
rdraskicTT wants to merge 1 commit intodevfrom
rdraskic/fix-flaky-penalty-test

Conversation

@rdraskicTT
Copy link

@rdraskicTT rdraskicTT commented Feb 13, 2026

Temperature is too low to create variability in response and show the effect of penalty

@rdraskicTT rdraskicTT marked this pull request as ready for review February 13, 2026 14:18
@rdraskicTT rdraskicTT requested a review from tchedaTT as a code owner February 13, 2026 14:18
@sraizada-tt
Copy link

why do we need this change?

@rdraskicTT
Copy link
Author

rdraskicTT commented Feb 13, 2026

why do we need this change?

This was the only test failing for llama 8b in vllm nightly. It was producing the same responses due to low temperature and short outputs.

@sraizada-tt
Copy link

hmm because 0.5 is bumping up the max token too much? okay

@rdraskicTT
Copy link
Author

hmm because 0.5 is bumping up the max token too much? okay

Yeah, I preferred this since it keeps the test shorter

@tchedaTT
Copy link

Alternative fix merged in #307

@tchedaTT tchedaTT closed this Feb 13, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants