Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix vLLM max tokens #570

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Conversation

Datta0
Copy link

@Datta0 Datta0 commented Feb 18, 2025

vLLM has max_tokens as Optional[int] but defaulting to 16 here
That means whenever sampling_params is created, it assumes the value 16 and hence this sampling_params.max_tokens ends up being always equal to 16
Then the lighteval benchmark goes on to warn that the output is not in the Gold Format ...

Copy link
Member

@NathanHB NathanHB left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi ! The behaviour we want is to use the max_new_tokens defined in the task config to be used by default except when it is set in the model's generation config.
In that case, the fix you would need to is set the sampling_params.max_tokens default value to None

@Datta0
Copy link
Author

Datta0 commented Feb 18, 2025

Is there a way to control sampling params while invoking lighteval from the CLI? I couldn't find it in the documentation unfortunately...

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants