-
Notifications
You must be signed in to change notification settings - Fork 212
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Allow passing batch_size via CLI #1025
Comments
Yeah. And within the tasks / evaluators, |
That has been the approach so far. I would actually probably prefer having encode_kwargs as an argument for the MTEB(model, encode_kwargs={"batch_size": 16}). Batch size seem very specific. This also allow passing e.g. length normalization. We can still add the batch_size argument to the CLI |
As @KennethEnevoldsen said you can already pass it in evaluation.run like e.g. here: https://github.com/ContextualAI/gritlm/blob/0cc9aeab83b90f2e22bcdd2b084d51507c624d95/evaluation/eval_mteb.py#L1206 Maybe it would help having that in the docs / one of the README examples so people know 🤔 |
Yea. I was actually hoping to transition away from using the batch_size arguments (essentially it is a model argument, but it is passed on to the task). I also don't think it is consistently implemented for all tasks types. |
I am more than happy to implement the encode_kwargs if people think it is a good idea. |
I think it'd make sense to allow passing the
batch_size
kwarg when using the cli argumentmteb run ...
? Some missing results are due to OOMs (#1014)The text was updated successfully, but these errors were encountered: