-
Notifications
You must be signed in to change notification settings - Fork 312
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Added benchmark object #878
Added benchmark object #878
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice, I think we also need to change the README which imports MTEB_MAIN_EN i.e. https://github.com/embeddings-benchmark/mteb?tab=readme-ov-file#dataset-selection
Added a test and it does run just fine (since it has an iter method). I also changed the type hint to iterable (weaker than sequence) to reflect that it is the only requirement. |
* Ensure result are consistently stored in the same way - (due to failing test): updated missing dataset references - (to test with more than one model) Added e5 models base and large - updated mteb.get_model to now include metadata in the model object - ensure that model name is always included when saving (with a default when it is not available) - use the ModelMeta for the model_meta.json * format * minor test fixes * docs: Minor updated to repro. workflow docs * fixed failing test * format * Apply suggestions from code review Co-authored-by: Isaac Chung <chungisaac1217@gmail.com> * docs: update PR template * fix: Added benchmark object (#878) * removed duplicate task * Added benchmark object * removed import for duplicate task * fix dataset references * added seb * Added test for running benchmarks * changed tasks to be an iterable * format * Apply suggestions from code review Co-authored-by: Niklas Muennighoff <n.muennighoff@gmail.com> Co-authored-by: Isaac Chung <chungisaac1217@gmail.com> --------- Co-authored-by: Isaac Chung <chungisaac1217@gmail.com> Co-authored-by: Niklas Muennighoff <n.muennighoff@gmail.com>
Assuming we want a selection menu for benchmarks in the future leaderboard I have now added a benchmark object. Generally it shouldn't change much otherwise.
The idea is that you would select the benchmark. The description will show + some references. Potentially with a dropdown for the citation.