Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Added benchmark object #878

Merged
merged 8 commits into from
Jun 5, 2024

Conversation

KennethEnevoldsen
Copy link
Contributor

Assuming we want a selection menu for benchmarks in the future leaderboard I have now added a benchmark object. Generally it shouldn't change much otherwise.

The idea is that you would select the benchmark. The description will show + some references. Potentially with a dropdown for the citation.

  • Added benchmark object (should work like a list)
  • removed a duplicate task
  • Added SEB (we should probably add a few more here as well)

Copy link
Contributor

@Muennighoff Muennighoff left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice, I think we also need to change the README which imports MTEB_MAIN_EN i.e. https://github.com/embeddings-benchmark/mteb?tab=readme-ov-file#dataset-selection

@KennethEnevoldsen
Copy link
Contributor Author

Nice, I think we also need to change the README which imports MTEB_MAIN_EN i.e. https://github.com/embeddings-benchmark/mteb?tab=readme-ov-file#dataset-selection

Added a test and it does run just fine (since it has an iter method). I also changed the type hint to iterable (weaker than sequence) to reflect that it is the only requirement.

@KennethEnevoldsen KennethEnevoldsen merged commit 46170d0 into result-normalization Jun 5, 2024
6 checks passed
@KennethEnevoldsen KennethEnevoldsen deleted the add_benchmark_handling branch June 5, 2024 08:03
KennethEnevoldsen added a commit that referenced this pull request Jun 5, 2024
* Ensure result are consistently stored in the same way

- (due to failing test): updated missing dataset references
- (to test with more than one model) Added e5 models base and large
- updated mteb.get_model to now include metadata in the model object
- ensure that model name is always included when saving (with a default when it is not available)
- use the ModelMeta for the model_meta.json

* format

* minor test fixes

* docs: Minor updated to repro. workflow docs

* fixed failing test

* format

* Apply suggestions from code review

Co-authored-by: Isaac Chung <chungisaac1217@gmail.com>

* docs: update PR template

* fix: Added benchmark object (#878)

* removed duplicate task

* Added benchmark object

* removed import for duplicate task

* fix dataset references

* added seb

* Added test for running benchmarks

* changed tasks to be an iterable

* format

* Apply suggestions from code review

Co-authored-by: Niklas Muennighoff <n.muennighoff@gmail.com>
Co-authored-by: Isaac Chung <chungisaac1217@gmail.com>

---------

Co-authored-by: Isaac Chung <chungisaac1217@gmail.com>
Co-authored-by: Niklas Muennighoff <n.muennighoff@gmail.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants