-
Notifications
You must be signed in to change notification settings - Fork 307
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
fix: Adding missing model meta (#1856)
* Added CDE models * Added bge-en-icl * Updated CDE to bge_full_data * Fixed public_training_data flag type to include boolean, as this is how all models are annotated * Added public training data link instead of bool to CDE and BGE * Added GME models * Changed Torch to PyTorch * Added metadata on LENS models * Added ember_v1 * Added metadata for amazon titan * Removed GME implementation
- Loading branch information
1 parent
fde446d
commit 692bd26
Showing
6 changed files
with
292 additions
and
0 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,54 @@ | ||
from __future__ import annotations | ||
|
||
import logging | ||
|
||
from mteb.model_meta import ModelMeta | ||
|
||
from .bge_models import bge_full_data | ||
|
||
logger = logging.getLogger(__name__) | ||
|
||
|
||
cde_small_v1 = ModelMeta( | ||
loader=None, # I will leave this at None for now, | ||
name="jxm/cde-small-v1", | ||
languages=["eng_Latn"], | ||
open_weights=True, | ||
revision="8d5736163718a8b65cd787b75ed61020d18bad3c", | ||
release_date="2024-09-24", | ||
n_parameters=int(281 * 1e6), # Though the second-stage model is only 140M | ||
max_tokens=512, | ||
embed_dim=768, | ||
license="mit", | ||
similarity_fn_name="cosine", | ||
framework=["Sentence Transformers"], | ||
reference="https://huggingface.co/jxm/cde-small-v1", | ||
use_instructions=True, | ||
adapted_from="nomic-ai/nomic-bert-2048", | ||
superseded_by="jxm/cde-small-v2", | ||
training_datasets=bge_full_data, | ||
public_training_code="https://github.com/jxmorris12/cde", | ||
public_training_data="https://huggingface.co/datasets/cfli/bge-full-data", | ||
) | ||
|
||
cde_small_v2 = ModelMeta( | ||
loader=None, # I will leave this at None for now, | ||
name="jxm/cde-small-v2", | ||
languages=["eng_Latn"], | ||
open_weights=True, | ||
revision="a7e5882ad52c27ea2831fc8258f24379c25cb459", | ||
release_date="2025-01-13", | ||
n_parameters=int(306 * 1e6), # Though the second-stage model is only 140M | ||
max_tokens=512, | ||
embed_dim=768, | ||
license="mit", | ||
similarity_fn_name="cosine", | ||
framework=["Sentence Transformers"], | ||
reference="https://huggingface.co/jxm/cde-small-v1", | ||
use_instructions=True, | ||
adapted_from="answerdotai/ModernBERT-base", | ||
superseded_by="jxm/cde-small-v2", | ||
training_datasets=bge_full_data, | ||
public_training_code="https://github.com/jxmorris12/cde", | ||
public_training_data="https://huggingface.co/datasets/cfli/bge-full-data", | ||
) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,63 @@ | ||
from __future__ import annotations | ||
|
||
import logging | ||
from functools import partial | ||
|
||
from mteb.model_meta import ModelMeta | ||
|
||
logger = logging.getLogger(__name__) | ||
|
||
|
||
gme_qwen2_vl_2b_instruct = ModelMeta( | ||
loader=None, | ||
name="Alibaba-NLP/gme-Qwen2-VL-2B-Instruct", | ||
languages=["eng_Latn"], | ||
open_weights=True, | ||
revision="cfeb66885b598de483cc04eb08c7d9da534d7afe", | ||
release_date="2024-12-21", | ||
n_parameters=int(2.21 * 1e9), | ||
max_tokens=32768, | ||
embed_dim=1536, | ||
license="mit", | ||
similarity_fn_name="cosine", | ||
framework=["PyTorch"], | ||
reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-2B-Instruct", | ||
use_instructions=True, | ||
adapted_from=None, | ||
superseded_by=None, | ||
training_datasets={ | ||
# Only annotating text data for now | ||
# source: https://arxiv.org/pdf/2412.16855 | ||
"MSMARCO": ["train"], | ||
"MSMARCO.v2": ["train"], | ||
}, | ||
public_training_code=None, | ||
public_training_data=None, | ||
) | ||
|
||
gme_qwen2_vl_7b_instruct = ModelMeta( | ||
loader=None, | ||
name="Alibaba-NLP/gme-Qwen2-VL-2B-Instruct", | ||
languages=["eng_Latn"], | ||
open_weights=True, | ||
revision="d42eca5a540526cfa982a349724b24b25c12a95e", | ||
release_date="2024-12-21", | ||
n_parameters=int(8.29 * 1e9), | ||
max_tokens=32768, | ||
embed_dim=3584, | ||
license="mit", | ||
similarity_fn_name="cosine", | ||
framework=["PyTorch"], | ||
reference="https://huggingface.co/Alibaba-NLP/gme-Qwen2-VL-7B-Instruct", | ||
use_instructions=True, | ||
adapted_from=None, | ||
superseded_by=None, | ||
training_datasets={ | ||
# Only annotating text data for now | ||
# source: https://arxiv.org/pdf/2412.16855 | ||
"MSMARCO": ["train"], | ||
"MSMARCO.v2": ["train"], | ||
}, | ||
public_training_code=None, | ||
public_training_data=None, | ||
) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,49 @@ | ||
from __future__ import annotations | ||
|
||
from functools import partial | ||
|
||
import torch | ||
|
||
from mteb.encoder_interface import PromptType | ||
from mteb.model_meta import ModelMeta, sentence_transformers_loader | ||
from mteb.models.instruct_wrapper import instruct_wrapper | ||
|
||
lens_d4000 = ModelMeta( | ||
loader=None, # TODO: implement this in the future | ||
name="yibinlei/LENS-d4000", | ||
languages=None, | ||
open_weights=True, | ||
revision="e473b33364e6c48a324796fd1411d3b93670c6fe", | ||
release_date="2025-01-17", | ||
n_parameters=int(7.11 * 1e9), | ||
embed_dim=4000, | ||
license="apache-2.0", | ||
reference="https://huggingface.co/yibinlei/LENS-d4000", | ||
similarity_fn_name="cosine", | ||
framework=["PyTorch"], | ||
use_instructions=True, | ||
public_training_code=None, | ||
public_training_data=None, | ||
training_datasets=None, | ||
max_tokens=32768, | ||
) | ||
|
||
lens_d8000 = ModelMeta( | ||
loader=None, # TODO: implement this in the future | ||
name="yibinlei/LENS-d8000", | ||
languages=None, | ||
open_weights=True, | ||
revision="a0b87bd91cb27b6f2f0b0fe22c28026da1d464ef", | ||
release_date="2025-01-17", | ||
n_parameters=int(7.11 * 1e9), | ||
embed_dim=8000, | ||
license="apache-2.0", | ||
reference="https://huggingface.co/yibinlei/LENS-d8000", | ||
similarity_fn_name="cosine", | ||
framework=["PyTorch"], | ||
use_instructions=True, | ||
public_training_code=None, | ||
public_training_data=None, | ||
training_datasets=None, | ||
max_tokens=32768, | ||
) |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters