Adding new benchmark MTEB-NL #3339
Replies: 3 comments 7 replies
-
|
Yes, you can start to integrate your benchmark now, but we're planning release |
Beta Was this translation helpful? Give feedback.
-
|
Sounds great @nikolay-banar - nice work on the article. As @Samoed says, I would wait a week until we have v2 ready. However, just a few questions:
|
Beta Was this translation helpful? Give feedback.
-
|
I have a question regarding prompts. In our experiments, we fed e5-style prompts directly to the models, but I guess it is not the best solution for MTEB. The default prompts (e.g. Does it make sense to add the prompts now to every dataset class? class ArguAnaNL(AbsTaskRetrieval):
prompt={"query": "Given a claim, find documents that refute the claim"}The results from MTEB-NL will not be affected, because I did not submit any instruct models. However, BEIR-NL has some instruct submissions. In this case, I will need to rerun some experiments. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi all,
Recently, we released Massive Text Embedding Benchmark for Dutch (MTEB-NL). You can find the paper here.
We are planning to submit it to some conferences, so it hasn’t been peer-reviewed yet.
Would it make sense to start integrating the benchmark into the MTEB leaderboard now, or should we wait until after peer review?
Sincerely,
Nicolae
Beta Was this translation helpful? Give feedback.
All reactions