mteb icon indicating copy to clipboard operation
mteb copied to clipboard

Finalizing MMTEB

Open KennethEnevoldsen opened this issue 1 year ago • 6 comments

This issue is to get an overview of what needs to be done before MMTEB can be finalized.

  1. Adding the last remaining datasets, notably:
    • [x] #641 #830
    • [x] #718
    • [x] #642 #833
  2. Speeding up the benchmark
    • [x] I believe we are only missing: #660
    • see also #836
    • see also #838
    • see also #835
  3. #705 (partly depends on 1, 2 as well as #879)
  4. Figuring out #752 (partly depends on 3)
  5. Deciding on meaningful benchmark subsets (depends on 3)
  • see #837
  1. #896 (depends on 3, 4 and 5) (see also #595)
  • see #839
  1. Updating leaderboard to new format https://github.com/embeddings-benchmark/mteb/discussions/674 (depends on 3-6)

Is there anything else that is needed?

KennethEnevoldsen avatar May 22 '24 08:05 KennethEnevoldsen

Construction of MMTEB-Lite? It will be a faster version of MMTEB. Two approaches that come to mind for implementing this are -

  1. Reducing the size document set of some retrieval benchmarks.
  2. Reducing the number of tasks

vaibhavad avatar May 22 '24 14:05 vaibhavad

Hey @KennethEnevoldsen I'd like to merge also this dataset in #773. 3 reasons: a) we don't seem to have brazilian dialect represeted, b) multilabel task doesn't have large language coverage c) I had it prepared for long time, but multilabel task got only merged last week when I was away. We only need to address a problem with stratification of the splits there.

dokato avatar May 23 '24 07:05 dokato

@vaibhavad yes, def. we need to construct the benchmarks and ideally think about downsampling some of the larger retrieval datasets. A solution might be to implement a downsample function for retrieval tasks.

Thanks @dokato - let us get it merged in as well. Looks to be in a reasonable state

KennethEnevoldsen avatar May 23 '24 08:05 KennethEnevoldsen

Hey @KennethEnevoldsen I read the list and I think I can help in Running models https://github.com/embeddings-benchmark/mteb/discussions/705

Ruqyai avatar May 24 '24 06:05 Ruqyai

@KennethEnevoldsen Is there anything meaningful new contributors can help with?

jordiclive avatar May 31 '24 11:05 jordiclive

Hi @jordiclive! I believe there are multiple avenues to take, but any of the outlines paper segment I believe is meaning (see the updated post above), implementing model (see e.g. #845, will finish it up either Monday or in the weekend), or starting work on 8)

KennethEnevoldsen avatar May 31 '24 13:05 KennethEnevoldsen

quick question: is there a script to select & run all MMTEB tasks? I'm a bit unclear about the difference between current development progress and how MMTEB is different from the current MTEB (in different languages).

Best

Bo

bwanglzu avatar Oct 29 '24 13:10 bwanglzu

@bwanglzu You can select benchmark like this:

import mteb
mteb.get_benchmark("MTEB(eng, classic)") # or get_benchmarks

Full list of bencharks here

Samoed avatar Oct 29 '24 14:10 Samoed

Will close this issue as MMTEB has been submitted, moving the public preprint release over to https://github.com/embeddings-benchmark/mteb/issues/1405

KennethEnevoldsen avatar Nov 07 '24 15:11 KennethEnevoldsen