prrao87 / mteb-validation

Compare different embedding models from MTEB leaderboard

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

MTEB embedding model validation

Comparing results of embedding models from the MTEB leaderboard.

The goal is to study the classes of embeddings models that top the MTEB leaderboard produces high-quality embeddings across domains.

The MTEB leaderboard, or the “Massive Text Embedding Benchmark”, ranks sentence embedding models over a total of 129 total datasets and 113 different languages. Because there are ever more models being added to this leaderboard, it becomes necessary to evaluate specific models on custom datasets of

The goal of this repo is to define and describe the target tasks that will be used to evaluate the results of embedding models from the MTEB leaderboard.

Datasets

Evaluating documents across various domains would be useful to understand the impact of using a particular model (off-the-shelf) prior to fine-tuning for a custom use case. This section lists the various datasets that could be used to evaluate the quality of retrievals from a vector database that stores the underlying embeddings from each of the models. For each of the cases listed below, a set of 20 questions will be formulated (via human inspection) from these datasets for retrieval via a vector database downstream.

About

Compare different embedding models from MTEB leaderboard


Languages

Language:Python 100.0%