Name | model2vec JSON |
Version |
0.4.0
JSON |
| download |
home_page | None |
Summary | Fast State-of-the-Art Static Embeddings |
upload_time | 2025-02-12 19:49:38 |
maintainer | None |
docs_url | None |
author | None |
requires_python | >=3.9 |
license | MIT License
Copyright (c) 2024 Thomas van Dongen
Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to deal
in the Software without restriction, including without limitation the rights
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all
copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
SOFTWARE.
|
keywords |
|
VCS |
 |
bugtrack_url |
|
requirements |
No requirements were recorded.
|
Travis-CI |
No Travis.
|
coveralls test coverage |
No coveralls.
|
<div align="center">
<picture>
<img width="35%" alt="Model2Vec logo" src="assets/images/logo_v2.png">
</picture>
</a>
</div>
<div align="center">
<h2>Fast State-of-the-Art Static Embeddings</h2>
</div>
<div align="center">
<h2>
<a href="https://huggingface.co/minishlab"><strong>π€ Models</strong></a> |
<a href="https://github.com/MinishLab/model2vec/tree/main/tutorials"><strong>π Tutorials</strong></a> |
<a href="https://minishlab.github.io/"><strong>π Blog</strong></a> |
<a href="https://github.com/MinishLab/model2vec/blob/main/results/README.md"><strong>π Results</strong></a> |
<a href="https://github.com/MinishLab/model2vec/blob/main/docs"><strong>π Docs</strong></a>
</h2>
</div>
<div align="center">
<h2>
<a href="https://pypi.org/project/model2vec/"><img src="https://img.shields.io/pypi/v/model2vec?color=%23007ec6&label=pypi%20package" alt="Package version"></a>
<a href="https://pypi.org/project/model2vec/"><img src="https://img.shields.io/pypi/pyversions/model2vec" alt="Supported Python versions"></a>
<a href="https://pepy.tech/project/model2vec">
<img src="https://static.pepy.tech/badge/model2vec" alt="Downloads">
</a>
<a href="https://app.codecov.io/gh/MinishLab/model2vec">
<img src="https://codecov.io/gh/MinishLab/model2vec/graph/badge.svg?token=21TWJ6B5ET" alt="Codecov">
</a>
</a>
<a href="https://github.com/MinishLab/model2vec/blob/main/LICENSE"><img src="https://img.shields.io/badge/license-MIT-green" alt="License - MIT"></a>
</h2>
</div>
Model2Vec is a technique to turn any sentence transformer into a really small static model, reducing model size by a factor up to 50 and making the models up to 500 times faster, with a small drop in performance. Our [best model](https://huggingface.co/minishlab/potion-base-8M) is the most performant static embedding model in the world. See our results [here](results/README.md), or dive in to see how it works.
<div align="center">
<h3>
[Quickstart](#quickstart) β’ [Updates & Announcements](#updates--announcements) β’ [Main Features](#main-features) β’ [Model List](#model-list)
</h3>
</div>
## Quickstart
Install the lightweight base package with:
```bash
pip install model2vec
```
You can start using Model2Vec by loading one of our [flagship models from the HuggingFace hub](https://huggingface.co/collections/minishlab/potion-6721e0abd4ea41881417f062). These models are pre-trained and ready to use. The following code snippet shows how to load a model and make embeddings, which you can use for any task, such as text classification, retrieval, clustering, or building a RAG system:
```python
from model2vec import StaticModel
# Load a model from the HuggingFace hub (in this case the potion-base-8M model)
model = StaticModel.from_pretrained("minishlab/potion-base-8M")
# Make embeddings
embeddings = model.encode(["It's dangerous to go alone!", "It's a secret to everybody."])
# Make sequences of token embeddings
token_embeddings = model.encode_as_sequence(["It's dangerous to go alone!", "It's a secret to everybody."])
```
Instead of using one of our models, you can also distill your own Model2Vec model from a Sentence Transformer model. First, install the `distillation` extras with:
```bash
pip install model2vec[distill]
```
Then, you can distill a model in ~30 seconds on a CPU with the following code snippet:
```python
from model2vec.distill import distill
# Distill a Sentence Transformer model, in this case the BAAI/bge-base-en-v1.5 model
m2v_model = distill(model_name="BAAI/bge-base-en-v1.5", pca_dims=256)
# Save the model
m2v_model.save_pretrained("m2v_model")
```
After distillation, you can also fine-tune your own classification models on top of the distilled model, or on a pre-trained model. First, make sure you install the `training` extras with:
```bash
pip install model2vec[training]
```
Then, you can fine-tune a model as follows:
```python
import numpy as np
from datasets import load_dataset
from model2vec.train import StaticModelForClassification
# Initialize a classifier from a pre-trained model
classifier = StaticModelForClassification.from_pretrained(model_name="minishlab/potion-base-32M")
# Load a dataset
ds = load_dataset("setfit/subj")
# Train the classifier on text (X) and labels (y)
classifier.fit(ds["train"]["text"], ds["train"]["label"])
# Evaluate the classifier
predictions = classifier.predict(ds["test"]["text"])
accuracy = np.mean(np.array(predictions) == np.array(ds["test"]["label"])) * 100
```
For advanced usage, please refer to our [usage documentation](https://github.com/MinishLab/model2vec/blob/main/docs/usage.md).
## Updates & Announcements
- **12/02/2024**: We released **Model2Vec training**, allowing you to fine-tune your own classification models on top of Model2Vec models. Find out more in our [training documentation](https://github.com/MinishLab/model2vec/blob/main/model2vec/train/README.md) and [results](results/README.md#training-results).
- **30/01/2024**: We released two new models: [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) and [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M). [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) is our most performant model to date, using a larger vocabulary and higher dimensions. [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M) is a finetune of [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) that is optimized for retrieval tasks, and is the best performing static retrieval model currently available.
- **30/10/2024**: We released three new models: [potion-base-8M](https://huggingface.co/minishlab/potion-base-8M), [potion-base-4M](https://huggingface.co/minishlab/potion-base-4M), and [potion-base-2M](https://huggingface.co/minishlab/potion-base-2M). These models are trained using [Tokenlearn](https://github.com/MinishLab/tokenlearn). Find out more in our [blog post](https://minishlab.github.io/tokenlearn_blogpost/). NOTE: for users of any of our old English M2V models, we recommend switching to these new models as they [perform better on all tasks](https://github.com/MinishLab/model2vec/tree/main/results).
## Main Features
- **State-of-the-Art Performance**: Model2Vec models outperform any other static embeddings (such as GLoVe and BPEmb) by a large margin, as can be seen in our [results](results/README.md).
- **Small**: Model2Vec reduces the size of a Sentence Transformer model by a factor of up to 50. Our [best model](https://huggingface.co/minishlab/potion-base-8M) is just ~30 MB on disk, and our smallest model just ~8 MB (making it the smallest model on [MTEB](https://huggingface.co/spaces/mteb/leaderboard)!).
- **Lightweight Dependencies**: the base package's only major dependency is `numpy`.
- **Lightning-fast Inference**: up to 500 times faster on CPU than the original model.
- **Fast, Dataset-free Distillation**: distill your own model in 30 seconds on a CPU, without a dataset.
- **Fine-tuning**: fine-tune your own classification models on top of Model2Vec models.
- **Integrated in many popular libraries**: Model2Vec is integrated direclty into popular libraries such as [Sentence Transformers](https://github.com/UKPLab/sentence-transformers) and [LangChain](https://github.com/langchain-ai/langchain). For more information, see our [integrations documentation](https://github.com/MinishLab/model2vec/blob/main/docs/integrations.md).
- **Tightly integrated with HuggingFace hub**: easily share and load models from the HuggingFace hub, using the familiar `from_pretrained` and `push_to_hub`. Our own models can be found [here](https://huggingface.co/minishlab).
## What is Model2Vec?
Model2vec creates a small, fast, and powerful model that outperforms other static embedding models by a large margin on all tasks we could find, while being much faster to create than traditional static embedding models such as GloVe. Like BPEmb, it can create subword embeddings, but with much better performance. Distillation doesn't need _any_ data, just a vocabulary and a model.
The core idea is to forward pass a vocabulary through a sentence transformer model, creating static embeddings for the indiviudal tokens. After this, there are a number of post-processing steps we do that results in our best models. For a more extensive deepdive, please refer to the following resources:
- Our initial [Model2Vec blog post](https://huggingface.co/blog/Pringled/model2vec). Note that, while this post gives a good overview of the core idea, we've made a number of substantial improvements since then.
- Our [Tokenlearn blog post](https://minishlab.github.io/tokenlearn_blogpost/). This post describes the Tokenlearn method we used to train our [potion models](https://huggingface.co/collections/minishlab/potion-6721e0abd4ea41881417f062).
- Our official [documentation](https://github.com/MinishLab/model2vec/blob/main/docs/what_is_model2vec.md). This document provides a high-level overview of how Model2Vec works.
## Documentation
Our official documentation can be found [here](https://github.com/MinishLab/model2vec/blob/main/docs/README.md). This includes:
- [Usage documentation](https://github.com/MinishLab/model2vec/blob/main/docs/usage.md): provides a technical overview of how to use Model2Vec.
- [Integrations documentation](https://github.com/MinishLab/model2vec/blob/main/docs/integrations.md): provides examples of how to use Model2Vec in various downstream libraries.
- [Model2Vec technical documentation](https://github.com/MinishLab/model2vec/blob/main/docs/what_is_model2vec.md): provides a high-level overview of how Model2Vec works.
## Model List
We provide a number of models that can be used out of the box. These models are available on the [HuggingFace hub](https://huggingface.co/collections/minishlab/model2vec-base-models-66fd9dd9b7c3b3c0f25ca90e) and can be loaded using the `from_pretrained` method. The models are listed below.
| Model | Language | Sentence Transformer | Params | Task |
|-----------------------------------------------------------------------|------------|-----------------------------------------------------------------|---------|-----------|
| [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 32.3M | General |
| [potion-base-8M](https://huggingface.co/minishlab/potion-base-8M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 7.5M | General |
| [potion-base-4M](https://huggingface.co/minishlab/potion-base-4M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 3.7M | General |
| [potion-base-2M](https://huggingface.co/minishlab/potion-base-2M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 1.8M | General |
| [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 32.3M | Retrieval |
| [M2V_multilingual_output](https://huggingface.co/minishlab/M2V_multilingual_output) | Multilingual | [LaBSE](https://huggingface.co/sentence-transformers/LaBSE) | 471M | General |
## Results
We have performed extensive experiments to evaluate the performance of Model2Vec models. The results are documented in the [results](results/README.md) folder. The results are presented in the following sections:
- [MTEB Results](results/README.md#mteb-results)
- [Training Results](results/README.md#training-results)
- [Ablations](results/README.md#ablations)
## License
MIT
## Citing
If you use Model2Vec in your research, please cite the following:
```bibtex
@software{minishlab2024model2vec,
authors = {Stephan Tulkens and Thomas van Dongen},
title = {Model2Vec: Fast State-of-the-Art Static Embeddings},
year = {2024},
url = {https://github.com/MinishLab/model2vec}
}
```
Raw data
{
"_id": null,
"home_page": null,
"name": "model2vec",
"maintainer": null,
"docs_url": null,
"requires_python": ">=3.9",
"maintainer_email": null,
"keywords": null,
"author": null,
"author_email": "St\u00e9phan Tulkens <stephantul@gmail.com>, Thomas van Dongen <thomas123@live.nl>",
"download_url": "https://files.pythonhosted.org/packages/83/e2/3fb7bd8c612f71ad3abded92e7401f97f1e71427d3a68a3fb85f39394b17/model2vec-0.4.0.tar.gz",
"platform": null,
"description": "\n<div align=\"center\">\n <picture>\n <img width=\"35%\" alt=\"Model2Vec logo\" src=\"assets/images/logo_v2.png\">\n </picture>\n </a>\n</div>\n\n<div align=\"center\">\n <h2>Fast State-of-the-Art Static Embeddings</h2>\n</div>\n\n<div align=\"center\">\n <h2>\n <a href=\"https://huggingface.co/minishlab\"><strong>\ud83e\udd17 Models</strong></a> |\n <a href=\"https://github.com/MinishLab/model2vec/tree/main/tutorials\"><strong>\ud83d\udcda Tutorials</strong></a> |\n <a href=\"https://minishlab.github.io/\"><strong>\ud83c\udf10 Blog</strong></a> |\n <a href=\"https://github.com/MinishLab/model2vec/blob/main/results/README.md\"><strong>\ud83c\udfc6 Results</strong></a> |\n <a href=\"https://github.com/MinishLab/model2vec/blob/main/docs\"><strong>\ud83d\udcd6 Docs</strong></a>\n </h2>\n</div>\n\n<div align=\"center\">\n <h2>\n <a href=\"https://pypi.org/project/model2vec/\"><img src=\"https://img.shields.io/pypi/v/model2vec?color=%23007ec6&label=pypi%20package\" alt=\"Package version\"></a>\n <a href=\"https://pypi.org/project/model2vec/\"><img src=\"https://img.shields.io/pypi/pyversions/model2vec\" alt=\"Supported Python versions\"></a>\n <a href=\"https://pepy.tech/project/model2vec\">\n <img src=\"https://static.pepy.tech/badge/model2vec\" alt=\"Downloads\">\n </a>\n <a href=\"https://app.codecov.io/gh/MinishLab/model2vec\">\n <img src=\"https://codecov.io/gh/MinishLab/model2vec/graph/badge.svg?token=21TWJ6B5ET\" alt=\"Codecov\">\n </a>\n </a>\n <a href=\"https://github.com/MinishLab/model2vec/blob/main/LICENSE\"><img src=\"https://img.shields.io/badge/license-MIT-green\" alt=\"License - MIT\"></a>\n\n </h2>\n</div>\n\n\n\n\nModel2Vec is a technique to turn any sentence transformer into a really small static model, reducing model size by a factor up to 50 and making the models up to 500 times faster, with a small drop in performance. Our [best model](https://huggingface.co/minishlab/potion-base-8M) is the most performant static embedding model in the world. See our results [here](results/README.md), or dive in to see how it works.\n\n<div align=\"center\">\n<h3>\n\n[Quickstart](#quickstart) \u2022 [Updates & Announcements](#updates--announcements) \u2022 [Main Features](#main-features) \u2022 [Model List](#model-list)\n</h3>\n</div>\n\n## Quickstart\n\nInstall the lightweight base package with:\n\n```bash\npip install model2vec\n```\n\nYou can start using Model2Vec by loading one of our [flagship models from the HuggingFace hub](https://huggingface.co/collections/minishlab/potion-6721e0abd4ea41881417f062). These models are pre-trained and ready to use. The following code snippet shows how to load a model and make embeddings, which you can use for any task, such as text classification, retrieval, clustering, or building a RAG system:\n```python\nfrom model2vec import StaticModel\n\n# Load a model from the HuggingFace hub (in this case the potion-base-8M model)\nmodel = StaticModel.from_pretrained(\"minishlab/potion-base-8M\")\n\n# Make embeddings\nembeddings = model.encode([\"It's dangerous to go alone!\", \"It's a secret to everybody.\"])\n\n# Make sequences of token embeddings\ntoken_embeddings = model.encode_as_sequence([\"It's dangerous to go alone!\", \"It's a secret to everybody.\"])\n```\n\nInstead of using one of our models, you can also distill your own Model2Vec model from a Sentence Transformer model. First, install the `distillation` extras with:\n\n```bash\npip install model2vec[distill]\n```\n\n\n Then, you can distill a model in ~30 seconds on a CPU with the following code snippet:\n\n```python\nfrom model2vec.distill import distill\n\n# Distill a Sentence Transformer model, in this case the BAAI/bge-base-en-v1.5 model\nm2v_model = distill(model_name=\"BAAI/bge-base-en-v1.5\", pca_dims=256)\n\n# Save the model\nm2v_model.save_pretrained(\"m2v_model\")\n```\n\nAfter distillation, you can also fine-tune your own classification models on top of the distilled model, or on a pre-trained model. First, make sure you install the `training` extras with:\n\n```bash\npip install model2vec[training]\n```\n\nThen, you can fine-tune a model as follows:\n\n```python\nimport numpy as np\nfrom datasets import load_dataset\nfrom model2vec.train import StaticModelForClassification\n\n# Initialize a classifier from a pre-trained model\nclassifier = StaticModelForClassification.from_pretrained(model_name=\"minishlab/potion-base-32M\")\n\n# Load a dataset\nds = load_dataset(\"setfit/subj\")\n\n# Train the classifier on text (X) and labels (y)\nclassifier.fit(ds[\"train\"][\"text\"], ds[\"train\"][\"label\"])\n\n# Evaluate the classifier\npredictions = classifier.predict(ds[\"test\"][\"text\"])\naccuracy = np.mean(np.array(predictions) == np.array(ds[\"test\"][\"label\"])) * 100\n```\n\nFor advanced usage, please refer to our [usage documentation](https://github.com/MinishLab/model2vec/blob/main/docs/usage.md).\n\n## Updates & Announcements\n\n- **12/02/2024**: We released **Model2Vec training**, allowing you to fine-tune your own classification models on top of Model2Vec models. Find out more in our [training documentation](https://github.com/MinishLab/model2vec/blob/main/model2vec/train/README.md) and [results](results/README.md#training-results).\n\n- **30/01/2024**: We released two new models: [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) and [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M). [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) is our most performant model to date, using a larger vocabulary and higher dimensions. [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M) is a finetune of [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) that is optimized for retrieval tasks, and is the best performing static retrieval model currently available.\n\n- **30/10/2024**: We released three new models: [potion-base-8M](https://huggingface.co/minishlab/potion-base-8M), [potion-base-4M](https://huggingface.co/minishlab/potion-base-4M), and [potion-base-2M](https://huggingface.co/minishlab/potion-base-2M). These models are trained using [Tokenlearn](https://github.com/MinishLab/tokenlearn). Find out more in our [blog post](https://minishlab.github.io/tokenlearn_blogpost/). NOTE: for users of any of our old English M2V models, we recommend switching to these new models as they [perform better on all tasks](https://github.com/MinishLab/model2vec/tree/main/results).\n\n## Main Features\n\n- **State-of-the-Art Performance**: Model2Vec models outperform any other static embeddings (such as GLoVe and BPEmb) by a large margin, as can be seen in our [results](results/README.md).\n- **Small**: Model2Vec reduces the size of a Sentence Transformer model by a factor of up to 50. Our [best model](https://huggingface.co/minishlab/potion-base-8M) is just ~30 MB on disk, and our smallest model just ~8 MB (making it the smallest model on [MTEB](https://huggingface.co/spaces/mteb/leaderboard)!).\n- **Lightweight Dependencies**: the base package's only major dependency is `numpy`.\n- **Lightning-fast Inference**: up to 500 times faster on CPU than the original model.\n- **Fast, Dataset-free Distillation**: distill your own model in 30 seconds on a CPU, without a dataset.\n- **Fine-tuning**: fine-tune your own classification models on top of Model2Vec models.\n- **Integrated in many popular libraries**: Model2Vec is integrated direclty into popular libraries such as [Sentence Transformers](https://github.com/UKPLab/sentence-transformers) and [LangChain](https://github.com/langchain-ai/langchain). For more information, see our [integrations documentation](https://github.com/MinishLab/model2vec/blob/main/docs/integrations.md).\n- **Tightly integrated with HuggingFace hub**: easily share and load models from the HuggingFace hub, using the familiar `from_pretrained` and `push_to_hub`. Our own models can be found [here](https://huggingface.co/minishlab).\n\n## What is Model2Vec?\n\nModel2vec creates a small, fast, and powerful model that outperforms other static embedding models by a large margin on all tasks we could find, while being much faster to create than traditional static embedding models such as GloVe. Like BPEmb, it can create subword embeddings, but with much better performance. Distillation doesn't need _any_ data, just a vocabulary and a model.\n\nThe core idea is to forward pass a vocabulary through a sentence transformer model, creating static embeddings for the indiviudal tokens. After this, there are a number of post-processing steps we do that results in our best models. For a more extensive deepdive, please refer to the following resources:\n- Our initial [Model2Vec blog post](https://huggingface.co/blog/Pringled/model2vec). Note that, while this post gives a good overview of the core idea, we've made a number of substantial improvements since then.\n- Our [Tokenlearn blog post](https://minishlab.github.io/tokenlearn_blogpost/). This post describes the Tokenlearn method we used to train our [potion models](https://huggingface.co/collections/minishlab/potion-6721e0abd4ea41881417f062).\n- Our official [documentation](https://github.com/MinishLab/model2vec/blob/main/docs/what_is_model2vec.md). This document provides a high-level overview of how Model2Vec works.\n\n## Documentation\n\nOur official documentation can be found [here](https://github.com/MinishLab/model2vec/blob/main/docs/README.md). This includes:\n- [Usage documentation](https://github.com/MinishLab/model2vec/blob/main/docs/usage.md): provides a technical overview of how to use Model2Vec.\n- [Integrations documentation](https://github.com/MinishLab/model2vec/blob/main/docs/integrations.md): provides examples of how to use Model2Vec in various downstream libraries.\n- [Model2Vec technical documentation](https://github.com/MinishLab/model2vec/blob/main/docs/what_is_model2vec.md): provides a high-level overview of how Model2Vec works.\n\n\n## Model List\n\nWe provide a number of models that can be used out of the box. These models are available on the [HuggingFace hub](https://huggingface.co/collections/minishlab/model2vec-base-models-66fd9dd9b7c3b3c0f25ca90e) and can be loaded using the `from_pretrained` method. The models are listed below.\n\n\n\n| Model | Language | Sentence Transformer | Params | Task |\n|-----------------------------------------------------------------------|------------|-----------------------------------------------------------------|---------|-----------|\n| [potion-base-32M](https://huggingface.co/minishlab/potion-base-32M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 32.3M | General |\n| [potion-base-8M](https://huggingface.co/minishlab/potion-base-8M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 7.5M | General |\n| [potion-base-4M](https://huggingface.co/minishlab/potion-base-4M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 3.7M | General |\n| [potion-base-2M](https://huggingface.co/minishlab/potion-base-2M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 1.8M | General |\n| [potion-retrieval-32M](https://huggingface.co/minishlab/potion-retrieval-32M) | English | [bge-base-en-v1.5](https://huggingface.co/BAAI/bge-base-en-v1.5) | 32.3M | Retrieval |\n| [M2V_multilingual_output](https://huggingface.co/minishlab/M2V_multilingual_output) | Multilingual | [LaBSE](https://huggingface.co/sentence-transformers/LaBSE) | 471M | General |\n\n\n## Results\n\nWe have performed extensive experiments to evaluate the performance of Model2Vec models. The results are documented in the [results](results/README.md) folder. The results are presented in the following sections:\n- [MTEB Results](results/README.md#mteb-results)\n- [Training Results](results/README.md#training-results)\n- [Ablations](results/README.md#ablations)\n\n## License\n\nMIT\n\n## Citing\n\nIf you use Model2Vec in your research, please cite the following:\n```bibtex\n@software{minishlab2024model2vec,\n authors = {Stephan Tulkens and Thomas van Dongen},\n title = {Model2Vec: Fast State-of-the-Art Static Embeddings},\n year = {2024},\n url = {https://github.com/MinishLab/model2vec}\n}\n```\n",
"bugtrack_url": null,
"license": "MIT License\n \n Copyright (c) 2024 Thomas van Dongen\n \n Permission is hereby granted, free of charge, to any person obtaining a copy\n of this software and associated documentation files (the \"Software\"), to deal\n in the Software without restriction, including without limitation the rights\n to use, copy, modify, merge, publish, distribute, sublicense, and/or sell\n copies of the Software, and to permit persons to whom the Software is\n furnished to do so, subject to the following conditions:\n \n The above copyright notice and this permission notice shall be included in all\n copies or substantial portions of the Software.\n \n THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE\n SOFTWARE.\n ",
"summary": "Fast State-of-the-Art Static Embeddings",
"version": "0.4.0",
"project_urls": {
"Bug Reports": "https://github.com/MinishLab/model2vec/issues",
"Homepage": "https://github.com/MinishLab",
"Source": "https://github.com/MinishLab/model2vec"
},
"split_keywords": [],
"urls": [
{
"comment_text": null,
"digests": {
"blake2b_256": "937d39ff093c4e45303a06e3c5825c6144cbd21f18a1393a154bbf93232b0f1a",
"md5": "bb3dff38d1a33f821f6a8fd22a33ea15",
"sha256": "df30685a55841c61c6638e4f329648e76b148507bd778801d7bfcd6b970a4f2f"
},
"downloads": -1,
"filename": "model2vec-0.4.0-py3-none-any.whl",
"has_sig": false,
"md5_digest": "bb3dff38d1a33f821f6a8fd22a33ea15",
"packagetype": "bdist_wheel",
"python_version": "py3",
"requires_python": ">=3.9",
"size": 38593,
"upload_time": "2025-02-12T19:49:35",
"upload_time_iso_8601": "2025-02-12T19:49:35.486317Z",
"url": "https://files.pythonhosted.org/packages/93/7d/39ff093c4e45303a06e3c5825c6144cbd21f18a1393a154bbf93232b0f1a/model2vec-0.4.0-py3-none-any.whl",
"yanked": false,
"yanked_reason": null
},
{
"comment_text": null,
"digests": {
"blake2b_256": "83e23fb7bd8c612f71ad3abded92e7401f97f1e71427d3a68a3fb85f39394b17",
"md5": "d8cc2bc22e050e7a62c4b7aad3569ff0",
"sha256": "48d4a3da040499b0090f736eb8f22ea0fdd35b67462d81d789c70004423adbae"
},
"downloads": -1,
"filename": "model2vec-0.4.0.tar.gz",
"has_sig": false,
"md5_digest": "d8cc2bc22e050e7a62c4b7aad3569ff0",
"packagetype": "sdist",
"python_version": "source",
"requires_python": ">=3.9",
"size": 2486998,
"upload_time": "2025-02-12T19:49:38",
"upload_time_iso_8601": "2025-02-12T19:49:38.414873Z",
"url": "https://files.pythonhosted.org/packages/83/e2/3fb7bd8c612f71ad3abded92e7401f97f1e71427d3a68a3fb85f39394b17/model2vec-0.4.0.tar.gz",
"yanked": false,
"yanked_reason": null
}
],
"upload_time": "2025-02-12 19:49:38",
"github": true,
"gitlab": false,
"bitbucket": false,
"codeberg": false,
"github_user": "MinishLab",
"github_project": "model2vec",
"travis_ci": false,
"coveralls": false,
"github_actions": true,
"lcname": "model2vec"
}