Check out the full blogpost if you'd like to 1) use these lightning-fast models or 2) learn how to train them with consumer-level hardware: https://hf.co/blog/static-embeddings
Or read more in this thread first ๐งต
Or read more in this thread first ๐งต
Comments
2๏ธโฃ an English Retrieval model and a general-purpose Multilingual similarity model (classification, clustering, etc.), both Apache 2.0. Fully integrated in Sentence Transformers, etc.
๐งต
๐ my training scripts, using the Sentence Transformers library
๐ my Weights & Biases reports with losses & metrics
๐ my list of 30 training and 13 evaluation datasets
๐งต
I apply this simple architecture, but train it like a modern embedding model: Contrastive Learning with Matryoshka support.
๐งต
๐๏ธ Extremely fast, e.g. 107500 sentences per second on a consumer CPU, compared to 270 for all-mpnet-base-v2 and 56 for gte-large-en-v1.5
๐ No maximum sequence length! Embed texts at any length (at your own risk)
๐งต
๐ช Matryoshka support: allow you to truncate embeddings with minimal performance loss (e.g. 4x smaller with a 0.56% performance decrease for English Similarity tasks)
๐งต
- @langchain.bsky.social