Static Embedding models have been around since before Transformers (e.g. GLoVe, word2vec), they work with pre-computed word embeddings from a mapping.
I apply this simple architecture, but train it like a modern embedding model: Contrastive Learning with Matryoshka support.
๐งต
I apply this simple architecture, but train it like a modern embedding model: Contrastive Learning with Matryoshka support.
๐งต
Comments
๐๏ธ Extremely fast, e.g. 107500 sentences per second on a consumer CPU, compared to 270 for all-mpnet-base-v2 and 56 for gte-large-en-v1.5
๐ No maximum sequence length! Embed texts at any length (at your own risk)
๐งต
๐ช Matryoshka support: allow you to truncate embeddings with minimal performance loss (e.g. 4x smaller with a 0.56% performance decrease for English Similarity tasks)
๐งต
- @langchain.bsky.social