An “embeddings model” is trained to convert a piece of text into a vector, which can later be rapidly compared to other vectors to determine similarity between the pieces of text. Embeddings models are typically much smaller than LLMs, and will be extremely fast and cheap in comparison. In Continue, embeddings are generated during indexing and then used by codebase awareness to perform similarity search over your codebase. You can addDocumentation Index
Fetch the complete documentation index at: https://continue-docs-dependabot-npm-and-yarn-docs-multi-c8c89d9539.mintlify.app/llms.txt
Use this file to discover all available pages before exploring further.
embed to a model’s roles to specify that it can be used to embed.
[Built-in model (VS Code only)]
transformers.js is used as a built-in
embeddings model in VS Code. In JetBrains, there currently is no built-in
embedder.Recommended embedding models
See our comprehensive model recommendations for the best embedding models comparison.
voyage-code-3, which is listed below along with the rest of the options for embeddings models.
If you want to generate embeddings locally, we recommend using nomic-embed-text with Ollama.
Voyage AI
After obtaining an API key from here, you can configure like this:- Hub
- YAML
- JSON
Ollama
See here for instructions on how to use Ollama for embeddings.Transformers.js (currently VS Code only)
Transformers.js is a JavaScript port of the popular Transformers library. It allows embeddings to be calculated entirely locally. The model used isall-MiniLM-L6-v2, which is shipped alongside the Continue extension.
- YAML
- JSON
config.yaml
Text Embeddings Inference
Hugging Face Text Embeddings Inference enables you to host your own embeddings endpoint. You can configure embeddings to use your endpoint as follows:- YAML
- JSON
config.yaml