Embedding Models

Discover the embedding models supported by Hyper.


Hyper enables the use of a variety of embedding models, each selected for its strength in processing different data types, from text to images, ensuring applications can leverage the most appropriate capabilities for their data analysis and query needs.

Supported Models

Hyper's model support is designed to cater to a wide range of use cases, facilitating the integration of models best suited to the specific characteristics of the data being processed.

Text Embeddings

ModelDeveloperEmbedding SizeHyper Slug
ada v2OpenAI1536OPENAI_ADA_V2
CurieOpenAI4096OPENAI_CURIE
BERT BaseGoogle768GOOGLE_BERT_BASE
GPT-3 SmallOpenAI125M parametersOPENAI_GPT3_SMALL
DistilBERTHugging Face768HF_DISTILBERT

Image Embeddings

ModelDeveloperEmbedding SizeHyper Slug
CLIPOpenAI512OPENAI_CLIP
Vision Transformer (ViT)Google768GOOGLE_VIT

Usage

To utilize a specific embedding model in Hyper, specify the embedding_model parameter in the POST body for the /embeddings and other related API endpoints. Hyper defaults to using the OPENAI ada v2 model if no model is explicitly specified.

For vector searches, Hyper considers only the files that have embeddings generated with the model specified in the query. This ensures that the search results are relevant and accurately reflect the data's semantic meaning as processed by the chosen model.

For example, if you have files A and B with embeddings from OPENAI_ADA_V2 and files C and D with GOOGLE_BERT_BASE, specifying OPENAI_ADA_V2 as the embedding_model in your query will limit the search results to include only files A and B.

It is crucial that all files targeted in a search have embeddings generated with the same model to maintain consistency and accuracy in the results.