bge_1-5_query_embeddings

Maintainer: center-for-curriculum-redesign

Total Score

5

Last updated 9/19/2024
AI model preview image
PropertyValue
Run this modelRun on Replicate
API specView on Replicate
Github linkView on Github
Paper linkView on Arxiv

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

The bge_1-5_query_embeddings model is a query embedding generator developed by the Center for Curriculum Redesign. It is built on top of BAAI's bge-large-en v1.5 embedding model, which is a powerful text encoding model for embedding text sequences. Similar models include the bge-large-en-v1.5 model, the bge-reranker-base model, and the multilingual-e5-large model.

Model inputs and outputs

The bge_1-5_query_embeddings model takes in a list of text queries and generates corresponding embedding vectors for retrieval and comparison purposes. The model automatically formats the input queries for retrieval, so users do not need to preprocess the text.

Inputs

  • Query Texts: A serialized JSON array of strings to be used as text queries for generating embeddings.
  • Normalize: A boolean flag to control whether the output embeddings are normalized to a magnitude of 1.
  • Precision: The numerical precision to use for the inference computations, either "full" or "half".
  • Batchtoken Max: The maximum number of kibiTokens (1 kibiToken = 1024 tokens) to include in a single batch, to avoid out-of-memory errors.

Outputs

  • Query Embeddings: An array of embedding vectors, where each vector corresponds to one of the input text queries.
  • Extra Metrics: Additional metrics or data associated with the embedding generation process.

Capabilities

The bge_1-5_query_embeddings model is capable of generating high-quality text embeddings that can be used for a variety of natural language processing tasks, such as information retrieval, text similarity comparison, and document clustering. The embeddings capture the semantic meaning of the input text, allowing for more effective downstream applications.

What can I use it for?

The bge_1-5_query_embeddings model can be used in a wide range of applications that require text encoding and comparison, such as search engines, recommendation systems, and content analysis tools. By generating embeddings for text queries, you can leverage the model's powerful encoding capabilities to improve the relevance and accuracy of your search or recommendation results.

Things to try

One interesting thing to try with the bge_1-5_query_embeddings model is to experiment with different levels of precision for the inference computations. Depending on your specific use case and hardware constraints, you may find that the "half" precision setting provides sufficient accuracy while requiring less computational resources. Additionally, you could explore how the model's performance varies when using different normalization strategies for the output embeddings.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

AI model preview image

bge-large-en-v1.5

nateraw

Total Score

202

The bge-large-en-v1.5 is a text embedding model created by BAAI (Beijing Academy of Artificial Intelligence). It is designed to generate high-quality embeddings for text sequences in English. This model builds upon BAAI's previous work on the bge-reranker-base and multilingual-e5-large models, which have shown strong performance on various language tasks. The bge-large-en-v1.5 model offers enhanced capabilities and is well-suited for a range of natural language processing applications. Model inputs and outputs The bge-large-en-v1.5 model takes text sequences as input and generates corresponding embeddings. Users can provide the text either as a path to a file containing JSONL data with a 'text' field, or as a JSON list of strings. The model also accepts a batch size parameter to control the processing of the input data. Additionally, users can choose to normalize the output embeddings and convert the results to a NumPy format. Inputs Path**: Path to a file containing text as JSONL with a 'text' field or a valid JSON string list. Texts**: Text to be embedded, formatted as a JSON list of strings. Batch Size**: Batch size to use when processing the text data. Convert To Numpy**: Option to return the output as a NumPy file instead of JSON. Normalize Embeddings**: Option to normalize the generated embeddings. Outputs The model outputs the text embeddings, which can be returned either as a JSON array or as a NumPy file, depending on the user's preference. Capabilities The bge-large-en-v1.5 model is capable of generating high-quality text embeddings that capture the semantic and contextual meaning of the input text. These embeddings can be utilized in a wide range of natural language processing tasks, such as text classification, semantic search, and content recommendation. The model's performance has been demonstrated in various benchmarks and real-world applications. What can I use it for? The bge-large-en-v1.5 model can be a valuable tool for developers and researchers working on natural language processing projects. The text embeddings generated by the model can be used as input features for downstream machine learning models, enabling more accurate and efficient text-based applications. For example, the embeddings could be used in sentiment analysis, topic modeling, or to power personalized content recommendations. Things to try To get the most out of the bge-large-en-v1.5 model, you can experiment with different input text formats, batch sizes, and normalization options to find the configuration that works best for your specific use case. You can also explore how the model's performance compares to other similar models, such as the bge-reranker-base and multilingual-e5-large models, to determine the most suitable approach for your needs.

Read more

Updated Invalid Date

AI model preview image

bge-reranker-base

ninehills

Total Score

8

The bge-reranker-base model from BAAI (Beijing Academy of Artificial Intelligence) is a cross-encoder model that can be used to re-rank the top-k documents returned by an embedding model. It is more accurate than embedding models like BGE-M3 or LLM Embedder, but less efficient. This model can be fine-tuned on your own data to improve performance on specific tasks. Model inputs and outputs Inputs pairs_json**: A JSON string containing input pairs, e.g. [["a", "b"], ["c", "d"]] Outputs scores**: An array of scores for the input pairs use_fp16**: A boolean indicating whether the model used FP16 inference model_name**: The name of the model used Capabilities The bge-reranker-base model can effectively re-rank the top-k documents returned by an embedding model, making the final ranking more accurate. This can be particularly useful when you need high-precision retrieval results, such as for question answering or knowledge-intensive tasks. What can I use it for? You can use the bge-reranker-base model to re-rank the results of an embedding model like BGE-M3 or LLM Embedder. This can help improve the accuracy of your retrieval system, especially for critical applications where precision is important. Things to try You can try fine-tuning the bge-reranker-base model on your own data to further improve its performance on your specific use case. The examples provided can be a good starting point for this.

Read more

Updated Invalid Date

🌀

bge-small-zh-v1.5

BAAI

Total Score

43

The bge-small-zh-v1.5 model from BAAI is a small-scale version of the BAAI General Embedding (BGE) model, which can map any text to a low-dimensional dense vector. Unlike previous BGE models, version 1.5 has a more reasonable similarity distribution, enhancing its retrieval ability without the need for instruction. The bge-small-zh-v1.5 model is competitive in performance compared to larger models, making it a good option for projects with computational constraints. Model inputs and outputs The bge-small-zh-v1.5 model takes in text as input and outputs a fixed-size embedding vector. This embedding can then be used for tasks like retrieval, classification, clustering, or semantic search. The model supports both Chinese and English text. Inputs Text**: The model can accept any Chinese or English text as input. Outputs Embedding vector**: The model outputs a fixed-size vector representation of the input text, which can be used for downstream tasks. Capabilities The bge-small-zh-v1.5 model is capable of generating high-quality text embeddings that can be used for a variety of natural language processing tasks. Its performance is competitive with larger BGE models, making it a good choice for projects with limited computational resources. The model's improved similarity distribution helps to better differentiate between similar and dissimilar text. What can I use it for? The bge-small-zh-v1.5 embedding can be used in a wide range of applications, such as: Semantic search**: Use the embeddings to find relevant passages or documents for a given query. Text classification**: Train a classifier on top of the embeddings to categorize text into different classes. Clustering**: Group similar text together based on the embeddings. Recommendation systems**: Use the embeddings to find similar items or content for recommendation. Things to try One interesting thing to try with the bge-small-zh-v1.5 model is to fine-tune it on your specific data and task. The examples provided by the maintainers show how to prepare data and fine-tune the model to improve performance on your use case. Additionally, you can experiment with using the model in conjunction with the provided reranker models to further enhance retrieval performance.

Read more

Updated Invalid Date

🌀

bge-base-zh-v1.5

BAAI

Total Score

51

The bge-base-zh-v1.5 model is a text embedding model developed by BAAI (Beijing Academy of Artificial Intelligence). It is part of the BAAI General Embedding (BGE) family of models, which can map any text to a low-dimensional dense vector. This can be used for tasks like retrieval, classification, clustering, or semantic search. The bge-base-zh-v1.5 model is the Chinese version of the base-scale BGE model, updated to version 1.5 to have a more reasonable similarity distribution compared to previous versions. The bge-base-zh-v1.5 model is similar in capability to the BAAI/bge-large-zh-v1.5 model, which is the large-scale Chinese BGE model, but the base-scale model has a smaller embedding size. The BAAI/bge-small-zh-v1.5 model is an even smaller-scale Chinese BGE model, with further reduced embedding size but still competitive performance. Model inputs and outputs Inputs Text**: The model can take any text as input, such as short queries or long passages. Outputs Embeddings**: The model outputs a low-dimensional dense vector representation (embedding) of the input text. Capabilities The bge-base-zh-v1.5 model can effectively map Chinese text to a semantic embedding space. It achieves state-of-the-art performance on the Chinese Massive Text Embedding Benchmark (C-MTEB), ranking 1st in multiple evaluation tasks. What can I use it for? The bge-base-zh-v1.5 embedding model can be used in a variety of natural language processing applications that require semantic understanding of text, such as: Retrieval**: Use the embeddings to find the most relevant passages or documents for a given query. Classification**: Train a classifier on top of the embeddings to categorize text into different classes. Clustering**: Group similar text together based on the proximity of their embeddings. Semantic search**: Find documents or passages that are semantically similar to a given query. The model can also be integrated into vector databases to support retrieval-augmented large language models (LLMs). Things to try One interesting aspect of the bge-base-zh-v1.5 model is that it has improved retrieval performance without using any instruction in the query, compared to previous versions that required an instruction. This makes it more convenient to use in many applications. You can experiment with using the model with and without instructions to see which setting works best for your specific task. Additionally, you can try fine-tuning the bge-base-zh-v1.5 model on your own data using the provided examples. This can help improve the model's performance on your domain-specific tasks.

Read more

Updated Invalid Date