airoboros-13b

Maintainer: jondurbin

Total Score

106

Last updated 5/28/2024

🐍

PropertyValue
Run this modelRun on HuggingFace
API specView on HuggingFace
Github linkNo Github link provided
Paper linkNo paper link provided

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

The airoboros-13b model is a fine-tuned 13-billion parameter LlaMa model, using completely synthetic training data created by jondurbin. It is an experimental model that the maintainer does not recommend using, as the outputs are not particularly great and it may contain "harmful" data due to jailbreaking. For a much better experience, the maintainer recommends using one of the updated airoboros models.

The model was evaluated by GPT-4, which provided scores across a variety of tasks. Compared to models like GPT-3.5, GPT-4-x-Alpasta-30B, Manticore-13B, Vicuna-13B-1.1, and Wizard-Vicuna-13B-Uncensored, the airoboros-13b model scored relatively lower on average, with some specific task scores being quite low.

Model inputs and outputs

Inputs

  • Text prompts: The model accepts text prompts as input, which can be used to generate continuations or responses.

Outputs

  • Text generation: The primary output of the model is the generation of text, which can be used for a variety of natural language processing tasks.

Capabilities

The airoboros-13b model is capable of generating text, but the maintainer has warned that the outputs are not particularly great and may contain "harmful" data. The model was outperformed by several other similar models across a range of tasks, as evidenced by the evaluation scores provided.

What can I use it for?

Given the maintainer's warning about the quality of the outputs and potential for harmful content, it is not recommended to use the airoboros-13b model for any production or real-world applications. The maintainer suggests using one of the updated airoboros models instead, which are likely to provide a much better experience.

Things to try

As the maintainer has advised against using this model, it is not recommended to experiment with it. Instead, users interested in similar language models should explore the updated airoboros models or other state-of-the-art models that have been more thoroughly vetted and validated.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

🎲

airoboros-l2-70b-gpt4-1.4.1

jondurbin

Total Score

48

The airoboros-l2-70b-gpt4-1.4.1 model is a large language model developed by Jon Durbin. It is a fine-tuned Llama 2 model with 70 billion parameters, trained on a dataset generated using the airoboros tool. This model builds on the previous Llama 65B GPT4 1.4 model from the same creator. Similar models include the airoboros-l2-70b-gpt4-1.4.1-GPTQ model, which provides GPTQ-quantized versions of the model for improved efficiency. Model inputs and outputs Inputs The model takes in text prompts as input. Outputs The model generates coherent and contextual text as output. Capabilities The airoboros-l2-70b-gpt4-1.4.1 model is capable of generating human-like text across a wide range of topics. It can be used for tasks like language generation, text summarization, and question answering. The model has been fine-tuned to provide more detailed and informative responses compared to the previous Llama 65B model. What can I use it for? The airoboros-l2-70b-gpt4-1.4.1 model could be used for a variety of natural language processing applications, such as chatbots, content generation, or data analysis. However, due to the licensing restrictions around the use of the OpenAI API data used to fine-tune the model, it is recommended to avoid using this model commercially. The model could be a useful research tool or for personal/non-commercial projects. Things to try Experiment with different prompting techniques to see how the model responds to more specific or creative inputs. Try providing the model with instructions or tasks to see how it can follow directions and generate relevant text. Additionally, explore the model's ability to engage in conversational exchanges by providing it with back-and-forth dialogues.

Read more

Updated Invalid Date

⛏️

airoboros-m-7b-3.1.2

jondurbin

Total Score

47

The airoboros-m-7b-3.1.2 model is an experimental language model created by jondurbin. It is based on the Mistral-7B-v0.1 base model and uses a dataset generated by the airoboros tool. This model focuses heavily on instruction following and has capabilities in areas like math problem solving, log information extraction, anonymization, and chat introspection. Model Inputs and Outputs Inputs The model accepts prompts in the "llama-2 chat format", which includes a system message and a user message. The system message sets up the assistant as a "helpful, unbiased, uncensored assistant", while the user message contains the actual prompt. For multi-turn conversations, the prompt format includes the previous prompts and responses. Outputs The model generates a response to the user's prompt, which can range from a few sentences to multiple paragraphs. The output can include structured data like MathJSON solutions, which are formatted between ` and ` tags. Capabilities The airoboros-m-7b-3.1.2 model excels at following instructions and generating helpful, detailed responses. It has been trained on a variety of tasks like math problem solving, log information extraction, and anonymization. The model can also handle multi-turn conversations and respond to questions about the chat history or characters. What Can I Use It For? This model could be useful for applications that require a versatile, instruction-following assistant. Some potential use cases include: Building chatbots or virtual assistants that can handle a wide range of tasks Automating data processing and extraction tasks from logs or other structured data Generating anonymized versions of text data for privacy-preserving applications Things to Try One interesting aspect of this model is its ability to handle structured data like MathJSON. You could try prompting the model with math problems and see how it generates the corresponding JSON solution. Another interesting area to explore is the model's multi-turn conversation capabilities - you could engage it in a back-and-forth dialogue and see how it maintains context and responds to questions about the chat history.

Read more

Updated Invalid Date

🤔

airoboros-l2-70B-gpt4-1.4.1-GPTQ

TheBloke

Total Score

57

airoboros-l2-70B-gpt4-1.4.1-GPTQ is a large language model created by TheBloke, a prominent AI model developer. This model is a quantized version of the original Airoboros Llama 2 70B GPT4 1.4.1 model, using the GPTQ technique to reduce the model size and memory footprint. The goal is to enable deployment on a wider range of hardware, from GPUs to CPUs. Similar quantized models provided by TheBloke include the Llama-2-70B-GPTQ, Llama-2-7B-GPTQ, and Llama-2-13B-GPTQ models, all based on Meta's open-source Llama 2 language models. Model inputs and outputs Inputs Text**: The model accepts natural language text as input, which it can then use to generate additional text. Outputs Generated text**: The model's primary output is new text that it generates based on the input prompt. This can range from short responses to long-form content. Capabilities The airoboros-l2-70B-gpt4-1.4.1-GPTQ model has been trained on a large corpus of text data, giving it broad knowledge and the ability to engage in a wide variety of language tasks. It can be used for tasks like question answering, summarization, translation, and open-ended text generation. The quantization process has reduced the model's size and memory footprint, enabling deployment on a wider range of hardware. What can I use it for? This model could be useful for developing chatbots, content generation tools, language learning applications, and other natural language processing projects. Developers could integrate it into their applications to provide intelligent language capabilities. Businesses could leverage it to automate text-based workflows, generate marketing content, or provide customer support. Things to try One interesting aspect of this model is the inclusion of an "uncensored" version, the llama2_70b_chat_uncensored-GPTQ model. This variant was fine-tuned on an uncensored dataset, providing more unfiltered and potentially controversial responses. Developers could experiment with this model to explore the challenges and implications of deploying such language models. Another idea would be to fine-tune the base airoboros-l2-70B-gpt4-1.4.1-GPTQ model on a specific domain or task, such as technical writing, legal analysis, or creative fiction, to see how it performs in those specialized areas.

Read more

Updated Invalid Date

↗️

OpenOrca-Platypus2-13B-GPTQ

TheBloke

Total Score

49

The OpenOrca-Platypus2-13B-GPTQ is a large language model created by Open-Orca and refined by TheBloke. It is based on the Llama 2 architecture and has been trained on a combination of the OpenOrca dataset and a custom dataset focused on STEM and logic tasks. This model builds on the previous OpenOrca Platypus2 13B model, incorporating improvements to its performance and capabilities. The OpenOrca-Platypus2-13B-GPTQ model is available in various quantized versions optimized for different hardware and performance requirements. These include 2-bit, 3-bit, 4-bit, 5-bit, 6-bit, and 8-bit GPTQ models, as well as 2-8 bit GGUF models for CPU and GPU inference. Model inputs and outputs Inputs Prompts**: The model takes in natural language prompts that describe a task or request. Instructions**: The model can also accept structured instruction-based prompts, such as the Alpaca-InstructOnly format. Outputs Text generation**: The primary output of the model is generated text, which can range from short responses to long-form narratives. Task completion**: The model is capable of understanding and completing a variety of tasks described in the input prompts. Capabilities The OpenOrca-Platypus2-13B-GPTQ model excels at a wide range of language tasks, including creative writing, question answering, code generation, and more. It has demonstrated strong performance on various benchmarks, including the HuggingFace Leaderboard, AGIEval, and BigBench-Hard. Compared to the original OpenOrca Platypus2 13B model, this version offers improved performance, lower hallucination rates, and longer responses. What can I use it for? The OpenOrca-Platypus2-13B-GPTQ model can be used for a variety of applications, such as: Content generation**: Create engaging stories, articles, or product descriptions. Conversational AI**: Build chatbots and virtual assistants that can engage in natural language interactions. Task completion**: Develop applications that can understand and complete complex instructions, such as code generation, math problem-solving, or creative tasks. Research and development**: Use the model as a starting point for further fine-tuning or as a benchmark for comparing language model performance. Things to try One interesting aspect of the OpenOrca-Platypus2-13B-GPTQ model is its ability to generate long, detailed responses while maintaining coherence and factual accuracy. You can try providing the model with open-ended prompts or instructions and see how it responds. For example, you could ask it to write a story about llamas or solve a complex logic puzzle. Another avenue to explore is the model's performance on specialized tasks, such as technical writing, scientific analysis, or legal document review. By fine-tuning the model on domain-specific data, you may be able to unlock new capabilities that are tailored to your specific needs. Verifying the responses for safety and factual accuracy is also an important consideration when using large language models. Developing robust testing and monitoring procedures can help ensure the model is behaving as expected and not producing harmful or inaccurate outputs.

Read more

Updated Invalid Date