Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GPTQ

Maintainer: TheBloke

Total Score

127

Last updated 5/27/2024

๐Ÿงช

PropertyValue
Run this modelRun on HuggingFace
API specView on HuggingFace
Github linkNo Github link provided
Paper linkNo paper link provided

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GPTQ model is a large language model created by TheBloke, who has generously provided a variety of quantized versions of the model for GPU and CPU inference. This model is based on Eric Hartford's Wizard Vicuna 13B Uncensored merged with Kaio Ken's SuperHOT 8K model. The key innovation is an increased context size of up to 8K, which is tested to work with ExLlama. TheBloke has also provided GPTQ and GGML quantized versions of the model for efficient inference on different hardware.

Model inputs and outputs

Inputs

  • Prompts: The model takes in free-form text prompts that can cover a wide range of topics. These prompts are used to initiate the model's generation of relevant and coherent responses.

Outputs

  • Generated text: The primary output of the model is free-form text, generated in response to the provided prompts. The model aims to produce helpful, detailed, and polite responses.

Capabilities

The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GPTQ model is a large, powerful language model that can be used for a variety of natural language processing tasks. It has been trained on a diverse dataset and can engage in open-ended conversations, answer questions, and generate human-like text on a wide range of subjects. The increased context size of up to 8K allows the model to maintain coherence and consistency over longer sequences.

What can I use it for?

This model could be useful for applications such as chatbots, virtual assistants, creative writing, summarization, and question-answering. The increased context size may be particularly beneficial for tasks that require maintaining context over longer interactions, such as task-oriented dialogues. Developers and researchers could explore using this model as a foundation for further fine-tuning or prompt engineering to create specialized AI applications.

Things to try

One interesting aspect of this model is the ability to control the generation process through parameters like temperature and top-k/top-p sampling. Experimenting with these settings can result in outputs with different levels of creativity, coherence, and diversity. Additionally, prompting the model with specific instructions or templates, as shown in the provided examples, can help elicit more targeted responses for certain use cases.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

๐Ÿง 

Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML

TheBloke

Total Score

68

The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML model is an uncensored version of Eric Hartford's Wizard Vicuna 13B model, with increased context length up to 8K using SuperHOT, a new system that employs RoPE to expand context. It was developed by kaiokendev. The model is available in GGML format for CPU and GPU inference using llama.cpp-compatible clients. Similar models include the Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GPTQ and Wizard-Vicuna-30B-Uncensored-GGML models created by the same maintainer, TheBloke. Model inputs and outputs The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML model is a text-to-text language model, taking in text prompts as input and generating text outputs. The model has been trained on a subset of the dataset used for the original Wizard Vicuna 13B model, with responses containing alignment or moralizing removed. Inputs Freeform text prompts that the model will use to generate a response. Outputs Coherent, multi-paragraph text responses generated by the model based on the input prompt. Capabilities The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML model can be used for a variety of natural language processing tasks, such as text generation, summarization, and question answering. The increased 8K context length enabled by SuperHOT allows the model to maintain coherence and consistency over longer passages of text. What can I use it for? The Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML model could be used for applications like creative writing, content generation, or task-oriented dialog. The uncensored nature of the model means it can generate more open-ended and less constrained text, which could be useful for certain use cases but also comes with increased responsibility. As with any powerful language model, users should be cautious about how they deploy and use this model. Things to try One interesting aspect of the Wizard-Vicuna-13B-Uncensored-SuperHOT-8K-GGML model is its ability to maintain context and coherence over longer passages of text due to the increased 8K context length. You could try providing the model with multi-paragraph prompts and see how it is able to generate coherent, extended responses that build upon the initial prompt. The model's uncensored nature also provides opportunities to explore more open-ended and creative applications, but users should be mindful of the potential risks and use responsibly.

Read more

Updated Invalid Date

๐Ÿ–ผ๏ธ

Wizard-Vicuna-13B-Uncensored-GPTQ

TheBloke

Total Score

302

The Wizard-Vicuna-13B-Uncensored-GPTQ is a large language model developed by Eric Hartford and maintained by TheBloke. It is a quantized version of the Wizard Vicuna 13B Uncensored model, using the GPTQ compression technique to reduce the model size while maintaining performance. This model is part of a suite of quantized models provided by TheBloke, including Wizard-Vicuna-30B-Uncensored-GPTQ and WizardLM-7B-uncensored-GPTQ. Model inputs and outputs The Wizard-Vicuna-13B-Uncensored-GPTQ model is a text-to-text model, capable of generating natural language responses given text prompts. The model follows the standard Vicuna prompt format, where the user's input is prefixed with "USER:" and the model's response is prefixed with "ASSISTANT:". Inputs Text prompts provided by the user, which the model uses to generate a response. Outputs Natural language text generated by the model in response to the user's input. Capabilities The Wizard-Vicuna-13B-Uncensored-GPTQ model is capable of engaging in open-ended dialogue, answering questions, and generating creative text. It has been fine-tuned to provide helpful, detailed, and polite responses, while avoiding harmful, unethical, or biased content. What can I use it for? The Wizard-Vicuna-13B-Uncensored-GPTQ model can be used for a variety of natural language processing tasks, such as building chatbots, virtual assistants, and text generation applications. Its large size and strong performance make it well-suited for tasks that require in-depth language understanding and generation. Developers can use this model as a starting point for further fine-tuning or deployment in their own applications. Things to try One interesting aspect of the Wizard-Vicuna-13B-Uncensored-GPTQ model is its ability to generate long, coherent responses. You can try providing the model with open-ended prompts and see how it develops a detailed, multi-paragraph answer. Additionally, you can experiment with different temperature and sampling settings to adjust the creativity and diversity of the model's outputs.

Read more

Updated Invalid Date

๐Ÿ

WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ

TheBloke

Total Score

47

The WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ model is a 13B parameter language model created by combining Eric Hartford's WizardLM 13B V1.0 Uncensored with Kaio Ken's SuperHOT 8K. The model has been quantized to 4-bit using the GPTQ-for-LLaMa tool, which allows for increased context size up to 8K tokens. This model is an experimental new GPTQ that offers expanded context compared to the original WizardLM 13B V1.0 Uncensored. Model inputs and outputs The WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ model takes text prompts as input and generates coherent, detailed responses. The model has been trained on a large corpus of online text data, allowing it to understand and converse on a wide range of topics. Inputs Text prompt**: A text prompt provided to the model to initiate the generation of a response. Outputs Generated text**: The model's response to the provided text prompt, which can be up to 8192 tokens in length. Capabilities The WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ model is a powerful language model capable of engaging in open-ended conversations, answering questions, and generating human-like text on a variety of subjects. Its expanded context size allows it to maintain coherence and provide more detailed responses compared to models with shorter context. What can I use it for? The WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ model can be used for a wide range of natural language processing tasks, such as chatbots, content generation, question answering, and creative writing. The increased context size makes it well-suited for applications that require longer-form, coherent responses. Things to try One interesting aspect of the WizardLM-13B-V1-0-Uncensored-SuperHOT-8K-GPTQ model is its ability to maintain context and narrative structure over longer text generation. Try providing the model with a multi-sentence prompt and see how it continues the story or expands on the initial ideas. The model's large knowledge base and generation capabilities make it well-suited for collaborative storytelling or worldbuilding exercises.

Read more

Updated Invalid Date

๐ŸŒ

Wizard-Vicuna-7B-Uncensored-GPTQ

TheBloke

Total Score

162

The Wizard-Vicuna-7B-Uncensored-GPTQ model is a quantized version of the open-source Wizard Vicuna 7B Uncensored language model created by Eric Hartford. It has been quantized using GPTQ techniques by TheBloke, who has provided several quantization options to choose from based on the user's hardware and performance requirements. Model inputs and outputs The Wizard-Vicuna-7B-Uncensored-GPTQ model is a text-to-text transformer model, which means it takes text as input and generates text as output. The input is typically a prompt or a partial message, and the output is the model's continuation or response. Inputs Text prompt or partial message Outputs Continued text, with the model responding to the input prompt in a contextual and coherent manner Capabilities The Wizard-Vicuna-7B-Uncensored-GPTQ model has broad language understanding and generation capabilities, allowing it to engage in open-ended conversations, answer questions, and assist with a variety of text-based tasks. It has been trained on a large corpus of text data, giving it the ability to produce human-like responses on a wide range of subjects. What can I use it for? The Wizard-Vicuna-7B-Uncensored-GPTQ model can be used for a variety of applications, such as building chatbots, virtual assistants, or creative writing tools. It could be used to generate responses for customer service inquiries, provide explanations for complex topics, or even help with ideation and brainstorming. Given its uncensored nature, users should exercise caution and responsibility when using this model. Things to try Users can experiment with the model by providing it with prompts on different topics and observing the generated responses. They can also try adjusting the temperature and other sampling parameters to see how it affects the creativity and coherence of the output. Additionally, users may want to explore the various quantization options provided by TheBloke to find the best balance between performance and accuracy for their specific use case.

Read more

Updated Invalid Date