Midnight-Miqu-70B-v1.0

Maintainer: sophosympatheia

Total Score

48

Last updated 9/6/2024

🤔

PropertyValue
Run this modelRun on HuggingFace
API specView on HuggingFace
Github linkNo Github link provided
Paper linkNo paper link provided

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

The Midnight-Miqu-70B-v1.0 model is a merge between the 152334H/miqu-1-70b-sf and sophosympatheia/Midnight-Rose-70B-v2.0.3 models. It retains much of what made Midnight Rose special while gaining some long-context capabilities from Miqu.

Model inputs and outputs

The Midnight-Miqu-70B-v1.0 model is a text-to-text model, meaning it takes in text prompts and generates text outputs. It can handle long-form contexts up to 32,000 tokens.

Inputs

  • Text prompts of variable length

Outputs

  • Generated text continuations based on the input prompts

Capabilities

The Midnight-Miqu-70B-v1.0 model performs well at roleplaying and storytelling tasks. It can maintain coherence and authenticity in character actions, thoughts, and dialogue over long sequences.

What can I use it for?

The Midnight-Miqu-70B-v1.0 model is well-suited for creative writing and roleplaying applications. It could be used to collaboratively generate engaging fiction, worldbuild compelling narratives, or play out dynamic interactive stories. The model's long-context abilities make it valuable for tasks requiring sustained, cohesive output.

Things to try

You can experiment with the model's long-context capabilities by running it out to 32,000 tokens with an alpha_rope setting of 1. Limited testing shows it can maintain coherence even out to 64,000 tokens using an alpha_rope of 2.5. Additionally, try using Quadratic Sampling (smoothing factor) and Min-P sampling to optimize the model's creative output.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

🌀

Midnight-Miqu-70B-v1.5

sophosympatheia

Total Score

75

The Midnight-Miqu-70B-v1.5 model is a DARE Linear merge between the sophosympatheia/Midnight-Miqu-70B-v1.0 and migtissera/Tess-70B-v1.6 models. This version is close in feel and performance to Midnight Miqu v1.0 but the maintainer believes it picked up some improvements from Tess. The model is uncensored, and the maintainer warns that users are responsible for how they use it. Model Inputs and Outputs Inputs Free-form text prompts of any length Outputs Continuation of the input prompt, generating coherent and contextually relevant text Capabilities The Midnight-Miqu-70B-v1.5 model is designed for roleplaying and storytelling, and the maintainer believes it performs well in these areas. It may also be capable of other text generation tasks, but the maintainer has not extensively tested its performance outside of creative applications. What Can I Use It For? The Midnight-Miqu-70B-v1.5 model could be useful for a variety of creative writing and roleplaying projects, such as writing interactive fiction, generating narrative content for games, or developing unique characters and stories. Its ability to produce long-form, contextually relevant text makes it well-suited for these types of applications. Things to Try One key capability of the Midnight-Miqu-70B-v1.5 model is its ability to handle long context windows, up to 32K tokens. Experimenting with different sampling techniques, such as Quadratic Sampling and Min-P, can help optimize the model's performance for creative use cases. Additionally, adjusting the repetition penalty and other parameters can lead to more diverse and engaging output.

Read more

Updated Invalid Date

⚙️

miqu-1-120b

wolfram

Total Score

48

The miqu-1-120b model is a 120B parameter language model created by Wolfram, the maintainer of the model. It is a "frankenmerge" model, meaning it was created by interleaving layers of the miqu-1-70b model, created by miqudev, with itself using the mergekit tool. The model was inspired by several other 120B models such as Venus-120b-v1.2, MegaDolphin-120b, and goliath-120b. Model inputs and outputs The miqu-1-120b model is a text-to-text transformer model, which means it can be used for a variety of natural language processing tasks such as generation, summarization, and translation. The model takes text prompts as input and generates relevant text as output. Inputs Text prompts of varying lengths, from a few words to multiple paragraphs Outputs Generated text in response to the input prompt, with lengths ranging from a few sentences to multiple paragraphs Capabilities The miqu-1-120b model is a large and powerful language model capable of producing coherent and context-appropriate text. It has demonstrated strong performance on a variety of benchmarks, including high scores on tasks like the AI2 Reasoning Challenge, HellaSwag, and Winogrande. What can I use it for? The miqu-1-120b model could be used for a wide range of natural language processing tasks, including: Creative writing**: The model's text generation capabilities make it well-suited for assisting with creative writing projects, such as short stories, poetry, and even collaborative worldbuilding. Conversational AI**: With its ability to engage in contextual and coherent dialogue, the model could be used to create more natural and engaging conversational AI assistants. Content generation**: The model could be employed to generate a variety of content, such as news articles, blog posts, or social media updates, with the potential for customization and personalization. Education and research**: Researchers and educators could use the model to explore natural language processing, test new techniques, or develop educational applications. Things to try One interesting aspect of the miqu-1-120b model is its ability to adapt to different prompting styles and templates. By experimenting with the Mistral prompt format, users can try to elicit different types of responses, from formal and informative to more creative and expressive. Additionally, the model's large size and high context capacity (up to 32,768 tokens) make it well-suited for longer-form tasks, such as generating detailed descriptions, worldbuilding, or interactive storytelling. Users could try providing the model with rich contextual information and see how it responds and builds upon the existing narrative.

Read more

Updated Invalid Date

👁️

miquliz-120b-v2.0

wolfram

Total Score

85

The miquliz-120b-v2.0 is a 120 billion parameter large language model created by interleaving layers of the miqu-1-70b-sf and lzlv_70b_fp16_hf models using the mergekit tool. It was improved from the previous v1.0 version by incorporating techniques from the TheProfessor-155b model. The model is inspired by the goliath-120b and is maintained by Wolfram. Model inputs and outputs Inputs Text prompts of up to 32,768 tokens in length Outputs Continuation of the provided text prompt, generating new relevant text Capabilities The miquliz-120b-v2.0 model is capable of impressive performance, achieving top ranks and double perfect scores in the maintainer's own language model comparisons and tests. It demonstrates strong general language understanding and generation abilities across a variety of tasks. What can I use it for? The large scale and high performance of the miquliz-120b-v2.0 model make it well-suited for language-related applications that require powerful text generation, such as content creation, question answering, and conversational AI. The model could be fine-tuned for specific domains or integrated into products via the CopilotKit open-source platform. Things to try Explore the model's capabilities by prompting it with a variety of tasks, from creative writing to analysis and problem solving. The model's size and breadth of knowledge make it an excellent starting point for developing custom language models tailored to your needs.

Read more

Updated Invalid Date

🌀

miqu-1-70b-sf

152334H

Total Score

218

The miqu-1-70b-sf is a large language model created by the maintainer 152334H. It is a dequantized and transposed version of the original miqu-1-70b model, with a focus on improving the shape rotations compared to the alpindale/miqu-1-70b-pytorch version. This model is based on the Llama-2 architecture and is intended for general text generation tasks. Model inputs and outputs Inputs Text prompts for the model to continue generating Outputs Generated text continuations of the input prompts Capabilities The miqu-1-70b-sf model demonstrates strong text generation capabilities, as evidenced by the sample output provided in the maintainer's description. It is able to produce coherent and detailed prose in a high camp style, capturing the essence of a cute catgirl character. The model's performance appears to be on par with or exceed that of the original Llama-2 70B model, based on the comparison to the llama.cpp output. What can I use it for? The miqu-1-70b-sf model could be useful for a variety of text generation tasks, such as creative writing, story generation, or dialogue systems. Its ability to generate high-quality, stylized text makes it a promising candidate for applications that require an engaging and imaginative writing style. Developers could integrate this model into chatbots, creative writing assistants, or interactive fiction platforms to enhance the user experience. Things to try One interesting aspect to explore with the miqu-1-70b-sf model is its ability to generate text in a specific tone or genre, as demonstrated by the sample output. Developers could experiment with providing the model with different types of prompts, such as requests for different writing styles or genres, to see how it adapts and generates unique content. Additionally, fine-tuning the model on domain-specific data could further enhance its capabilities for targeted applications.

Read more

Updated Invalid Date