Anthracite-org

Models by this creator

🏷️

magnum-72b-v1

anthracite-org

Total Score

158

The magnum-72b-v1 model is the first in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of the Qwen-2 72B Instruct model. Model inputs and outputs The magnum-72b-v1 model has been Instruct tuned with the ChatML formatting. A typical input would look like this: """user Hi there! assistant Nice to meet you! user Can I ask a question? assistant """ Inputs Text prompts using the ChatML formatting Outputs Coherent, high-quality generated text responses Capabilities The magnum-72b-v1 model is designed to produce prose of a similar quality to the Claude 3 models. It can engage in open-ended conversation, answer questions, and generate creative text. What can I use it for? The magnum-72b-v1 model could be used for a variety of natural language tasks, such as chatbots, content generation, and creative writing assistance. As it is designed to replicate the quality of the Claude 3 models, it may be particularly well-suited for applications that require a more refined language output. Things to try One interesting aspect of the magnum-72b-v1 model is its fine-tuning on the Qwen-2 72B Instruct model. This could allow it to excel at following instructions and completing task-oriented prompts, in addition to open-ended conversation. Experimenters may want to try giving the model a variety of instructional prompts to see how it performs.

Read more

Updated 8/29/2024

🏷️

magnum-v1-72b

anthracite-org

Total Score

158

The magnum-v1-72b model is the first in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This 72B parameter model is fine-tuned on top of the Qwen-2 72B Instruct model. It was created by the anthracite-org team. Model inputs and outputs The magnum-v1-72b model uses a typical Instruct tuned input format with ChatML formatting. A common input would look like this: """user Hi there! assistant Nice to meet you! user Can I ask a question? assistant """ The model's outputs are generated in a similar ChatML format. Inputs User prompt**: The user's initial prompt or query System prompt**: An optional system-level prompt that sets the context Outputs Assistant response**: The model's generated response to the user's prompt Capabilities The magnum-v1-72b model aims to generate high-quality prose on par with the Claude 3 models. It has been trained on 55 million tokens of high-quality RP data to achieve this level of writing ability. What can I use it for? The magnum-v1-72b model could be useful for a variety of text generation tasks that require a high level of writing quality, such as: Creative writing assistance Generating detailed and coherent story passages Producing polished and persuasive arguments or essays Drafting professional-sounding business communications Things to try One interesting aspect of the magnum-v1-72b model is its use of the ChatML input format. This allows for more structured and context-rich prompts, which could enable the model to better understand and respond to complex queries or multi-turn conversations. Experimenting with different ChatML prompting techniques could be a fruitful area of exploration.

Read more

Updated 9/18/2024

👀

magnum-v2-12b

anthracite-org

Total Score

67

The magnum-v2-12b model is the fourth in a series of large language models created by the Anthracite organization. It is designed to replicate the high-quality prose of the Claude 3 models, specifically the Sonnet and Opus models. The model is fine-tuned on top of the Mistral-Nemo-Base-2407 model, incorporating datasets like the Stheno dataset (filtered), Opus_Instruct_25k, Opus_WritingStruct, and a subset of the Sonnet3.5-SlimOrcaDedupCleaned dataset. This model is part of a larger effort by the Anthracite team to develop high-quality language models. Model inputs and outputs The magnum-v2-12b model is an Instruct-tuned language model that accepts text input and generates text output. It uses the ChatML formatting, where the input is structured with system and user prompts enclosed in ` and ` tags. Inputs Text prompts**: The model accepts text prompts that can include instructions, questions, or other information for the model to generate a response. Outputs Generated text**: The model will generate text in response to the input prompt, aiming to produce high-quality, coherent prose. Capabilities The magnum-v2-12b model is capable of generating human-like text on a variety of topics, with a focus on producing content with a similar level of quality and style as the Claude 3 models. It can be used for tasks such as creative writing, content generation, and language modeling. What can I use it for? The magnum-v2-12b model can be used for a variety of natural language processing tasks, including: Content Generation**: Use the model to generate articles, stories, or other long-form content with a high level of coherence and quality. Conversational AI**: Integrate the model into a chatbot or virtual assistant to engage in natural conversations. Language Modeling**: Fine-tune the model on domain-specific data to create specialized language models for various applications. Things to try One interesting aspect of the magnum-v2-12b model is its ability to generate text with a distinct narrative voice and style. Try prompting the model with open-ended questions or writing prompts and see how it responds, exploring the range of tones and perspectives it can take on.

Read more

Updated 9/18/2024

👀

magnum-12b-v2

anthracite-org

Total Score

63

magnum-12b-v2 is the fourth in a series of models designed to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of Mistral-Nemo-Base-2407. Similar models in this series include magnum-72b-v1 and mini-magnum-12b-v1.1, which are also designed to replicate the prose quality of the Claude 3 models. Model Inputs and Outputs magnum-12b-v2 is an Instruct-tuned language model that has been fine-tuned using the ChatML formatting. This allows the model to engage in multi-turn chat-style dialogues, with the user providing prompts and the model generating responses. Inputs Prompts in ChatML format, with the user's message denoted by user and ``. System prompts that provide additional context or instructions for the model, denoted by system and ``. Outputs Responses generated by the model, denoted by assistant and ``. Capabilities magnum-12b-v2 is capable of generating high-quality prose, with a strong focus on coherence, fluency, and tone. The model can engage in a wide range of tasks, from creative writing to analytical tasks, and can adapt its language to suit the user's needs. What Can I Use It For? magnum-12b-v2 can be a valuable tool for projects that require natural language generation, such as content creation, dialogue systems, or language-based AI assistants. The model's ability to generate coherent and engaging prose makes it well-suited for tasks like creative writing, article generation, or even chatbots. Things to Try One interesting aspect of magnum-12b-v2 is its ability to maintain a consistent persona and voice across multiple turns of dialogue. Try engaging the model in a longer conversation and see how it adapts its responses to the context and flow of the discussion. You can also experiment with different types of prompts, from open-ended questions to more specific instructions, to explore the model's versatility.

Read more

Updated 9/14/2024

magnum-v2-123b

anthracite-org

Total Score

53

magnum-v2-123b is the sixth in a series of models designed by the team at anthracite-org to replicate the prose quality of the Claude 3 models, specifically Sonnet and Opus. This model is fine-tuned on top of the Mistral-Large-Instruct-2407 model. The model has been trained on a variety of high-quality datasets, including the Stheno-Data-Filtered, kalo-opus-instruct-22k-no-refusal, and nopm_claude_writing_fixed datasets. Model inputs and outputs The magnum-v2-123b model is a text-to-text AI model, meaning it takes text as input and generates text as output. The model has been fine-tuned for instruction following, and a typical input would look like this: [INST] SYSTEM MESSAGE\nUSER MESSAGE[/INST] ASSISTANT MESSAGE[INST] USER MESSAGE[/INST] The model also supports SillyTavern presets for Context and Instruct prompting. Capabilities The magnum-v2-123b model has been designed to produce high-quality, coherent prose that replicates the style of the Claude 3 models. It has been fine-tuned on a variety of datasets to improve its ability to generate natural-sounding text across a range of topics. What can I use it for? The magnum-v2-123b model could be used for a variety of text-generation tasks, such as creative writing, article generation, or task-oriented dialogue. Given its focus on replicating the style of the Claude 3 models, it may be particularly well-suited for applications that require a more formal or literary tone, such as academic or professional writing. Things to try One interesting aspect of the magnum-v2-123b model is its sensitivity to learning rate adjustments, which the maintainers hypothesize is due to the narrow and low-variance weight distributions typical of Mistral-derived models. This suggests that careful hyperparameter tuning may be necessary to get the best performance from the model, and users may want to experiment with different learning rates and other training parameters to find the optimal configuration for their specific use case.

Read more

Updated 9/19/2024