NexusRaven-13B

Maintainer: Nexusflow

Total Score

97

Last updated 5/28/2024

👁️

PropertyValue
Run this modelRun on HuggingFace
API specView on HuggingFace
Github linkNo Github link provided
Paper linkNo paper link provided

Create account to get full access

or

If you already have an account, we'll log you in

Model overview

NexusRaven-13B is an open-source and commercially viable function calling language model developed by Nexusflow that surpasses the state-of-the-art in function calling capabilities. It was fine-tuned from the codellama/CodeLlama-13b-Instruct-hf model. Compared to GPT-4, NexusRaven-13B achieves a 95% success rate in using cybersecurity tools like CVE/CPE Search and VirusTotal, while GPT-4 achieves 64%. It has significantly lower cost and faster inference speed. NexusRaven-13B also generalizes well to tools never seen during training, achieving performance comparable to GPT-3.5 in zero-shot settings, outperforming other open-source LLMs of similar sizes.

Model inputs and outputs

NexusRaven-13B is a function calling language model that takes in a list of Python functions with their docstrings and generates JSON outputs with the function name and arguments. The model works best when provided with well-documented functions that have arguments, whether required or optional.

Inputs

  • Functions: A list of Python functions with their docstrings
  • User Query: A prompt for the model to generate a function call response to

Outputs

  • Function Call: A JSON object with the function name and argument values
  • Explanation (optional): A detailed explanation of the generated function call

Capabilities

NexusRaven-13B is capable of generating single function calls, nested calls, and parallel calls in many challenging cases. It can also provide detailed explanations for the function calls it generates, which can be turned off to save tokens during inference.

What can I use it for?

NexusRaven-13B can be used in a variety of applications that require interacting with APIs or executing functions based on user prompts. For example, you could use it to build a chatbot that can perform web scraping, make API calls, or execute other programmatic tasks on demand. The model's strong performance on cybersecurity tools makes it a promising candidate for building security-focused applications.

Things to try

One interesting thing to try with NexusRaven-13B is to provide it with a set of functions that interact with external APIs, such as fetching weather data or geolocating a city. You can then prompt the model to generate function calls that combine these capabilities to answer complex user queries, like "What's the weather like in Seattle right now?". The model's ability to chain together function calls and provide detailed explanations can make it a powerful tool for building conversational AI applications.



This summary was produced with help from an AI and may contain inaccuracies - check out the links to read the original source documents!

Related Models

🧠

NexusRaven-V2-13B

Nexusflow

Total Score

417

The NexusRaven-V2-13B is an open-source and commercially viable large language model (LLM) developed by Nexusflow that surpasses the state-of-the-art in function calling capabilities. It is capable of generating single function calls, nested calls, and parallel calls across many challenging cases. The model has been fine-tuned on a large corpus of function calls and can provide detailed explanations for the function calls it generates. Compared to the GPT-4 model, NexusRaven-V2-13B achieves a 7% higher function calling success rate on human-generated use cases involving nested and composite functions. Notably, the model has never been trained on the specific functions used in the evaluation, demonstrating strong generalization to the unseen. The training data for the model does not include any proprietary data from models like GPT-4, giving users full control when deploying it in commercial applications. Model Inputs and Outputs Inputs List of Python functions**: The model accepts a list of Python functions as input. The functions can perform any task, including sending GET/POST requests to external APIs. Function signatures and docstrings**: To enable the model to generate function calls, the input must include the Python function signature and an appropriate docstring. Function arguments**: The model performs best on functions that require arguments, so users should provide functions with arguments. Outputs Function calls**: The primary output of the model is function calls, which can be single, nested, or parallel. Detailed explanations**: The model can also generate detailed explanations for the function calls it produces, though this behavior can be turned off to save tokens during inference. Capabilities The NexusRaven-V2-13B model excels at zero-shot function calling, surpassing the performance of GPT-4 by a significant margin. It can handle a wide range of function call types, from simple single calls to complex nested and parallel calls. The model's ability to generalize to unseen functions is particularly impressive, as it demonstrates its versatility and potential for real-world applications. What Can I Use it For? The NexusRaven-V2-13B model is well-suited for a variety of applications that require function calling capabilities, such as: Automated software development**: The model can be used to assist developers in writing and orchestrating complex software systems by generating function calls on-the-fly. Intelligent virtual assistants**: The model's function calling abilities can be leveraged to build virtual assistants that can perform a wide range of tasks by dynamically calling relevant functions. Data processing and analysis**: The model's function calling capabilities can be used to build pipelines for data processing and analysis, automating complex workflows. Things to Try One interesting thing to try with the NexusRaven-V2-13B model is to provide it with a diverse set of custom functions and observe how it handles the function calling process. You can experiment with different types of functions, including those that interact with external APIs, to see the model's versatility and adaptability. Additionally, you can explore the model's ability to generate detailed explanations for the function calls it produces and how this feature can be leveraged in various applications.

Read more

Updated Invalid Date

💬

NexusRaven-V2-13B-GGUF

TheBloke

Total Score

48

The NexusRaven-V2-13B-GGUF is a large language model created by Nexusflow and quantized in the GGUF format by TheBloke. It is based on the original NexusRaven V2 13B model. The GGUF format offers improved tokenization and support for special tokens compared to the previous GGML format. Model inputs and outputs Inputs Text prompt**: The model accepts natural language text prompts as input. Outputs Text generation**: The model can generate coherent and contextual text continuations of the input prompt. Capabilities The NexusRaven-V2-13B-GGUF model demonstrates strong natural language understanding and generation capabilities. It can engage in open-ended conversations, summarize information, and answer questions on a wide range of topics. The model's capabilities make it well-suited for tasks like chatbots, content generation, and language-based AI assistants. What can I use it for? The NexusRaven-V2-13B-GGUF model could be used for a variety of natural language processing applications. Some potential use cases include: Conversational AI**: Integrating the model into a chatbot or virtual assistant to engage in open-ended conversations and assist users with a range of tasks. Content generation**: Using the model to generate articles, stories, scripts, or other forms of written content. Summarization**: Leveraging the model's text summarization capabilities to condense long-form text into concise summaries. Question answering**: Deploying the model to answer questions on a variety of topics, drawing upon its broad knowledge base. Things to try Experiment with providing the model with different types of prompts, such as open-ended questions, creative writing prompts, or task-oriented instructions. Observe how the model responds and generates text, noting its coherence, contextual awareness, and ability to stay on topic. Additionally, try varying the model parameters, like temperature and repetition penalty, to see how they affect the output.

Read more

Updated Invalid Date

⛏️

Starling-LM-7B-beta

Nexusflow

Total Score

318

Starling-LM-7B-beta is an open large language model (LLM) developed by the Nexusflow team. It is trained using Reinforcement Learning from AI Feedback (RLAIF) and finetuned from the Openchat-3.5-0106 model, which is based on the Mistral-7B-v0.1 model. The model uses the berkeley-nest/Nectar ranking dataset and the Nexusflow/Starling-RM-34B reward model, along with the Fine-Tuning Language Models from Human Preferences (PPO) policy optimization method. This results in an improved score of 8.12 on the MT Bench evaluation with GPT-4 as the judge, compared to the 7.81 score of the original Openchat-3.5-0106 model. Model inputs and outputs Inputs A conversational prompt following the exact chat template provided for the Openchat-3.5-0106 model. Outputs A natural language response to the input prompt. Capabilities Starling-LM-7B-beta is a capable language model that can engage in open-ended conversations, provide informative responses, and assist with a variety of tasks. It has demonstrated strong performance on benchmarks like MT Bench, outperforming several other prominent language models. What can I use it for? Starling-LM-7B-beta can be used for a wide range of applications, such as: Conversational AI**: The model can be used to power chatbots and virtual assistants that engage in natural conversations. Content generation**: The model can be used to generate written content like articles, stories, or scripts. Question answering**: The model can be used to answer questions on a variety of topics. Task assistance**: The model can be used to help with tasks like summarization, translation, and code generation. Things to try One interesting aspect of Starling-LM-7B-beta is its ability to perform well while maintaining a consistent conversational format. By adhering to the prescribed chat template, the model is able to produce coherent and on-topic responses without deviating from the expected structure. This can be particularly useful in applications where a specific interaction style is required, such as in customer service or educational chatbots.

Read more

Updated Invalid Date

xLAM-7b-fc-r

Salesforce

Total Score

60

The xLAM-7b-fc-r model is part of the xLAM model family developed by Salesforce. xLAMs are advanced large language models designed to enhance decision-making and translate user intentions into executable actions that interact with the world. The xLAM-7b-fc-r model is optimized for function-calling capability, providing fast, accurate, and structured responses based on input queries and available APIs. This model is fine-tuned based on the deepseek-coder models and is designed to be small enough for deployment on personal devices. The model series also includes the xLAM-1b-fc-r and xLAM-7b-fc-r versions, which vary in size and context length to cater to different applications. These models are part of the broader LAM (Large Action Model) family, which aims to serve as the "brains" of AI agents by autonomously planning and executing tasks to achieve specific goals. Model inputs and outputs Inputs Natural language queries**: The model can accept a wide range of natural language inputs, from simple questions to complex instructions, which it then translates into executable actions. Available APIs**: The model can utilize information about available APIs and services to generate appropriate responses and actions. Outputs Structured responses**: The model provides detailed, step-by-step responses that outline the actions to be taken, often in a format that can be directly integrated with external systems or applications. Executable actions**: Beyond just generating text, the model can produce executable actions, such as API calls, that can be directly integrated into workflow processes. Capabilities The xLAM-7b-fc-r model excels at translating user intentions into executable actions, enabling the automation of various workflow processes. For example, it can assist with tasks like scheduling appointments, managing to-do lists, or even programming simple APIs - all through natural language interactions. The model's function-calling capabilities make it particularly useful for enhancing productivity and streamlining operations across a wide range of domains. What can I use it for? The xLAM-7b-fc-r model and the broader xLAM series have numerous applications, including: Workflow automation**: Integrate the model into business processes to automate repetitive tasks and enhance productivity. Personal digital assistants**: Leverage the model's natural language understanding and action-generation capabilities to build intelligent virtual assistants. No-code/low-code development**: Utilize the model's function-calling abilities to enable non-technical users to create custom applications and integrations. Intelligent process automation**: Combine the model's decision-making and action-planning skills with robotic process automation (RPA) for end-to-end workflow optimization. Things to try One interesting aspect of the xLAM-7b-fc-r model is its ability to handle multi-step tasks and break them down into structured, executable steps. Try providing the model with complex instructions or a series of related queries, and observe how it plans and responds to achieve the desired outcome. The model's versatility in translating natural language into effective actions makes it a powerful tool for streamlining various workflows and automating repetitive processes.

Read more

Updated Invalid Date