LogoLogo
  • 👋Welcome to Arcee AI Docs
  • Arcee Orchestra
    • Introduction to Arcee Orchestra
    • Getting Started
    • Workflows
      • Workflow Components
        • Model Node
        • Code Node
        • Integrations
        • Knowledge Retrieval
        • Conditional Node
      • Passing Variables
      • API Invocation
        • List Available Workflows API
        • Workflow Execution API
        • Workflow Execution Steps API
        • Execution History API
        • Workflow Diagram API
        • API Code Examples
        • Upload Workflow JSON API
        • Workflow Runs API
    • Workflow Library
      • Research Automation
      • Real Time Financial Analysis
      • Blog Writer
      • Code Improvement
      • Energy Domain Assistant
    • Chat Interface
    • FAQ
  • ARCEE CONDUCTOR
    • Introduction to Arcee Conductor
    • Getting Started
    • Features & Functionality
      • Auto Mode
      • Auto Reasoning Mode
      • Auto Tools Mode
      • Compare
      • Direct Model Invocation
      • Usage
      • API
    • Arcee Small Language Models
      • Model Selection
      • Model Performance
    • Pricing
Powered by GitBook
On this page
  1. Arcee Orchestra

FAQ

PreviousChat InterfaceNextIntroduction to Arcee Conductor

Last updated 1 month ago

Technical FAQs

  1. What is the SLA for Arcee Orchestra when consumed via cloud SaaS?

    1. Uptime Commitment: 99.95% uptime, allowing for approximately 9 hours of downtime per year. Orchestra's platform's architecture ensures that updates or upgrades to individual services within the cluster do not disrupt other functionalities.

    2. Response Times: 24/7 on-call dev support with a response time of 60 minutes for critical issues. Resolution times will depend on complexity but are expected to be within hours, not days.

    3. Credits for SLA Breaches: Credits will be applied to account for workflow executions lost during periods of SLA breaches.

  2. Runtime limitations for Arcee Orchestra

    1. Workflows: Workflows execute along the graph node by node so there are no node or timeout limitations.

    2. Integrations: the provided integrations access external systems through APIs. There are no limitations on Arcee’s platform to restrict the number of requests to an external source; however, some companies build limitations into their own APIs which must be followed.

    3. Code Node: Not all PyPi packages are enabled by default, but we can make most libraries that customers need available.

  3. What are all the available built-in integrations?

    1. You can find a list of the available integrations .

  4. Is there any portability of the system?

    1. For each workflow you create, the configuration can be downloaded. This means you'll have the prompt templates for each model, what integrations you're connecting with, and any code written in the code nodes. This allows for portability of information and configurations.

  5. With the Enterprise "Shared" tier, could the performance suffer due to multiple tenants using at the same time?

    1. Our infrastructure supports multiple simultaneous workflows across numerous tenants. The application is built on Kubernetes, allowing for auto-scaling to accommodate high levels of traffic.

  6. Is there any guarantee of input/output token speed?

    1. For Arcee AI Small Lanaguage Models, the token speed will vary depending on which model size is selected. 10B, 14B, 32B, and 72B parameter models typically average 100, 70, 50, and 30 tokens per second (TPS) respectively. However, these speeds will vary based on factors such as input token count and concurrent requests. If you bring in a 3rd party model, Arcee cannot guarantee token speeds.

  7. What if we want to upgrade to "dedicated", is there a path for an upgrade?

    1. Yes, users can upgrade to dedicated hosting. This process involves cloning the data in the current environment and moving it to the new environment. This process is typically completed within five days.

here
Page cover image