LogoLogo
IntroInstalling Dashboard
  • 🌐Distribute
    • Introducing DistributeAI
    • FAQ
  • 🪙Distribute For $DIS
    • Public Presale
    • Tokenomics
  • 📱Distribute For Consumers
    • Distribute.ai Discord Bot
      • Adding the bot to a server
      • Adding the bot as a user app
      • Bot Commands
      • Generation Buttons
      • Prompting Guide
    • Distribute.ai Dashboard
      • Statistics
      • Referrals
      • Account Settings
  • 🖥️Distribute For Inference Providers
    • Chrome Extension
      • Installing Extension
        • Enabling Chrome Flags
        • Chrome GPU Setup Guide
      • Updating the Extension
    • Desktop App
      • Installing Desktop App
      • Desktop App Debugging
        • Windows Debugging
        • MacOS Debugging
        • Linux Debugging
      • Desktop System Requirements
  • 🏢Distribute For Enterprise
    • Enterprise Inference API
      • Authentication
      • OpenAI Compatible API
        • Images
          • Generations
        • Chat
          • Completions
      • Async API
        • Chat
          • Chat Create
          • Chat Result
        • Images
          • Images Create
          • Images Result
    • Enterprise Compute Providers
Powered by GitBook
On this page
  1. Distribute For Enterprise
  2. Enterprise Inference API
  3. Async API

Chat

PreviousAsync APINextChat Create

Last updated 2 days ago

The Asynchronous Chat Completion API allows you to submit chat-based LLM tasks that are processed in the background and retrieved later. Ideal for high-latency or large-scale workloads, this custom implementation lets you queue chat requests via a POST endpoint and retrieve results using a unique task ID. It supports robust job tracking, delayed responses, and retry-safe workflows—making it well-suited for batch processing, long-running prompts, and serverless environments where real-time responses aren't required.

🏢
Chat Create
Chat Result