# Chat

The OpenAI-compatible chat completion endpoint enables interactive conversations with large language models using the familiar `/v1/chat/completions` format. Designed for drop-in compatibility with OpenAI’s API, it supports role-based messages (`system`, `user`, `assistant`), streaming responses, and model-specific parameters like `temperature`,  and  `top_p`. Perfect for building chatbots, assistants, or any LLM-powered applications without needing to change your existing OpenAI integration.

{% content-ref url="chat/completions" %}
[completions](https://docs.distribute.ai/distribute-for-enterprise/enterprise-inference-api/openai-compatible-api/chat/completions)
{% endcontent-ref %}
