POST /chat/{assistant_name}/chat/completions

Chat with an assistant. This endpoint is based on the OpenAI Chat Completion API, a commonly used and adopted API.

It is useful if you need inline citations or OpenAI-compatible responses, but has limited functionality compared to the standard chat interface.

For guidance and examples, see Chat with an assistant.

Servers

Path parameters

Name Type Required Description
assistant_name String Yes

The name of the assistant to be described.

Request headers

Name Type Required Description
Content-Type String Yes The media type of the request body.

Default value: "application/json"

X-Pinecone-Api-Version String Yes

Required date-based version header

Default value: "2025-10"

Request body fields

Name Type Required Description
stream Boolean No

If false, the assistant will return a single JSON response. If true, the assistant will return a stream of responses.

Default value: false

messages[] Array Yes
messages[].content String No

Content of the message

messages[].role String No

Role of the message such as 'user' or 'assistant'

temperature Number No

Controls the randomness of the model's output: lower values make responses more deterministic, while higher values increase creativity and variability. If the model does not support a temperature parameter, the parameter will be ignored.

Default value: 0.0

filter Object No

Optionally filter which documents can be retrieved using the following metadata fields.

model String No

The large language model to use for answer generation

Default value: "gpt-4o"

How to start integrating

  1. Add HTTP Task to your workflow definition.
  2. Search for the API you want to integrate with and click on the name.
    • This loads the API reference documentation and prepares the Http request settings.
  3. Click Test request to test run your request to the API and see the API's response.