After uploading files to an assistant, you can chat with the assistant.
This page shows you how to chat with an assistant using the OpenAI-compatible chat interface. This interface is based on the OpenAI Chat Completion API, a commonly used and adopted API. It is useful if you need inline citations or OpenAI-compatible responses, but has limited functionality compared to the standard chat interface.
The standard chat interface is the recommended way to chat with an assistant, as it offers more functionality and control over the assistant’s responses and references.
The following example sends a message and requests a response in the default format:
The content parameter in the request cannot be empty.
Copy
# To use the Python SDK, install the plugin:# pip install --upgrade pinecone pinecone-plugin-assistantfrom pinecone import Pineconefrom pinecone_plugins.assistant.models.chat import Messagepc = Pinecone(api_key="YOUR_API_KEY")# Get your assistant.assistant = pc.assistant.Assistant( assistant_name="example-assistant", )# Chat with the assistant.chat_context = [Message(role="user", content='What is the maximum height of a red pine?')]response = assistant.chat_completions(messages=chat_context)
The example above returns a result like the following:
Copy
{"chat_completion": { "id":"chatcmpl-9OtJCcR0SJQdgbCDc9JfRZy8g7VJR", "choices":[ { "finish_reason":"stop", "index":0, "message":{ "role":"assistant", "content":"The maximum height of a red pine (Pinus resinosa) is up to 25 meters." } } ], "model":"my_assistant" }}
The following example sends a messages and requests a streaming response:
The content parameter in the request cannot be empty.
Copy
# To use the Python SDK, install the plugin:# pip install --upgrade pinecone pinecone-plugin-assistantfrom pinecone import Pineconefrom pinecone_plugins.assistant.models.chat import Messagepc = Pinecone(api_key="YOUR_API_KEY")# Get your assistant.assistant = pc.assistant.Assistant( assistant_name="example-assistant")# Streaming chat with the Assistant.chat_context = [Message(role="user", content="What is the maximum height of a red pine?")]response = assistant.chat_completions(messages=[chat_context], stream=True, model="gpt-4o")for data in response: if data: print(data)
The example above returns a result like the following:
The assistant’s response is returned in a JSON response object along with other information. The message string is contained in the following JSON object:
choices.[0].message.content for the default chat response
choices[0].delta.content for the streaming chat response
You can extract the message content and print it to the console:
Copy
import sys# Print the assistant's response to the console.print(str(response.choices[0].message.content))
This creates output like the following:
Copy
A red pine, scientifically known as *Pinus resinosa*, is a medium-sized tree that can grow up to 25 meters high and 75 centimeters in diameter. [1, pp. 1]
Copy
import sys# Print the assistant's response to the console.print(str(response.choices[0].message.content))
This creates output like the following:
Copy
A red pine, scientifically known as *Pinus resinosa*, is a medium-sized tree that can grow up to 25 meters high and 75 centimeters in diameter. [1, pp. 1]
Copy
import sys# Store streaming response.response = assistant.chat_completions(messages=chat_context, stream=True)for data in response: if data: print(str(data.choices[0].delta.content))
This creates output like the following:
Streaming response
Copy
The maximum height of a red pine (Pinus resinosa) is up to twenty-five meters [1, pp. 1].