Create Response
Create an LLM response.
Supports both streaming and non-streaming modes via the stream parameter.
Request
This endpoint expects an object.
input
Conversation history
model
Model ID in format ‘provider/model-name’. Optional if a routing policy is configured.
tools
Available tools
tool_choice
Tool choice policy: ‘auto’, ‘none’, ‘required’, or {‘type’: ‘function’, ‘function’: {‘name’: ’…’}}
max_tokens
Maximum tokens to generate
temperature
Sampling temperature
top_p
Nucleus sampling parameter
stop
Stop sequences
response_format
Response format
stream
Whether to stream the response
project_id
Optional project ID (UUID) to associate with this request
routing_policy_id
Override the default routing policy (for testing)
include_routing_metadata
Include detailed routing metadata in response
Response
Successful Response
id
created_at
model
The model that generated the response, in ‘provider/model-name’ format
output
usage
Token usage statistics.
object
provider_request_id
The upstream provider's request ID
routing
Routing metadata (only present when include_routing_metadata=true)
Errors
422
Unprocessable Entity Error