input array contains message objects with a role (user or assistant) and content field. The model processes the input and returns a response with the generated text.
Basic Request
Response Format
The response includes the generated text in theoutput array, along with token usage information.
Multi-turn Conversations
For conversations with multiple messages, include both user and assistant messages in theinput array:
Request Parameters
The model identifier to use for generation. Example:
openai/gpt-5.3-codexSee the full list of available models in the Models section.Array of input messages. Each message object must contain:
type: Always"message"role: Either"user","assistant", or"system"content: The message text content
Maximum number of tokens to generate. Controls the length of the output.
Controls randomness in generation. Range:
0.0 to 2.0. Lower values produce more focused output; higher values produce more varied output. Default: 1.0Controls diversity via nucleus sampling. Range:
0.0 to 1.0. Not recommended to use alongside temperature. Default: 1.0Response Fields
Unique identifier for this response.
Always
"response" for this endpoint.The model used to generate the response.
Array containing the generated response. Each item has:
type:"message"role:"assistant"content: Array withtype: "output_text"andtextfields
Token usage statistics with
input_tokens and output_tokens.