responseFormat property
- @_CreateAssistantRequestResponseFormatConverter()
- @JsonKey(name: 'response_format', includeIfNull: false)
Specifies the format that the model must output. Compatible with
GPT-4o,
GPT-4 Turbo, and all GPT-3.5 Turbo models
since gpt-3.5-turbo-1106
.
Setting to { "type": "json_schema", "json_schema": {...} }
enables Structured Outputs which ensures
the model will match your supplied JSON schema. Learn more in the
Structured Outputs guide.
Setting to { "type": "json_object" }
enables JSON mode, which ensures the message the model generates
is valid JSON.
Important: when using JSON mode, you must also instruct the model to produce JSON yourself via a
system or user message. Without this, the model may generate an unending stream of whitespace until the
generation reaches the token limit, resulting in a long-running and seemingly "stuck" request. Also note
that the message content may be partially cut off if finish_reason="length"
, which indicates the
generation exceeded max_tokens
or the conversation exceeded the max context length.
Implementation
@_CreateAssistantRequestResponseFormatConverter()
@JsonKey(name: 'response_format', includeIfNull: false)
CreateAssistantRequestResponseFormat? get responseFormat =>
throw _privateConstructorUsedError;