diff --git a/crates/coverage-report/src/requests_expected_differences.json b/crates/coverage-report/src/requests_expected_differences.json index 0da07891..55067103 100644 --- a/crates/coverage-report/src/requests_expected_differences.json +++ b/crates/coverage-report/src/requests_expected_differences.json @@ -843,6 +843,14 @@ { "pattern": "messages[*].content", "reason": "Empty assistant content (no parts) from truncated Gemini response gains a placeholder empty text part when roundtripped through other providers" }, { "pattern": "messages.length", "reason": "Empty assistant content from truncated Gemini response may change message count when roundtripped through providers that split/merge messages" } ] + }, + { + "testCase": "exclusiveMinimumToolParam", + "source": "ChatCompletions", + "target": "Google", + "fields": [ + { "pattern": "params.tools[0].parameters.properties.max_tokens.exclusiveMinimum", "reason": "Gemini's Schema proto does not support JSON Schema Draft-07 keywords such as exclusiveMinimum; they are stripped during conversion to avoid a 400 INVALID_ARGUMENT error" } + ] } ] } diff --git a/crates/lingua/src/providers/google/adapter.rs b/crates/lingua/src/providers/google/adapter.rs index a43ec8a0..ad7f0013 100644 --- a/crates/lingua/src/providers/google/adapter.rs +++ b/crates/lingua/src/providers/google/adapter.rs @@ -338,7 +338,8 @@ impl ProviderAdapter for GoogleAdapter { ); } - // Add tools if present + // Add tools if present. FunctionDeclaration::try_from strips unsupported JSON Schema + // keywords (e.g. `exclusiveMinimum`) from parametersJsonSchema during conversion. if let Some(tools) = &req.params.tools { let google_tools = as TryFromLLM>>::try_from(tools.clone()) .map_err(|e| TransformError::FromUniversalFailed(e.to_string()))?; @@ -769,6 +770,7 @@ fn fill_tool_names_from_context(messages: &mut [Message]) { #[cfg(test)] mod tests { use super::*; + use crate::providers::google::GenerateContentRequest; use crate::serde_json::json; use crate::universal::request::ToolChoiceMode; @@ -1023,6 +1025,56 @@ mod tests { assert_eq!(parts.len(), 1); } + #[test] + fn test_exclusive_minimum_stripped_from_google_request() { + let adapter = GoogleAdapter; + let req = UniversalRequest { + model: None, + messages: vec![Message::User { + content: UserContent::String("Hello".into()), + }], + params: UniversalParams { + tools: Some(vec![crate::universal::tools::UniversalTool::function( + "my_tool", + Some("A tool".to_string()), + Some(json!({ + "type": "object", + "properties": { + "count": { + "type": "integer", + "exclusiveMinimum": 0 + } + } + })), + None, + )]), + ..Default::default() + }, + }; + + let payload = adapter.request_from_universal(&req).unwrap(); + + // Deserialize into the typed request struct so we navigate via struct + // fields rather than raw Value map access. + let typed: GenerateContentRequest = serde_json::from_value(payload).unwrap(); + let decl = &typed.tools.as_ref().unwrap()[0] + .function_declarations + .as_ref() + .unwrap()[0]; + let schema = decl + .parameters_json_schema + .as_ref() + .expect("parametersJsonSchema should be present"); + + // parameters_json_schema is an opaque JSON value; verify absence of the + // unsupported keyword by checking the serialized form. + let schema_str = serde_json::to_string(schema).unwrap(); + assert!( + !schema_str.contains("exclusiveMinimum"), + "exclusiveMinimum must be stripped from Google request" + ); + } + #[test] fn test_google_stream_from_universal_usage_only_emits_placeholder_candidate() { let adapter = GoogleAdapter; diff --git a/crates/lingua/src/providers/google/convert.rs b/crates/lingua/src/providers/google/convert.rs index fabc2f29..dec4cf2f 100644 --- a/crates/lingua/src/providers/google/convert.rs +++ b/crates/lingua/src/providers/google/convert.rs @@ -810,17 +810,49 @@ fn normalize_google_schema_types(mut value: Value) -> Value { value } +/// Recursively strip `exclusiveMinimum` fields from a JSON schema value in-place. +/// +/// Google's `Schema` proto does not recognise `exclusiveMinimum` (a JSON Schema Draft 6+ +/// keyword) and returns a 400 INVALID_ARGUMENT error when it is present. We strip it +/// here — at the conversion boundary — so that callers never have to remember to do so +/// themselves. The original schema is preserved in the caller's `UniversalTool.parameters` +/// field and can be round-tripped via the adapter's extras mechanism. +fn strip_exclusive_minimum(value: &mut Value) { + match value { + Value::Object(map) => { + map.remove("exclusiveMinimum"); + for v in map.values_mut() { + strip_exclusive_minimum(v); + } + } + Value::Array(arr) => { + for v in arr.iter_mut() { + strip_exclusive_minimum(v); + } + } + _ => {} + } +} + impl TryFrom<&UniversalTool> for FunctionDeclaration { type Error = ConvertError; fn try_from(tool: &UniversalTool) -> Result { match &tool.tool_type { - UniversalToolType::Function => Ok(FunctionDeclaration { - name: Some(tool.name.clone()), - description: tool.description.clone(), - parameters_json_schema: tool.parameters.clone(), - ..Default::default() - }), + UniversalToolType::Function => { + // Strip JSON Schema keywords unsupported by Google's Schema proto (e.g. + // `exclusiveMinimum`) before embedding the schema in the declaration. + let parameters_json_schema = tool.parameters.clone().map(|mut p| { + strip_exclusive_minimum(&mut p); + p + }); + Ok(FunctionDeclaration { + name: Some(tool.name.clone()), + description: tool.description.clone(), + parameters_json_schema, + ..Default::default() + }) + } UniversalToolType::Custom { .. } => Err(ConvertError::UnsupportedToolType { tool_name: tool.name.clone(), tool_type: "custom".to_string(), diff --git a/payloads/cases/params.ts b/payloads/cases/params.ts index bd249345..2a1af361 100644 --- a/payloads/cases/params.ts +++ b/payloads/cases/params.ts @@ -13,6 +13,7 @@ import { OPENAI_NON_REASONING_MODEL, ANTHROPIC_MODEL, ANTHROPIC_OPUS_MODEL, + GOOGLE_MODEL, GOOGLE_GEMINI_3_MODEL, GOOGLE_IMAGE_MODEL, GOOGLE_TTS_MODEL, @@ -2105,4 +2106,88 @@ export const paramsCases: TestCaseCollection = { }, bedrock: null, }, + + exclusiveMinimumToolParam: { + "chat-completions": { + model: OPENAI_CHAT_COMPLETIONS_MODEL, + messages: [{ role: "user", content: "Configure the LLM." }], + tools: [ + { + type: "function", + function: { + name: "configure_llm", + description: "Configure LLM generation parameters", + parameters: { + type: "object", + properties: { + max_tokens: { + type: "number", + exclusiveMinimum: 0, + description: "Maximum number of tokens to generate", + }, + }, + required: ["max_tokens"], + additionalProperties: false, + }, + }, + }, + ], + }, + responses: { + model: OPENAI_RESPONSES_MODEL, + input: [{ role: "user", content: "Configure the LLM." }], + tools: [ + { + type: "function", + name: "configure_llm", + description: "Configure LLM generation parameters", + parameters: { + type: "object", + properties: { + max_tokens: { + type: "number", + exclusiveMinimum: 0, + description: "Maximum number of tokens to generate", + }, + }, + required: ["max_tokens"], + additionalProperties: false, + }, + strict: false, + }, + ], + }, + anthropic: null, + google: (() => { + // Assigned to a variable first so TypeScript applies structural (not + // excess-property) checking when it lands in Record. + // exclusiveMinimum is not in Gemini's Schema type but IS passed here + // deliberately to capture the resulting 400 INVALID_ARGUMENT error. + const maxTokensSchema = { + type: Type.NUMBER, + exclusiveMinimum: 0, + description: "Maximum number of tokens to generate", + }; + return { + model: GOOGLE_MODEL, + contents: [{ role: "user", parts: [{ text: "Configure the LLM." }] }], + tools: [ + { + functionDeclarations: [ + { + name: "configure_llm", + description: "Configure LLM generation parameters", + parameters: { + type: Type.OBJECT, + properties: { max_tokens: maxTokensSchema }, + required: ["max_tokens"], + }, + }, + ], + }, + ], + }; + })(), + bedrock: null, + }, }; diff --git a/payloads/scripts/transforms/__snapshots__/transforms.test.ts.snap b/payloads/scripts/transforms/__snapshots__/transforms.test.ts.snap index 1a0711d4..e8b426a6 100644 --- a/payloads/scripts/transforms/__snapshots__/transforms.test.ts.snap +++ b/payloads/scripts/transforms/__snapshots__/transforms.test.ts.snap @@ -6714,6 +6714,79 @@ Now I'll total up for each digit across all 1440 minutes in a day (24 × 60): } `; +exports[`chat-completions → anthropic > exclusiveMinimumToolParam > request 1`] = ` +{ + "max_tokens": 4096, + "messages": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "claude-sonnet-4-5-20250929", + "tools": [ + { + "description": "Configure LLM generation parameters", + "input_schema": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "exclusiveMinimum": 0, + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "name": "configure_llm", + }, + ], +} +`; + +exports[`chat-completions → anthropic > exclusiveMinimumToolParam > response 1`] = ` +{ + "choices": [ + { + "finish_reason": "stop", + "index": 0, + "message": { + "annotations": [], + "content": "I'd be happy to help configure the LLM! However, I need to know what value you'd like to set for the \`max_tokens\` parameter. + +The \`max_tokens\` parameter controls the maximum number of tokens (pieces of text) that the LLM can generate in its response. + +Could you please specify: +- **max_tokens**: The maximum number of tokens you want the LLM to generate (must be greater than 0) + +For example, common values might be: +- 100-500 for short responses +- 1000-2000 for medium-length responses +- 4000+ for longer, detailed responses + +What value would you like to use?", + "role": "assistant", + }, + }, + ], + "created": 0, + "id": "chatcmpl-01GG7qfqnsDh9GxCJn5VWgPx", + "model": "claude-sonnet-4-5-20250929", + "object": "chat.completion", + "usage": { + "completion_tokens": 154, + "prompt_tokens": 596, + "prompt_tokens_details": { + "cached_tokens": 0, + }, + "total_tokens": 750, + }, +} +`; + exports[`chat-completions → anthropic > frequencyPenaltyParam > request 1`] = ` { "max_tokens": 4096, @@ -9193,6 +9266,75 @@ Now, let's count the occurrences of each digit systematically. We can do this by } `; +exports[`chat-completions → google > exclusiveMinimumToolParam > request 1`] = ` +{ + "contents": [ + { + "parts": [ + { + "text": "Configure the LLM.", + }, + ], + "role": "user", + }, + ], + "model": "gemini-2.5-flash", + "tools": [ + { + "functionDeclarations": [ + { + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": null, + "parametersJsonSchema": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "response": null, + }, + ], + }, + ], +} +`; + +exports[`chat-completions → google > exclusiveMinimumToolParam > response 1`] = ` +{ + "choices": [ + { + "finish_reason": "stop", + "index": 0, + "message": { + "annotations": [], + "content": "I can help with that. What would you like to set the maximum number of tokens to?", + "role": "assistant", + }, + }, + ], + "created": 0, + "id": "chatcmpl-transformed", + "model": "gemini-2.5-flash", + "object": "chat.completion", + "usage": { + "completion_tokens": 67, + "completion_tokens_details": { + "reasoning_tokens": 48, + }, + "prompt_tokens": 53, + "total_tokens": 120, + }, +} +`; + exports[`chat-completions → google > frequencyPenaltyParam > request 1`] = ` { "contents": [ @@ -11978,6 +12120,87 @@ Brief note on method: counted digit occurrences separately for hour tens (H1), h } `; +exports[`chat-completions → responses > exclusiveMinimumToolParam > request 1`] = ` +{ + "input": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "gpt-5-nano", + "tools": [ + { + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "exclusiveMinimum": 0, + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "type": "function", + }, + ], +} +`; + +exports[`chat-completions → responses > exclusiveMinimumToolParam > response 1`] = ` +{ + "choices": [ + { + "finish_reason": "stop", + "index": 0, + "message": { + "annotations": [], + "reasoning": "", + "role": "assistant", + }, + }, + { + "finish_reason": "stop", + "index": 1, + "message": { + "annotations": [], + "content": "Sure. I can configure the LLM, but I need a max_tokens value to apply. + +Options (balanced defaults): +- 512 – concise responses +- 1024 – short to medium +- 2048 – balanced +- 4096 – longer responses + +If you’d like, I can apply 2048 tokens right now as a balanced default. Would you like me to set max_tokens to 2048 (or choose another value)? Also, note that this configuration step only covers max_tokens; tell me if you want me to note any other preferences, and I can track them for you.", + "role": "assistant", + }, + }, + ], + "created": 0, + "id": "chatcmpl-0038cce6fceb0e88006a01faa00e5c8195921786d0852993cd", + "model": "gpt-5-nano-2025-08-07", + "object": "chat.completion", + "usage": { + "completion_tokens": 768, + "completion_tokens_details": { + "reasoning_tokens": 576, + }, + "prompt_tokens": 66, + "prompt_tokens_details": { + "cached_tokens": 0, + }, + "total_tokens": 834, + }, +} +`; + exports[`chat-completions → responses > frequencyPenaltyParam > request 1`] = ` { "input": [ @@ -14709,6 +14932,71 @@ Each digit 0-9 appears 6 times per hour (once in each group of 10 minutes) } `; +exports[`google → anthropic > exclusiveMinimumToolParam > request 1`] = ` +{ + "max_tokens": 4096, + "messages": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "claude-sonnet-4-5-20250929", + "tools": [ + { + "description": "Configure LLM generation parameters", + "input_schema": { + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "name": "configure_llm", + }, + ], +} +`; + +exports[`google → anthropic > exclusiveMinimumToolParam > response 1`] = ` +{ + "candidates": [ + { + "content": { + "parts": [ + { + "text": "I'd be happy to help configure the LLM! However, I need to know what value you'd like to set for the maximum number of tokens to generate. + +Could you please specify the \`max_tokens\` value you want? For example: +- 256 for shorter responses +- 512 for medium-length responses +- 1024 or higher for longer responses +- 4096 for very detailed responses + +What maximum token limit would you like to set?", + }, + ], + "role": "model", + }, + "finishReason": "STOP", + "index": 0, + }, + ], + "modelVersion": "claude-sonnet-4-5-20250929", + "usageMetadata": { + "cachedContentTokenCount": 0, + "candidatesTokenCount": 105, + "promptTokenCount": 579, + "totalTokenCount": 684, + }, +} +`; + exports[`google → anthropic > googleResponseSchemaPropertyOrderingParam > request 1`] = ` { "max_tokens": 128, @@ -17010,6 +17298,73 @@ Summary: } `; +exports[`google → chat-completions > exclusiveMinimumToolParam > request 1`] = ` +{ + "messages": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "gpt-5-nano", + "tools": [ + { + "function": { + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + }, + "type": "function", + }, + ], +} +`; + +exports[`google → chat-completions > exclusiveMinimumToolParam > response 1`] = ` +{ + "candidates": [ + { + "content": { + "parts": [ + { + "text": "I can configure the LLM, but I need the max_tokens value (the maximum number of tokens in responses). + +Options: +- 512 (short responses) +- 1024 (default) +- 2048 (longer responses) + +Would you like me to set max_tokens to 1024 now, or provide a different value? If you’re unsure, I can start with 1024 and adjust later.", + }, + ], + "role": "model", + }, + "finishReason": "STOP", + "index": 0, + }, + ], + "modelVersion": "gpt-5-nano-2025-08-07", + "usageMetadata": { + "cachedContentTokenCount": 0, + "candidatesTokenCount": 93, + "promptTokenCount": 139, + "thoughtsTokenCount": 704, + "totalTokenCount": 936, + }, +} +`; + exports[`google → chat-completions > googleResponseSchemaPropertyOrderingParam > request 1`] = ` { "max_completion_tokens": 128, @@ -19177,6 +19532,82 @@ So: } `; +exports[`google → responses > exclusiveMinimumToolParam > request 1`] = ` +{ + "input": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "gpt-5-nano", + "tools": [ + { + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "type": "function", + }, + ], +} +`; + +exports[`google → responses > exclusiveMinimumToolParam > response 1`] = ` +{ + "candidates": [ + { + "content": { + "parts": [ + { + "text": "", + "thought": true, + }, + ], + "role": "model", + }, + "finishReason": "STOP", + "index": 0, + }, + { + "content": { + "parts": [ + { + "functionCall": { + "args": { + "max_tokens": 1024, + }, + "name": "configure_llm", + }, + }, + ], + "role": "model", + }, + "finishReason": "STOP", + "index": 1, + }, + ], + "modelVersion": "gpt-5-nano-2025-08-07", + "usageMetadata": { + "cachedContentTokenCount": 0, + "candidatesTokenCount": 75, + "promptTokenCount": 57, + "thoughtsTokenCount": 512, + "totalTokenCount": 644, + }, +} +`; + exports[`google → responses > googleResponseSchemaPropertyOrderingParam > request 1`] = ` { "input": [ @@ -22450,6 +22881,97 @@ The digits 0 and 1 dominate because they appear frequently in both hours (00-19) } `; +exports[`responses → anthropic > exclusiveMinimumToolParam > request 1`] = ` +{ + "max_tokens": 4096, + "messages": [ + { + "content": "Configure the LLM.", + "role": "user", + }, + ], + "model": "claude-sonnet-4-5-20250929", + "tools": [ + { + "description": "Configure LLM generation parameters", + "input_schema": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "exclusiveMinimum": 0, + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "name": "configure_llm", + "strict": false, + }, + ], +} +`; + +exports[`responses → anthropic > exclusiveMinimumToolParam > response 1`] = ` +{ + "created_at": 0, + "id": "resp_01ERkvPjbQ7ZLjFZns3hskpK", + "incomplete_details": null, + "model": "claude-sonnet-4-5-20250929", + "object": "response", + "output": [ + { + "content": [ + { + "annotations": [], + "text": "I'd be happy to help configure the LLM! To do this, I need to know what value you'd like to set for the **max_tokens** parameter, which controls the maximum number of tokens the LLM can generate in its responses. + +Could you please specify: +- How many max_tokens would you like to configure? (This should be a positive number) + +For context, common values might be: +- 100-500 for short responses +- 500-2000 for medium-length responses +- 2000+ for longer, more detailed responses", + "type": "output_text", + }, + ], + "id": "msg_transformed_item_0", + "role": "assistant", + "status": "completed", + "type": "message", + }, + ], + "output_text": "I'd be happy to help configure the LLM! To do this, I need to know what value you'd like to set for the **max_tokens** parameter, which controls the maximum number of tokens the LLM can generate in its responses. + +Could you please specify: +- How many max_tokens would you like to configure? (This should be a positive number) + +For context, common values might be: +- 100-500 for short responses +- 500-2000 for medium-length responses +- 2000+ for longer, more detailed responses", + "parallel_tool_calls": false, + "status": "completed", + "tool_choice": "none", + "tools": [], + "usage": { + "input_tokens": 596, + "input_tokens_details": { + "cached_tokens": 0, + }, + "output_tokens": 127, + "output_tokens_details": { + "reasoning_tokens": 0, + }, + "total_tokens": 723, + }, +} +`; + exports[`responses → anthropic > instructionsParam > request 1`] = ` { "max_tokens": 4096, @@ -25436,6 +25958,88 @@ Total for H2: } `; +exports[`responses → google > exclusiveMinimumToolParam > request 1`] = ` +{ + "contents": [ + { + "parts": [ + { + "text": "Configure the LLM.", + }, + ], + "role": "user", + }, + ], + "model": "gemini-2.5-flash", + "tools": [ + { + "functionDeclarations": [ + { + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": null, + "parametersJsonSchema": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number", + }, + }, + "required": [ + "max_tokens", + ], + "type": "object", + }, + "response": null, + }, + ], + }, + ], +} +`; + +exports[`responses → google > exclusiveMinimumToolParam > response 1`] = ` +{ + "created_at": 0, + "id": "resp_transformed", + "incomplete_details": null, + "model": "gemini-2.5-flash", + "object": "response", + "output": [ + { + "content": [ + { + "annotations": [], + "text": "Please provide the \`max_tokens\` parameter to configure the LLM.", + "type": "output_text", + }, + ], + "id": "msg_transformed_item_0", + "role": "assistant", + "status": "completed", + "type": "message", + }, + ], + "output_text": "Please provide the \`max_tokens\` parameter to configure the LLM.", + "parallel_tool_calls": false, + "status": "completed", + "tool_choice": "none", + "tools": [], + "usage": { + "input_tokens": 53, + "input_tokens_details": { + "cached_tokens": 0, + }, + "output_tokens": 74, + "output_tokens_details": { + "reasoning_tokens": 59, + }, + "total_tokens": 127, + }, +} +`; + exports[`responses → google > instructionsParam > request 1`] = ` { "contents": [ diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-request.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-request.json new file mode 100644 index 00000000..6dcec1ac --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-request.json @@ -0,0 +1,53 @@ +{ + "model": "gpt-5-nano", + "messages": [ + { + "role": "user", + "content": "Configure the LLM." + }, + { + "role": "assistant", + "content": null, + "tool_calls": [ + { + "id": "call_U4A1cymd03k0mIgSaB4RwJKs", + "type": "function", + "function": { + "name": "configure_llm", + "arguments": "{\"max_tokens\":4096}" + } + } + ], + "refusal": null, + "annotations": [] + }, + { + "role": "tool", + "tool_call_id": "call_U4A1cymd03k0mIgSaB4RwJKs", + "content": "71 degrees" + } + ], + "tools": [ + { + "type": "function", + "function": { + "name": "configure_llm", + "description": "Configure LLM generation parameters", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + } + } + } + ] +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response-streaming.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response-streaming.json new file mode 100644 index 00000000..6d410204 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response-streaming.json @@ -0,0 +1,1100 @@ +[ + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "role": "assistant", + "content": "", + "refusal": null + }, + "finish_reason": null + } + ], + "obfuscation": "yi" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "Configuration" + }, + "finish_reason": null + } + ], + "obfuscation": "lRtUaf4" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " updated" + }, + "finish_reason": null + } + ], + "obfuscation": "pCDQCKdkPmcb" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "." + }, + "finish_reason": null + } + ], + "obfuscation": "0mY" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " Maximum" + }, + "finish_reason": null + } + ], + "obfuscation": "4iz5bcCV0ALy" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " tokens" + }, + "finish_reason": null + } + ], + "obfuscation": "qVnsXq61UXSMZ" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " set" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " to" + }, + "finish_reason": null + } + ], + "obfuscation": "U" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " " + }, + "finish_reason": null + } + ], + "obfuscation": "4Qm" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "409" + }, + "finish_reason": null + } + ], + "obfuscation": "p" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "6" + }, + "finish_reason": null + } + ], + "obfuscation": "3E3" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": ".\n\n" + }, + "finish_reason": null + } + ], + "obfuscation": "vnsKeuZgtZj59xb" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "Note" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": ":" + }, + "finish_reason": null + } + ], + "obfuscation": "F3Z" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " the" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " tool" + }, + "finish_reason": null + } + ], + "obfuscation": "qVT5UAzV7dbbiDQ" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " returned" + }, + "finish_reason": null + } + ], + "obfuscation": "zphiBnjhgDh" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " an" + }, + "finish_reason": null + } + ], + "obfuscation": "B" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " odd" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " \"" + }, + "finish_reason": null + } + ], + "obfuscation": "L" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "71" + }, + "finish_reason": null + } + ], + "obfuscation": "zi" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " degrees" + }, + "finish_reason": null + } + ], + "obfuscation": "FegNL7Lj7x5o" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "\"" + }, + "finish_reason": null + } + ], + "obfuscation": "Ks" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " message" + }, + "finish_reason": null + } + ], + "obfuscation": "eCuBvWQcdieA" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " earlier" + }, + "finish_reason": null + } + ], + "obfuscation": "sNGcTVZme7d3" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "—" + }, + "finish_reason": null + } + ], + "obfuscation": "Uoa" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "please" + }, + "finish_reason": null + } + ], + "obfuscation": "zPkdKCh5Xd8OLo" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " ignore" + }, + "finish_reason": null + } + ], + "obfuscation": "L0mpOCn3aDC3h" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " that" + }, + "finish_reason": null + } + ], + "obfuscation": "F6z1wvlTPCsdWEG" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " as" + }, + "finish_reason": null + } + ], + "obfuscation": "z" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " a" + }, + "finish_reason": null + } + ], + "obfuscation": "z0" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " glitch" + }, + "finish_reason": null + } + ], + "obfuscation": "Evv1Tuu5klbCN" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "." + }, + "finish_reason": null + } + ], + "obfuscation": "rTi" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " I" + }, + "finish_reason": null + } + ], + "obfuscation": "9L" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "’ve" + }, + "finish_reason": null + } + ], + "obfuscation": "y" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " applied" + }, + "finish_reason": null + } + ], + "obfuscation": "9Xl6GyzfwboE" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " the" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " requested" + }, + "finish_reason": null + } + ], + "obfuscation": "43Kj4Isl00" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " max" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "_tokens" + }, + "finish_reason": null + } + ], + "obfuscation": "M1df8BUshEK1q" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " setting" + }, + "finish_reason": null + } + ], + "obfuscation": "mIg9MMqZmv08" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": ".\n\n" + }, + "finish_reason": null + } + ], + "obfuscation": "aIrrqqJ5MtaAKPX" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "Would" + }, + "finish_reason": null + } + ], + "obfuscation": "yzvIgC4EzVmvpVl" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " you" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " like" + }, + "finish_reason": null + } + ], + "obfuscation": "zufQvjyPSxiaMsE" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " me" + }, + "finish_reason": null + } + ], + "obfuscation": "e" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " to" + }, + "finish_reason": null + } + ], + "obfuscation": "Z" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " adjust" + }, + "finish_reason": null + } + ], + "obfuscation": "Up5EiHZqOcaXK" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " any" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " other" + }, + "finish_reason": null + } + ], + "obfuscation": "X3P3guH2nbDDa5" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " settings" + }, + "finish_reason": null + } + ], + "obfuscation": "rupBS0ISoOi" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " or" + }, + "finish_reason": null + } + ], + "obfuscation": "U" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " run" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " a" + }, + "finish_reason": null + } + ], + "obfuscation": "Pw" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " quick" + }, + "finish_reason": null + } + ], + "obfuscation": "FJ1gJlaMOA4UTd" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " test" + }, + "finish_reason": null + } + ], + "obfuscation": "9maTTUimNpONS05" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " to" + }, + "finish_reason": null + } + ], + "obfuscation": "1" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " verify" + }, + "finish_reason": null + } + ], + "obfuscation": "LkK2J7C0wLg7r" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " behavior" + }, + "finish_reason": null + } + ], + "obfuscation": "ePvsro3WbiY" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "?" + }, + "finish_reason": null + } + ], + "obfuscation": "uxU" + }, + { + "id": "chatcmpl-DeNAxeK3v7jlvWzTe3MhI8UWYLGgc", + "object": "chat.completion.chunk", + "created": 1778514579, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": {}, + "finish_reason": "stop" + } + ], + "obfuscation": "SZ0t0qZufFCb0A" + } +] \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response.json new file mode 100644 index 00000000..400a5d02 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/followup-response.json @@ -0,0 +1,35 @@ +{ + "id": "chatcmpl-DeNAwoDvTl3WsKiwzcrzloKYnQCx8", + "object": "chat.completion", + "created": 1778514578, + "model": "gpt-5-nano-2025-08-07", + "choices": [ + { + "index": 0, + "message": { + "role": "assistant", + "content": "LLM configuration confirmed: max_tokens set to 4096.\n\nNote: The tool returned “71 degrees,” which seems unrelated to configuration. I’ve treated the setting as 4096 tokens.\n\nWould you like to adjust any other parameters (e.g., temperature, top_p) or run a quick test prompt to verify behavior?", + "refusal": null, + "annotations": [] + }, + "finish_reason": "stop" + } + ], + "usage": { + "prompt_tokens": 182, + "completion_tokens": 395, + "total_tokens": 577, + "prompt_tokens_details": { + "cached_tokens": 0, + "audio_tokens": 0 + }, + "completion_tokens_details": { + "reasoning_tokens": 320, + "audio_tokens": 0, + "accepted_prediction_tokens": 0, + "rejected_prediction_tokens": 0 + } + }, + "service_tier": "default", + "system_fingerprint": null +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/request.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/request.json new file mode 100644 index 00000000..aa626ae7 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/request.json @@ -0,0 +1,32 @@ +{ + "model": "gpt-5-nano", + "messages": [ + { + "role": "user", + "content": "Configure the LLM." + } + ], + "tools": [ + { + "type": "function", + "function": { + "name": "configure_llm", + "description": "Configure LLM generation parameters", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + } + } + } + ] +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response-streaming.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response-streaming.json new file mode 100644 index 00000000..8b607b22 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response-streaming.json @@ -0,0 +1,1208 @@ +[ + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "role": "assistant", + "content": "", + "refusal": null + }, + "finish_reason": null + } + ], + "obfuscation": "gm" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "Sure" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "—" + }, + "finish_reason": null + } + ], + "obfuscation": "U96" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "what" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " max" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "_tokens" + }, + "finish_reason": null + } + ], + "obfuscation": "zIobYwYQSUN2c" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " would" + }, + "finish_reason": null + } + ], + "obfuscation": "xZFgRQnz3GG1mX" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " you" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " like" + }, + "finish_reason": null + } + ], + "obfuscation": "xxGQMIeLvNToydA" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " me" + }, + "finish_reason": null + } + ], + "obfuscation": "I" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " to" + }, + "finish_reason": null + } + ], + "obfuscation": "k" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " configure" + }, + "finish_reason": null + } + ], + "obfuscation": "s07XI9jWn3" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "?\n\n" + }, + "finish_reason": null + } + ], + "obfuscation": "KISY9mhT2qq11VO" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "Common" + }, + "finish_reason": null + } + ], + "obfuscation": "qQgPeF4OJTDnyP" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " options" + }, + "finish_reason": null + } + ], + "obfuscation": "Q0lDW7jncT1L" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": ":\n" + }, + "finish_reason": null + } + ], + "obfuscation": "I" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "-" + }, + "finish_reason": null + } + ], + "obfuscation": "JWx" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " " + }, + "finish_reason": null + } + ], + "obfuscation": "CNr" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "256" + }, + "finish_reason": null + } + ], + "obfuscation": "5" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "\n" + }, + "finish_reason": null + } + ], + "obfuscation": "vM" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "-" + }, + "finish_reason": null + } + ], + "obfuscation": "hdx" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " " + }, + "finish_reason": null + } + ], + "obfuscation": "l4m" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "512" + }, + "finish_reason": null + } + ], + "obfuscation": "h" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "\n" + }, + "finish_reason": null + } + ], + "obfuscation": "xo" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "-" + }, + "finish_reason": null + } + ], + "obfuscation": "o6W" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " " + }, + "finish_reason": null + } + ], + "obfuscation": "vyX" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "102" + }, + "finish_reason": null + } + ], + "obfuscation": "2" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "4" + }, + "finish_reason": null + } + ], + "obfuscation": "mSi" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "\n" + }, + "finish_reason": null + } + ], + "obfuscation": "qr" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "-" + }, + "finish_reason": null + } + ], + "obfuscation": "b18" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " a" + }, + "finish_reason": null + } + ], + "obfuscation": "eA" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " custom" + }, + "finish_reason": null + } + ], + "obfuscation": "dLRR25bXMOxzO" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " number" + }, + "finish_reason": null + } + ], + "obfuscation": "f4MtH7D9IeYa8" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " (" + }, + "finish_reason": null + } + ], + "obfuscation": "zG" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "please" + }, + "finish_reason": null + } + ], + "obfuscation": "hN3DjuL7EFrIZp" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " specify" + }, + "finish_reason": null + } + ], + "obfuscation": "bz4jpaZIWkNJ" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": ")\n\n" + }, + "finish_reason": null + } + ], + "obfuscation": "cs7wT8lIqfkfIsA" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "If" + }, + "finish_reason": null + } + ], + "obfuscation": "Y2" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " you" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "’d" + }, + "finish_reason": null + } + ], + "obfuscation": "wC" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " like" + }, + "finish_reason": null + } + ], + "obfuscation": "vh1eabcAJz0zyxc" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " a" + }, + "finish_reason": null + } + ], + "obfuscation": "VT" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " starting" + }, + "finish_reason": null + } + ], + "obfuscation": "aY3R5bIa0vN" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " point" + }, + "finish_reason": null + } + ], + "obfuscation": "1xEjdnOffWrDjA" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "," + }, + "finish_reason": null + } + ], + "obfuscation": "cpW" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " I" + }, + "finish_reason": null + } + ], + "obfuscation": "Gu" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " can" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " set" + }, + "finish_reason": null + } + ], + "obfuscation": "" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " it" + }, + "finish_reason": null + } + ], + "obfuscation": "o" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " to" + }, + "finish_reason": null + } + ], + "obfuscation": "8" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " " + }, + "finish_reason": null + } + ], + "obfuscation": "WZl" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "512" + }, + "finish_reason": null + } + ], + "obfuscation": "G" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " tokens" + }, + "finish_reason": null + } + ], + "obfuscation": "ZtWYrU2HNG46A" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " as" + }, + "finish_reason": null + } + ], + "obfuscation": "E" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " a" + }, + "finish_reason": null + } + ], + "obfuscation": "Wf" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " balanced" + }, + "finish_reason": null + } + ], + "obfuscation": "YZXZUWI3PfA" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " default" + }, + "finish_reason": null + } + ], + "obfuscation": "qVyGyY869cZv" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "." + }, + "finish_reason": null + } + ], + "obfuscation": "GZV" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " Please" + }, + "finish_reason": null + } + ], + "obfuscation": "SDUMIgHim60Bu" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " confirm" + }, + "finish_reason": null + } + ], + "obfuscation": "8sboLBsKROic" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " or" + }, + "finish_reason": null + } + ], + "obfuscation": "o" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " provide" + }, + "finish_reason": null + } + ], + "obfuscation": "xZEbePX3MyOv" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " your" + }, + "finish_reason": null + } + ], + "obfuscation": "1Ibtw1idb6Hoo69" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " preferred" + }, + "finish_reason": null + } + ], + "obfuscation": "eAc1GLLqbF" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": " value" + }, + "finish_reason": null + } + ], + "obfuscation": "6QVbXk7K7D1C92" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": { + "content": "." + }, + "finish_reason": null + } + ], + "obfuscation": "HSH" + }, + { + "id": "chatcmpl-DeNArigX7wQuWaYcDt3PwMkxSFGuB", + "object": "chat.completion.chunk", + "created": 1778514573, + "model": "gpt-5-nano-2025-08-07", + "service_tier": "default", + "system_fingerprint": null, + "choices": [ + { + "index": 0, + "delta": {}, + "finish_reason": "stop" + } + ], + "obfuscation": "sqaPw56MU2KpbL" + } +] \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response.json b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response.json new file mode 100644 index 00000000..84080699 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/chat-completions/response.json @@ -0,0 +1,45 @@ +{ + "id": "chatcmpl-DeNAsr92qNnlhKlo4VBHEZ7oRucKa", + "object": "chat.completion", + "created": 1778514574, + "model": "gpt-5-nano-2025-08-07", + "choices": [ + { + "index": 0, + "message": { + "role": "assistant", + "content": null, + "tool_calls": [ + { + "id": "call_U4A1cymd03k0mIgSaB4RwJKs", + "type": "function", + "function": { + "name": "configure_llm", + "arguments": "{\"max_tokens\":4096}" + } + } + ], + "refusal": null, + "annotations": [] + }, + "finish_reason": "tool_calls" + } + ], + "usage": { + "prompt_tokens": 148, + "completion_tokens": 411, + "total_tokens": 559, + "prompt_tokens_details": { + "cached_tokens": 0, + "audio_tokens": 0 + }, + "completion_tokens_details": { + "reasoning_tokens": 384, + "audio_tokens": 0, + "accepted_prediction_tokens": 0, + "rejected_prediction_tokens": 0 + } + }, + "service_tier": "default", + "system_fingerprint": null +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/google/error.json b/payloads/snapshots/exclusiveMinimumToolParam/google/error.json new file mode 100644 index 00000000..6b1bf75c --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/google/error.json @@ -0,0 +1,3 @@ +{ + "error": "Error: ApiError: {\n \"error\": {\n \"code\": 400,\n \"message\": \"Invalid JSON payload received. Unknown name \\\"exclusiveMinimum\\\" at 'tools[0].function_declarations[0].parameters.properties[0].value': Cannot find field.\",\n \"status\": \"INVALID_ARGUMENT\",\n \"details\": [\n {\n \"@type\": \"type.googleapis.com/google.rpc.BadRequest\",\n \"fieldViolations\": [\n {\n \"field\": \"tools[0].function_declarations[0].parameters.properties[0].value\",\n \"description\": \"Invalid JSON payload received. Unknown name \\\"exclusiveMinimum\\\" at 'tools[0].function_declarations[0].parameters.properties[0].value': Cannot find field.\"\n }\n ]\n }\n ]\n }\n}\n" +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/google/request.json b/payloads/snapshots/exclusiveMinimumToolParam/google/request.json new file mode 100644 index 00000000..12de61aa --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/google/request.json @@ -0,0 +1,36 @@ +{ + "model": "gemini-2.5-flash", + "contents": [ + { + "role": "user", + "parts": [ + { + "text": "Configure the LLM." + } + ] + } + ], + "tools": [ + { + "functionDeclarations": [ + { + "name": "configure_llm", + "description": "Configure LLM generation parameters", + "parameters": { + "type": "OBJECT", + "properties": { + "max_tokens": { + "type": "NUMBER", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ] + } + } + ] + } + ] +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-request.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-request.json new file mode 100644 index 00000000..26b60180 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-request.json @@ -0,0 +1,54 @@ +{ + "model": "gpt-5-nano", + "input": [ + { + "role": "user", + "content": "Configure the LLM." + }, + { + "id": "rs_0ad9a5f3cf5786be006a01fa8e58ec81979fa3b0f6ce5c327e", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_0ad9a5f3cf5786be006a01fa92bdd48197bad67756895852a3", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. What max_tokens value would you like me to configure? Common options: 256, 512, 1024, 2048, 4096. If you prefer, I can set 2048 by default." + } + ], + "role": "assistant" + }, + { + "role": "user", + "content": "What should I do next?" + } + ], + "tools": [ + { + "type": "function", + "name": "configure_llm", + "description": "Configure LLM generation parameters", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ] +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response-streaming.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response-streaming.json new file mode 100644 index 00000000..051d0873 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response-streaming.json @@ -0,0 +1,1705 @@ +[ + { + "type": "response.created", + "response": { + "id": "resp_0ad9a5f3cf5786be006a01fa960f84819787abf6e8f7efb23e", + "object": "response", + "created_at": 1778514582, + "status": "in_progress", + "background": false, + "completed_at": null, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "auto", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": null, + "user": null, + "metadata": {} + }, + "sequence_number": 0 + }, + { + "type": "response.in_progress", + "response": { + "id": "resp_0ad9a5f3cf5786be006a01fa960f84819787abf6e8f7efb23e", + "object": "response", + "created_at": 1778514582, + "status": "in_progress", + "background": false, + "completed_at": null, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "auto", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": null, + "user": null, + "metadata": {} + }, + "sequence_number": 1 + }, + { + "type": "response.output_item.added", + "item": { + "id": "rs_0ad9a5f3cf5786be006a01fa969fa481979637869b6cfcbe3b", + "type": "reasoning", + "summary": [] + }, + "output_index": 0, + "sequence_number": 2 + }, + { + "type": "response.output_item.done", + "item": { + "id": "rs_0ad9a5f3cf5786be006a01fa969fa481979637869b6cfcbe3b", + "type": "reasoning", + "summary": [] + }, + "output_index": 0, + "sequence_number": 3 + }, + { + "type": "response.output_item.added", + "item": { + "id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "type": "message", + "status": "in_progress", + "content": [], + "role": "assistant" + }, + "output_index": 1, + "sequence_number": 4 + }, + { + "type": "response.content_part.added", + "content_index": 0, + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "output_index": 1, + "part": { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "" + }, + "sequence_number": 5 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "Great", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "cuHwrtyAJZB", + "output_index": 1, + "sequence_number": 6 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " question", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "eazUCFU", + "output_index": 1, + "sequence_number": 7 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "MFsfM9N0gmRq9Zi", + "output_index": 1, + "sequence_number": 8 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " Here", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "OjkJf2t6zob", + "output_index": 1, + "sequence_number": 9 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " are", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "AHwBFlTNqfV5", + "output_index": 1, + "sequence_number": 10 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " practical", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "eJrq09", + "output_index": 1, + "sequence_number": 11 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " next", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "CcwkJMyV8dW", + "output_index": 1, + "sequence_number": 12 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " steps", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "76f5tKsF9A", + "output_index": 1, + "sequence_number": 13 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":\n\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Yk6expDlgwFJa", + "output_index": 1, + "sequence_number": 14 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "1PELa51uy7KB0l3", + "output_index": 1, + "sequence_number": 15 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " Decide", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Nyv4cvq9P", + "output_index": 1, + "sequence_number": 16 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " on", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "We0ZTTHbhx0yT", + "output_index": 1, + "sequence_number": 17 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " max", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "5J4wyclW447x", + "output_index": 1, + "sequence_number": 18 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "_tokens", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "toYFhrURn", + "output_index": 1, + "sequence_number": 19 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " for", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "UDRjPVZsR8Yc", + "output_index": 1, + "sequence_number": 20 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "GMBvqqGYO1Lv", + "output_index": 1, + "sequence_number": 21 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " generated", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "hBmoEv", + "output_index": 1, + "sequence_number": 22 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " output", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "VKAxrC1fe", + "output_index": 1, + "sequence_number": 23 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "BqQkMc1rUNQJiw", + "output_index": 1, + "sequence_number": 24 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "FYOcA9ZX1WLnqh5", + "output_index": 1, + "sequence_number": 25 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " -", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "naM85dAJXCGxZd", + "output_index": 1, + "sequence_number": 26 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "rHkpOag8q4JWq3U", + "output_index": 1, + "sequence_number": 27 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "204", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "4wwwAHA5DtFdw", + "output_index": 1, + "sequence_number": 28 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "8", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "pDdpPMIuvx98BgL", + "output_index": 1, + "sequence_number": 29 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "jgql4nJWGzr74HL", + "output_index": 1, + "sequence_number": 30 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " balanced", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "v3U6fhz", + "output_index": 1, + "sequence_number": 31 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " default", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "9OBfanYf", + "output_index": 1, + "sequence_number": 32 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " for", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "fmZh7NHtRcJO", + "output_index": 1, + "sequence_number": 33 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " longer", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "MwtFj81nK", + "output_index": 1, + "sequence_number": 34 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ",", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "xX8hKqaMLN11rcf", + "output_index": 1, + "sequence_number": 35 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " yet", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "YAuc9sBhZFZO", + "output_index": 1, + "sequence_number": 36 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " still", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "YpBT3Rwsvh", + "output_index": 1, + "sequence_number": 37 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " responsive", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Dg2Mm", + "output_index": 1, + "sequence_number": 38 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " responses", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "96pOIH", + "output_index": 1, + "sequence_number": 39 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "0mXZWR8Bu6LE2l", + "output_index": 1, + "sequence_number": 40 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "JZGrxbV5AMbroGM", + "output_index": 1, + "sequence_number": 41 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " -", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "q8kGt7rAm887rQ", + "output_index": 1, + "sequence_number": 42 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Yydxr8RBoeLfAA7", + "output_index": 1, + "sequence_number": 43 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "102", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "84ltPrOZIhP0B", + "output_index": 1, + "sequence_number": 44 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "4", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "R34Rt3pmRdCSMGC", + "output_index": 1, + "sequence_number": 45 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "wKqSvf7KYaaxX89", + "output_index": 1, + "sequence_number": 46 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " quicker", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "FD1xS3M1", + "output_index": 1, + "sequence_number": 47 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " responses", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "TGLe1B", + "output_index": 1, + "sequence_number": 48 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " with", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "vvv6QVRq5Kz", + "output_index": 1, + "sequence_number": 49 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " shorter", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "AS7f7ajn", + "output_index": 1, + "sequence_number": 50 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " outputs", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "q7JEXirU", + "output_index": 1, + "sequence_number": 51 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Kk2sYNhTqTyn38", + "output_index": 1, + "sequence_number": 52 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "uuFiS9aVfbTIaxg", + "output_index": 1, + "sequence_number": 53 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " -", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "ErdBPOkwFFGdFq", + "output_index": 1, + "sequence_number": 54 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "uo5BPn76qHXfIiH", + "output_index": 1, + "sequence_number": 55 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "409", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "3ziyS3sCmwWOK", + "output_index": 1, + "sequence_number": 56 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "6", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "NNmn6IKvsKulnpy", + "output_index": 1, + "sequence_number": 57 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "uYlKBYDoK3SbUa9", + "output_index": 1, + "sequence_number": 58 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " for", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "GgJ06exMGalY", + "output_index": 1, + "sequence_number": 59 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " very", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "gTgNvCgTFMP", + "output_index": 1, + "sequence_number": 60 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " long", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "DAKk7boqZcg", + "output_index": 1, + "sequence_number": 61 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " outputs", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "pP91ZlUX", + "output_index": 1, + "sequence_number": 62 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "WJhM7mHArB1Xzl", + "output_index": 1, + "sequence_number": 63 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "if", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "l731hYRGREBl9z", + "output_index": 1, + "sequence_number": 64 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " your", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "gsaxasbjf7f", + "output_index": 1, + "sequence_number": 65 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " system", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "gsi24TxOk", + "output_index": 1, + "sequence_number": 66 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " allows", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "5jWqycP04", + "output_index": 1, + "sequence_number": 67 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ").\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "qS5P7MQ4wY6Aj", + "output_index": 1, + "sequence_number": 68 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "lEnsVWKAnOrhss1", + "output_index": 1, + "sequence_number": 69 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " Remember", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "pX6lWk0", + "output_index": 1, + "sequence_number": 70 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "LwNvqiNMS3kswPT", + "output_index": 1, + "sequence_number": 71 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " this", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "HaN1E4T95jk", + "output_index": 1, + "sequence_number": 72 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " controls", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Y4LpAgV", + "output_index": 1, + "sequence_number": 73 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " how", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "ZlAs8gmPRn0a", + "output_index": 1, + "sequence_number": 74 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " many", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "S8iygStWUaZ", + "output_index": 1, + "sequence_number": 75 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " tokens", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "8FbgrN34k", + "output_index": 1, + "sequence_number": 76 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "auSZ4eZEpjlK", + "output_index": 1, + "sequence_number": 77 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " model", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "sV3b8C1N8D", + "output_index": 1, + "sequence_number": 78 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " is", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "GcSuHyq3JFgKv", + "output_index": 1, + "sequence_number": 79 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " allowed", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "EVHSKWqF", + "output_index": 1, + "sequence_number": 80 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " to", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "jdy3UCfWP6LGd", + "output_index": 1, + "sequence_number": 81 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " generate", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "pq4ZTBr", + "output_index": 1, + "sequence_number": 82 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " in", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "RO25hf3dOobpx", + "output_index": 1, + "sequence_number": 83 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " one", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "X9kW6HCD3xRr", + "output_index": 1, + "sequence_number": 84 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " response", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "DN2KCwR", + "output_index": 1, + "sequence_number": 85 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "isZJYmlv0wVUey", + "output_index": 1, + "sequence_number": 86 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "not", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "np1IZZfbigfYJ", + "output_index": 1, + "sequence_number": 87 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "bZ3leEVSSrcm", + "output_index": 1, + "sequence_number": 88 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " input", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "OzMIoIWOWK", + "output_index": 1, + "sequence_number": 89 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " length", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "ez3SVt56t", + "output_index": 1, + "sequence_number": 90 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ").\n\n", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "jRknCbJhMToN", + "output_index": 1, + "sequence_number": 91 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "Because", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Y3jPXjl79", + "output_index": 1, + "sequence_number": 92 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "LIUj3F7WYkN7", + "output_index": 1, + "sequence_number": 93 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " tool", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "YKDTcYC9wIz", + "output_index": 1, + "sequence_number": 94 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " only", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "0UiTTcmqSkk", + "output_index": 1, + "sequence_number": 95 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " config", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "ZnAgbTrqW", + "output_index": 1, + "sequence_number": 96 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "ures", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "vOeMM2RnFMh2", + "output_index": 1, + "sequence_number": 97 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " max", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Cdv3i1FRRR32", + "output_index": 1, + "sequence_number": 98 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "_tokens", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "VPnYpIJkq", + "output_index": 1, + "sequence_number": 99 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " right", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "fuiSolL6PR", + "output_index": 1, + "sequence_number": 100 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " now", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "nqDVGC9eLvd1", + "output_index": 1, + "sequence_number": 101 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ",", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "uSn325MhS7xsZ4N", + "output_index": 1, + "sequence_number": 102 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " I", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "NtxQ7T1q4JtAYH", + "output_index": 1, + "sequence_number": 103 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " suggest", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "qArCjNcC", + "output_index": 1, + "sequence_number": 104 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " starting", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "8e0cmzZ", + "output_index": 1, + "sequence_number": 105 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " with", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "PnplDit08pJ", + "output_index": 1, + "sequence_number": 106 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "3r2Snq89BG9ci2K", + "output_index": 1, + "sequence_number": 107 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "204", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "OuTPpig8t5aUc", + "output_index": 1, + "sequence_number": 108 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "8", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "AtSe18a93368pzK", + "output_index": 1, + "sequence_number": 109 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " as", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "y65TyYooGdx7B", + "output_index": 1, + "sequence_number": 110 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " a", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "pWHD5bv2KXe7AR", + "output_index": 1, + "sequence_number": 111 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " safe", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "c7XA41HXxuv", + "output_index": 1, + "sequence_number": 112 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " default", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "T6t6Y0Fb", + "output_index": 1, + "sequence_number": 113 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "4DJIuWEdJkoVS4e", + "output_index": 1, + "sequence_number": 114 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " Would", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "2mK1hWrY5v", + "output_index": 1, + "sequence_number": 115 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " you", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "GeseVYcFfOf1", + "output_index": 1, + "sequence_number": 116 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " like", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "4moD9FvU9FG", + "output_index": 1, + "sequence_number": 117 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " me", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "B9BacP8807fcD", + "output_index": 1, + "sequence_number": 118 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " to", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "gNOih3pltzdTZ", + "output_index": 1, + "sequence_number": 119 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " set", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "nXQ5DvS2tsI0", + "output_index": 1, + "sequence_number": 120 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " max", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "0jGe1QIDChOu", + "output_index": 1, + "sequence_number": 121 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "_tokens", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "eokYw7aMw", + "output_index": 1, + "sequence_number": 122 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " to", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Tjt141uLcZJsI", + "output_index": 1, + "sequence_number": 123 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "DhJsYrwiSZMBhMm", + "output_index": 1, + "sequence_number": 124 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "204", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "5d7vzBxjDqU4y", + "output_index": 1, + "sequence_number": 125 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "8", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "UldzHlBmhPIKrrn", + "output_index": 1, + "sequence_number": 126 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " now", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "CVaC7bALzTUh", + "output_index": 1, + "sequence_number": 127 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "?", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "UJWyzjhCjYNlfIi", + "output_index": 1, + "sequence_number": 128 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " If", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "Nf5Ym6Wyrgsvb", + "output_index": 1, + "sequence_number": 129 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " you", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "NRt6Me744OK5", + "output_index": 1, + "sequence_number": 130 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " prefer", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "m9urhj3bl", + "output_index": 1, + "sequence_number": 131 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " a", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "9gtmySrS4H8xuE", + "output_index": 1, + "sequence_number": 132 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " different", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "08VY3m", + "output_index": 1, + "sequence_number": 133 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " value", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "mI4cerUTHP", + "output_index": 1, + "sequence_number": 134 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ",", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "whdOeQKeEQd6vd3", + "output_index": 1, + "sequence_number": 135 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " tell", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "EzqJej9vZdN", + "output_index": 1, + "sequence_number": 136 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " me", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "tG9MIWFLynRoR", + "output_index": 1, + "sequence_number": 137 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " and", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "ULbJVmxQR2dq", + "output_index": 1, + "sequence_number": 138 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " I", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "7JKNuIDFCJXpTZ", + "output_index": 1, + "sequence_number": 139 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "’ll", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "AcfukFsWL5zis", + "output_index": 1, + "sequence_number": 140 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " apply", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "v2tHjdPaQt", + "output_index": 1, + "sequence_number": 141 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " it", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "uP10xY5Vk7Nbh", + "output_index": 1, + "sequence_number": 142 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".", + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "obfuscation": "J9NrustlA6MgPy5", + "output_index": 1, + "sequence_number": 143 + }, + { + "type": "response.output_text.done", + "content_index": 0, + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "logprobs": [], + "output_index": 1, + "sequence_number": 144, + "text": "Great question. Here are practical next steps:\n\n- Decide on max_tokens for the generated output:\n - 2048: balanced default for longer, yet still responsive responses.\n - 1024: quicker responses with shorter outputs.\n - 4096: for very long outputs (if your system allows).\n- Remember: this controls how many tokens the model is allowed to generate in one response (not the input length).\n\nBecause the tool only configures max_tokens right now, I suggest starting with 2048 as a safe default. Would you like me to set max_tokens to 2048 now? If you prefer a different value, tell me and I’ll apply it." + }, + { + "type": "response.content_part.done", + "content_index": 0, + "item_id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "output_index": 1, + "part": { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Great question. Here are practical next steps:\n\n- Decide on max_tokens for the generated output:\n - 2048: balanced default for longer, yet still responsive responses.\n - 1024: quicker responses with shorter outputs.\n - 4096: for very long outputs (if your system allows).\n- Remember: this controls how many tokens the model is allowed to generate in one response (not the input length).\n\nBecause the tool only configures max_tokens right now, I suggest starting with 2048 as a safe default. Would you like me to set max_tokens to 2048 now? If you prefer a different value, tell me and I’ll apply it." + }, + "sequence_number": 145 + }, + { + "type": "response.output_item.done", + "item": { + "id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Great question. Here are practical next steps:\n\n- Decide on max_tokens for the generated output:\n - 2048: balanced default for longer, yet still responsive responses.\n - 1024: quicker responses with shorter outputs.\n - 4096: for very long outputs (if your system allows).\n- Remember: this controls how many tokens the model is allowed to generate in one response (not the input length).\n\nBecause the tool only configures max_tokens right now, I suggest starting with 2048 as a safe default. Would you like me to set max_tokens to 2048 now? If you prefer a different value, tell me and I’ll apply it." + } + ], + "role": "assistant" + }, + "output_index": 1, + "sequence_number": 146 + }, + { + "type": "response.completed", + "response": { + "id": "resp_0ad9a5f3cf5786be006a01fa960f84819787abf6e8f7efb23e", + "object": "response", + "created_at": 1778514582, + "status": "completed", + "background": false, + "completed_at": 1778514590, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_0ad9a5f3cf5786be006a01fa969fa481979637869b6cfcbe3b", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_0ad9a5f3cf5786be006a01fa9d01488197b392aec5c49addb2", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Great question. Here are practical next steps:\n\n- Decide on max_tokens for the generated output:\n - 2048: balanced default for longer, yet still responsive responses.\n - 1024: quicker responses with shorter outputs.\n - 4096: for very long outputs (if your system allows).\n- Remember: this controls how many tokens the model is allowed to generate in one response (not the input length).\n\nBecause the tool only configures max_tokens right now, I suggest starting with 2048 as a safe default. Would you like me to set max_tokens to 2048 now? If you prefer a different value, tell me and I’ll apply it." + } + ], + "role": "assistant" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 129, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 994, + "output_tokens_details": { + "reasoning_tokens": 832 + }, + "total_tokens": 1123 + }, + "user": null, + "metadata": {} + }, + "sequence_number": 147 + } +] \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response.json new file mode 100644 index 00000000..9c0f565f --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/followup-response.json @@ -0,0 +1,99 @@ +{ + "id": "resp_0ad9a5f3cf5786be006a01fa96cb688197bc6358a913c2627f", + "object": "response", + "created_at": 1778514582, + "status": "completed", + "background": false, + "billing": { + "payer": "developer" + }, + "completed_at": 1778514591, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_0ad9a5f3cf5786be006a01fa9783048197aec499ed761b01c6", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_0ad9a5f3cf5786be006a01fa9e4d7c81978c523b36caf80554", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Here’s the next step:\n\n- Decide the max_tokens you want for responses. This controls how long the model’s answers can be.\n - Short/concise: 256 or 512\n - Balanced: 1024\n - Longer answers: 2048 or 4096\n\nWould you like me to set 2048 as the default max_tokens now? If you prefer a different value, tell me and I’ll configure it. After configuring, we can test with a prompt to confirm the length works as you expect." + } + ], + "role": "assistant" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 129, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 966, + "output_tokens_details": { + "reasoning_tokens": 832 + }, + "total_tokens": 1095 + }, + "user": null, + "metadata": {}, + "output_text": "Here’s the next step:\n\n- Decide the max_tokens you want for responses. This controls how long the model’s answers can be.\n - Short/concise: 256 or 512\n - Balanced: 1024\n - Longer answers: 2048 or 4096\n\nWould you like me to set 2048 as the default max_tokens now? If you prefer a different value, tell me and I’ll configure it. After configuring, we can test with a prompt to confirm the length works as you expect." +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/request.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/request.json new file mode 100644 index 00000000..b1811630 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/request.json @@ -0,0 +1,31 @@ +{ + "model": "gpt-5-nano", + "input": [ + { + "role": "user", + "content": "Configure the LLM." + } + ], + "tools": [ + { + "type": "function", + "name": "configure_llm", + "description": "Configure LLM generation parameters", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ] +} \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/response-streaming.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/response-streaming.json new file mode 100644 index 00000000..d0ed1fbe --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/response-streaming.json @@ -0,0 +1,1115 @@ +[ + { + "type": "response.created", + "response": { + "id": "resp_016df566c9f94231006a01fa8da39081979444292d91c535df", + "object": "response", + "created_at": 1778514573, + "status": "in_progress", + "background": false, + "completed_at": null, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "auto", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": null, + "user": null, + "metadata": {} + }, + "sequence_number": 0 + }, + { + "type": "response.in_progress", + "response": { + "id": "resp_016df566c9f94231006a01fa8da39081979444292d91c535df", + "object": "response", + "created_at": 1778514573, + "status": "in_progress", + "background": false, + "completed_at": null, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "auto", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": null, + "user": null, + "metadata": {} + }, + "sequence_number": 1 + }, + { + "type": "response.output_item.added", + "item": { + "id": "rs_016df566c9f94231006a01fa8e945c8197894a63776c91baff", + "type": "reasoning", + "summary": [] + }, + "output_index": 0, + "sequence_number": 2 + }, + { + "type": "response.output_item.done", + "item": { + "id": "rs_016df566c9f94231006a01fa8e945c8197894a63776c91baff", + "type": "reasoning", + "summary": [] + }, + "output_index": 0, + "sequence_number": 3 + }, + { + "type": "response.output_item.added", + "item": { + "id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "type": "message", + "status": "in_progress", + "content": [], + "role": "assistant" + }, + "output_index": 1, + "sequence_number": 4 + }, + { + "type": "response.content_part.added", + "content_index": 0, + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "output_index": 1, + "part": { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "" + }, + "sequence_number": 5 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "Sure", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "McyDl7RV2Ndc", + "output_index": 1, + "sequence_number": 6 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "dTIYpbKVX8XJi84", + "output_index": 1, + "sequence_number": 7 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " What", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "pYgx4XcoLJW", + "output_index": 1, + "sequence_number": 8 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " max", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "j82ZlyUAXjKm", + "output_index": 1, + "sequence_number": 9 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "_tokens", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "2ygsm83GI", + "output_index": 1, + "sequence_number": 10 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " would", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "GhHICaYWVV", + "output_index": 1, + "sequence_number": 11 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " you", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "ng7GFyBkCg4l", + "output_index": 1, + "sequence_number": 12 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " like", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "kVbSZKPanE9", + "output_index": 1, + "sequence_number": 13 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " me", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "iD6HVo4wXWWAN", + "output_index": 1, + "sequence_number": 14 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " to", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "OCquCIrZK0GCp", + "output_index": 1, + "sequence_number": 15 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " set", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "zzhDIq8Rcu6M", + "output_index": 1, + "sequence_number": 16 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " for", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "WufTyJtQOXre", + "output_index": 1, + "sequence_number": 17 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " generation", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "AWqDl", + "output_index": 1, + "sequence_number": 18 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "?\n\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "DYbgzJFZI55Ls", + "output_index": 1, + "sequence_number": 19 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "Some", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "IepMYWpxuI8b", + "output_index": 1, + "sequence_number": 20 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " common", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "rQ7vN9Z63", + "output_index": 1, + "sequence_number": 21 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " options", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "KJXT9e1V", + "output_index": 1, + "sequence_number": 22 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ":\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "obagg7F2EhyNon", + "output_index": 1, + "sequence_number": 23 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "oz2zBKLoWOlWftf", + "output_index": 1, + "sequence_number": 24 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "1OsgiRTh9Yab5hx", + "output_index": 1, + "sequence_number": 25 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "256", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "2NB4IdGSC30Gt", + "output_index": 1, + "sequence_number": 26 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "eizxSEr7v5qceG", + "output_index": 1, + "sequence_number": 27 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "short", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "BQ26MFwBY4t", + "output_index": 1, + "sequence_number": 28 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " answers", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "ACgfYIyg", + "output_index": 1, + "sequence_number": 29 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ")\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "M6Ee8l5GQ4UCoo", + "output_index": 1, + "sequence_number": 30 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "TloJicXISwtEMa8", + "output_index": 1, + "sequence_number": 31 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "nznmrzAvvLxVh4k", + "output_index": 1, + "sequence_number": 32 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "512", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "g4zTg8ovautc1", + "output_index": 1, + "sequence_number": 33 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "irCZQfUJtJAfNo", + "output_index": 1, + "sequence_number": 34 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "moder", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "f8l50ji9ljc", + "output_index": 1, + "sequence_number": 35 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "ate", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "U3C5tHaQF8FaS", + "output_index": 1, + "sequence_number": 36 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " length", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "KBAif3KJv", + "output_index": 1, + "sequence_number": 37 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ")\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "LpXhDfCIwSHvou", + "output_index": 1, + "sequence_number": 38 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "cR6DTuAIPR3e3Mo", + "output_index": 1, + "sequence_number": 39 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "7cErGRNBKioIFzM", + "output_index": 1, + "sequence_number": 40 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "102", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "o9UyMxlrmpzOL", + "output_index": 1, + "sequence_number": 41 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "4", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "K5Udsfl0GvB3FFq", + "output_index": 1, + "sequence_number": 42 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "R2Xvm8XRhZSwbX", + "output_index": 1, + "sequence_number": 43 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "d", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "SMGau4mUmOX5ina", + "output_index": 1, + "sequence_number": 44 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "etailed", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "gVzTb8DkO", + "output_index": 1, + "sequence_number": 45 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " responses", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "xj9YXD", + "output_index": 1, + "sequence_number": 46 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ")\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "SZ4jFpgyThqQCd", + "output_index": 1, + "sequence_number": 47 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "GjyL71ccETWamQe", + "output_index": 1, + "sequence_number": 48 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "aAejchT0RcKqJBM", + "output_index": 1, + "sequence_number": 49 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "204", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "qLZMgP7vNPrHF", + "output_index": 1, + "sequence_number": 50 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "8", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "4v0Lrm5lDSEqeGy", + "output_index": 1, + "sequence_number": 51 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "v9zPTbMbuDd2fM", + "output_index": 1, + "sequence_number": 52 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "long", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "6cKIxc5Z4HdI", + "output_index": 1, + "sequence_number": 53 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "-form", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "Vz1DiGXFmJM", + "output_index": 1, + "sequence_number": 54 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ")\n\n", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "spTQ6qvBUlfG4", + "output_index": 1, + "sequence_number": 55 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "If", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "3D1hVSREvWdizQ", + "output_index": 1, + "sequence_number": 56 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " you", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "6EDrvhz8VFFl", + "output_index": 1, + "sequence_number": 57 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "’re", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "xEcayq5p68lXH", + "output_index": 1, + "sequence_number": 58 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " not", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "ELOP8x1yrjWK", + "output_index": 1, + "sequence_number": 59 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " sure", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "VGvSnqJNgRG", + "output_index": 1, + "sequence_number": 60 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ",", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "gOoWSeD4GnER13p", + "output_index": 1, + "sequence_number": 61 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " I", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "nm0TcMKWhMZtcK", + "output_index": 1, + "sequence_number": 62 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " can", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "627kZKMdxb2E", + "output_index": 1, + "sequence_number": 63 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " set", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "D4tRhZ0wG9dt", + "output_index": 1, + "sequence_number": 64 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " a", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "aQnpLz5biy4024", + "output_index": 1, + "sequence_number": 65 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " balanced", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "gAUv9iP", + "output_index": 1, + "sequence_number": 66 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " default", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "quyMCBGS", + "output_index": 1, + "sequence_number": 67 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " of", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "QU8IwxHpsgqVE", + "output_index": 1, + "sequence_number": 68 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " ", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "S5k1boRlBrs9BIc", + "output_index": 1, + "sequence_number": 69 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "102", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "aJfuArG59zJBT", + "output_index": 1, + "sequence_number": 70 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "4", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "W2ccWfkm0IR18A6", + "output_index": 1, + "sequence_number": 71 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ".", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "TrryZMym50vD2VD", + "output_index": 1, + "sequence_number": 72 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " Tell", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "jgLvT6uVwcB", + "output_index": 1, + "sequence_number": 73 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " me", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "Qzzh3ey25P53x", + "output_index": 1, + "sequence_number": 74 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "pgOdlxoF2dWc", + "output_index": 1, + "sequence_number": 75 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " value", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "z1oHVE0RZT", + "output_index": 1, + "sequence_number": 76 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " you", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "JbM4Iz7gQDsR", + "output_index": 1, + "sequence_number": 77 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " want", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "1nqEfoXC5Nr", + "output_index": 1, + "sequence_number": 78 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " (", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "pjFsxCyYrqrZop", + "output_index": 1, + "sequence_number": 79 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": "or", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "F8SV2mn3ZV3Z5I", + "output_index": 1, + "sequence_number": 80 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " confirm", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "vrtQ3dLW", + "output_index": 1, + "sequence_number": 81 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " the", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "3ZbX7do7L8y2", + "output_index": 1, + "sequence_number": 82 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": " default", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "oMKqKPrx", + "output_index": 1, + "sequence_number": 83 + }, + { + "type": "response.output_text.delta", + "content_index": 0, + "delta": ").", + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "obfuscation": "CkaVr8mq7AN5qz", + "output_index": 1, + "sequence_number": 84 + }, + { + "type": "response.output_text.done", + "content_index": 0, + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "logprobs": [], + "output_index": 1, + "sequence_number": 85, + "text": "Sure. What max_tokens would you like me to set for generation?\n\nSome common options:\n- 256 (short answers)\n- 512 (moderate length)\n- 1024 (detailed responses)\n- 2048 (long-form)\n\nIf you’re not sure, I can set a balanced default of 1024. Tell me the value you want (or confirm the default)." + }, + { + "type": "response.content_part.done", + "content_index": 0, + "item_id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "output_index": 1, + "part": { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. What max_tokens would you like me to set for generation?\n\nSome common options:\n- 256 (short answers)\n- 512 (moderate length)\n- 1024 (detailed responses)\n- 2048 (long-form)\n\nIf you’re not sure, I can set a balanced default of 1024. Tell me the value you want (or confirm the default)." + }, + "sequence_number": 86 + }, + { + "type": "response.output_item.done", + "item": { + "id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. What max_tokens would you like me to set for generation?\n\nSome common options:\n- 256 (short answers)\n- 512 (moderate length)\n- 1024 (detailed responses)\n- 2048 (long-form)\n\nIf you’re not sure, I can set a balanced default of 1024. Tell me the value you want (or confirm the default)." + } + ], + "role": "assistant" + }, + "output_index": 1, + "sequence_number": 87 + }, + { + "type": "response.completed", + "response": { + "id": "resp_016df566c9f94231006a01fa8da39081979444292d91c535df", + "object": "response", + "created_at": 1778514573, + "status": "completed", + "background": false, + "completed_at": 1778514581, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_016df566c9f94231006a01fa8e945c8197894a63776c91baff", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_016df566c9f94231006a01fa94f16c8197a4bb0e5d62d5ea4e", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. What max_tokens would you like me to set for generation?\n\nSome common options:\n- 256 (short answers)\n- 512 (moderate length)\n- 1024 (detailed responses)\n- 2048 (long-form)\n\nIf you’re not sure, I can set a balanced default of 1024. Tell me the value you want (or confirm the default)." + } + ], + "role": "assistant" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 66, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 885, + "output_tokens_details": { + "reasoning_tokens": 768 + }, + "total_tokens": 951 + }, + "user": null, + "metadata": {} + }, + "sequence_number": 88 + } +] \ No newline at end of file diff --git a/payloads/snapshots/exclusiveMinimumToolParam/responses/response.json b/payloads/snapshots/exclusiveMinimumToolParam/responses/response.json new file mode 100644 index 00000000..6fe8b2f6 --- /dev/null +++ b/payloads/snapshots/exclusiveMinimumToolParam/responses/response.json @@ -0,0 +1,99 @@ +{ + "id": "resp_0ad9a5f3cf5786be006a01fa8da28c819797a13e41babc185e", + "object": "response", + "created_at": 1778514573, + "status": "completed", + "background": false, + "billing": { + "payer": "developer" + }, + "completed_at": 1778514579, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_0ad9a5f3cf5786be006a01fa8e58ec81979fa3b0f6ce5c327e", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_0ad9a5f3cf5786be006a01fa92bdd48197bad67756895852a3", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. What max_tokens value would you like me to configure? Common options: 256, 512, 1024, 2048, 4096. If you prefer, I can set 2048 by default." + } + ], + "role": "assistant" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "type": "object", + "properties": { + "max_tokens": { + "type": "number", + "exclusiveMinimum": 0, + "description": "Maximum number of tokens to generate" + } + }, + "required": [ + "max_tokens" + ], + "additionalProperties": false + }, + "strict": false + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 66, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 603, + "output_tokens_details": { + "reasoning_tokens": 512 + }, + "total_tokens": 669 + }, + "user": null, + "metadata": {}, + "output_text": "Sure. What max_tokens value would you like me to configure? Common options: 256, 512, 1024, 2048, 4096. If you prefer, I can set 2048 by default." +} \ No newline at end of file diff --git a/payloads/transforms/chat-completions_to_anthropic/exclusiveMinimumToolParam.json b/payloads/transforms/chat-completions_to_anthropic/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..f7433188 --- /dev/null +++ b/payloads/transforms/chat-completions_to_anthropic/exclusiveMinimumToolParam.json @@ -0,0 +1,27 @@ +{ + "model": "claude-sonnet-4-5-20250929", + "id": "msg_01GG7qfqnsDh9GxCJn5VWgPx", + "type": "message", + "role": "assistant", + "content": [ + { + "type": "text", + "text": "I'd be happy to help configure the LLM! However, I need to know what value you'd like to set for the `max_tokens` parameter.\n\nThe `max_tokens` parameter controls the maximum number of tokens (pieces of text) that the LLM can generate in its response.\n\nCould you please specify:\n- **max_tokens**: The maximum number of tokens you want the LLM to generate (must be greater than 0)\n\nFor example, common values might be:\n- 100-500 for short responses\n- 1000-2000 for medium-length responses\n- 4000+ for longer, detailed responses\n\nWhat value would you like to use?" + } + ], + "stop_reason": "end_turn", + "stop_sequence": null, + "stop_details": null, + "usage": { + "input_tokens": 596, + "cache_creation_input_tokens": 0, + "cache_read_input_tokens": 0, + "cache_creation": { + "ephemeral_5m_input_tokens": 0, + "ephemeral_1h_input_tokens": 0 + }, + "output_tokens": 154, + "service_tier": "standard", + "inference_geo": "not_available" + } +} \ No newline at end of file diff --git a/payloads/transforms/chat-completions_to_google/exclusiveMinimumToolParam.json b/payloads/transforms/chat-completions_to_google/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..a68fcfee --- /dev/null +++ b/payloads/transforms/chat-completions_to_google/exclusiveMinimumToolParam.json @@ -0,0 +1,32 @@ +{ + "candidates": [ + { + "content": { + "parts": [ + { + "text": "I can help with that. What would you like to set the maximum number of tokens to?", + "thoughtSignature": "Cu4BAQw51sfM8F31TciEgkRESm+9CBEfVVvQodbTNDK5zpNisnX40nuUDHIlGCOYO8LrBiGdNTCfQYFNBGqhHebYUKcu3lJTIHYSjGHcFhpCpYhuxFV+SXot99S+HqRHrGrCDVy4UoDk2+LInrKrXFb+j9iNfv8lC3MlQZ2qlV2+Orell/YC80wfgGZz3a7TJOGwjeF7ghsu/1mzqHd4tZDZ611uYrxSUH3NnZTwdt8HYNbU1GIsZbDz/biAqwG6FdcjnQ4Hxy2LdCKasUJoVTcGc13PdZ64cHJf3KGqQ6fDEzaZl6IqzedPQCuIgJpKLQ==" + } + ], + "role": "model" + }, + "finishReason": "STOP", + "index": 0 + } + ], + "usageMetadata": { + "promptTokenCount": 53, + "candidatesTokenCount": 19, + "totalTokenCount": 120, + "promptTokensDetails": [ + { + "modality": "TEXT", + "tokenCount": 53 + } + ], + "thoughtsTokenCount": 48, + "serviceTier": "standard" + }, + "modelVersion": "gemini-2.5-flash", + "responseId": "cvsBaqjuBrOw1MkP1srUuQ8" +} \ No newline at end of file diff --git a/payloads/transforms/chat-completions_to_responses/exclusiveMinimumToolParam.json b/payloads/transforms/chat-completions_to_responses/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..3e3dd2be --- /dev/null +++ b/payloads/transforms/chat-completions_to_responses/exclusiveMinimumToolParam.json @@ -0,0 +1,99 @@ +{ + "id": "resp_0038cce6fceb0e88006a01faa00e5c8195921786d0852993cd", + "object": "response", + "created_at": 1778514592, + "status": "completed", + "background": false, + "billing": { + "payer": "developer" + }, + "completed_at": 1778514601, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_0038cce6fceb0e88006a01faa1d9d08195aa33d7aff4728a07", + "type": "reasoning", + "summary": [] + }, + { + "id": "msg_0038cce6fceb0e88006a01faa7a4c48195937d5eacf3744da5", + "type": "message", + "status": "completed", + "content": [ + { + "type": "output_text", + "annotations": [], + "logprobs": [], + "text": "Sure. I can configure the LLM, but I need a max_tokens value to apply.\n\nOptions (balanced defaults):\n- 512 – concise responses\n- 1024 – short to medium\n- 2048 – balanced\n- 4096 – longer responses\n\nIf you’d like, I can apply 2048 tokens right now as a balanced default. Would you like me to set max_tokens to 2048 (or choose another value)? Also, note that this configuration step only covers max_tokens; tell me if you want me to note any other preferences, and I can track them for you." + } + ], + "role": "assistant" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "additionalProperties": false, + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "exclusiveMinimum": 0, + "type": "number" + } + }, + "required": [ + "max_tokens" + ], + "type": "object" + }, + "strict": true + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 66, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 768, + "output_tokens_details": { + "reasoning_tokens": 576 + }, + "total_tokens": 834 + }, + "user": null, + "metadata": {}, + "output_text": "Sure. I can configure the LLM, but I need a max_tokens value to apply.\n\nOptions (balanced defaults):\n- 512 – concise responses\n- 1024 – short to medium\n- 2048 – balanced\n- 4096 – longer responses\n\nIf you’d like, I can apply 2048 tokens right now as a balanced default. Would you like me to set max_tokens to 2048 (or choose another value)? Also, note that this configuration step only covers max_tokens; tell me if you want me to note any other preferences, and I can track them for you." +} \ No newline at end of file diff --git a/payloads/transforms/google_to_anthropic/exclusiveMinimumToolParam.json b/payloads/transforms/google_to_anthropic/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..069802dd --- /dev/null +++ b/payloads/transforms/google_to_anthropic/exclusiveMinimumToolParam.json @@ -0,0 +1,27 @@ +{ + "model": "claude-sonnet-4-5-20250929", + "id": "msg_01Ej1s2o98kH3B4KEjTa7uTR", + "type": "message", + "role": "assistant", + "content": [ + { + "type": "text", + "text": "I'd be happy to help configure the LLM! However, I need to know what value you'd like to set for the maximum number of tokens to generate.\n\nCould you please specify the `max_tokens` value you want? For example:\n- 256 for shorter responses\n- 512 for medium-length responses\n- 1024 or higher for longer responses\n- 4096 for very detailed responses\n\nWhat maximum token limit would you like to set?" + } + ], + "stop_reason": "end_turn", + "stop_sequence": null, + "stop_details": null, + "usage": { + "input_tokens": 579, + "cache_creation_input_tokens": 0, + "cache_read_input_tokens": 0, + "cache_creation": { + "ephemeral_5m_input_tokens": 0, + "ephemeral_1h_input_tokens": 0 + }, + "output_tokens": 105, + "service_tier": "standard", + "inference_geo": "not_available" + } +} \ No newline at end of file diff --git a/payloads/transforms/google_to_chat-completions/exclusiveMinimumToolParam.json b/payloads/transforms/google_to_chat-completions/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..6ec9c618 --- /dev/null +++ b/payloads/transforms/google_to_chat-completions/exclusiveMinimumToolParam.json @@ -0,0 +1,35 @@ +{ + "id": "chatcmpl-DeNBBnm25aBmS3HfNYyqsq5Gtl4XU", + "object": "chat.completion", + "created": 1778514593, + "model": "gpt-5-nano-2025-08-07", + "choices": [ + { + "index": 0, + "message": { + "role": "assistant", + "content": "I can configure the LLM, but I need the max_tokens value (the maximum number of tokens in responses).\n\nOptions:\n- 512 (short responses)\n- 1024 (default)\n- 2048 (longer responses)\n\nWould you like me to set max_tokens to 1024 now, or provide a different value? If you’re unsure, I can start with 1024 and adjust later.", + "refusal": null, + "annotations": [] + }, + "finish_reason": "stop" + } + ], + "usage": { + "prompt_tokens": 139, + "completion_tokens": 797, + "total_tokens": 936, + "prompt_tokens_details": { + "cached_tokens": 0, + "audio_tokens": 0 + }, + "completion_tokens_details": { + "reasoning_tokens": 704, + "audio_tokens": 0, + "accepted_prediction_tokens": 0, + "rejected_prediction_tokens": 0 + } + }, + "service_tier": "default", + "system_fingerprint": null +} \ No newline at end of file diff --git a/payloads/transforms/google_to_responses/exclusiveMinimumToolParam.json b/payloads/transforms/google_to_responses/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..2880463c --- /dev/null +++ b/payloads/transforms/google_to_responses/exclusiveMinimumToolParam.json @@ -0,0 +1,92 @@ +{ + "id": "resp_0c73324acbddd2b0006a01faa3fa5c8193b4c07d00b43d01c4", + "object": "response", + "created_at": 1778514596, + "status": "completed", + "background": false, + "billing": { + "payer": "developer" + }, + "completed_at": 1778514600, + "error": null, + "frequency_penalty": 0, + "incomplete_details": null, + "instructions": null, + "max_output_tokens": null, + "max_tool_calls": null, + "model": "gpt-5-nano-2025-08-07", + "moderation": null, + "output": [ + { + "id": "rs_0c73324acbddd2b0006a01faa44ff08193b878fb061da79bf7", + "type": "reasoning", + "summary": [] + }, + { + "id": "fc_0c73324acbddd2b0006a01faa8069481939eaaa83abc7b561e", + "type": "function_call", + "status": "completed", + "arguments": "{\"max_tokens\": 1024}", + "call_id": "call_GSbXCURhGXAkZqWeKF5syU0P", + "name": "configure_llm" + } + ], + "parallel_tool_calls": true, + "presence_penalty": 0, + "previous_response_id": null, + "prompt_cache_key": null, + "prompt_cache_retention": "in_memory", + "reasoning": { + "effort": "medium", + "summary": null + }, + "safety_identifier": null, + "service_tier": "default", + "store": true, + "temperature": 1, + "text": { + "format": { + "type": "text" + }, + "verbosity": "medium" + }, + "tool_choice": "auto", + "tools": [ + { + "type": "function", + "description": "Configure LLM generation parameters", + "name": "configure_llm", + "parameters": { + "properties": { + "max_tokens": { + "description": "Maximum number of tokens to generate", + "type": "number" + } + }, + "required": [ + "max_tokens" + ], + "type": "object", + "additionalProperties": false + }, + "strict": true + } + ], + "top_logprobs": 0, + "top_p": 1, + "truncation": "disabled", + "usage": { + "input_tokens": 57, + "input_tokens_details": { + "cached_tokens": 0 + }, + "output_tokens": 587, + "output_tokens_details": { + "reasoning_tokens": 512 + }, + "total_tokens": 644 + }, + "user": null, + "metadata": {}, + "output_text": "" +} \ No newline at end of file diff --git a/payloads/transforms/responses_to_anthropic/exclusiveMinimumToolParam.json b/payloads/transforms/responses_to_anthropic/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..8ac40c8b --- /dev/null +++ b/payloads/transforms/responses_to_anthropic/exclusiveMinimumToolParam.json @@ -0,0 +1,27 @@ +{ + "model": "claude-sonnet-4-5-20250929", + "id": "msg_01ERkvPjbQ7ZLjFZns3hskpK", + "type": "message", + "role": "assistant", + "content": [ + { + "type": "text", + "text": "I'd be happy to help configure the LLM! To do this, I need to know what value you'd like to set for the **max_tokens** parameter, which controls the maximum number of tokens the LLM can generate in its responses.\n\nCould you please specify:\n- How many max_tokens would you like to configure? (This should be a positive number)\n\nFor context, common values might be:\n- 100-500 for short responses\n- 500-2000 for medium-length responses\n- 2000+ for longer, more detailed responses" + } + ], + "stop_reason": "end_turn", + "stop_sequence": null, + "stop_details": null, + "usage": { + "input_tokens": 596, + "cache_creation_input_tokens": 0, + "cache_read_input_tokens": 0, + "cache_creation": { + "ephemeral_5m_input_tokens": 0, + "ephemeral_1h_input_tokens": 0 + }, + "output_tokens": 127, + "service_tier": "standard", + "inference_geo": "not_available" + } +} \ No newline at end of file diff --git a/payloads/transforms/responses_to_google/exclusiveMinimumToolParam.json b/payloads/transforms/responses_to_google/exclusiveMinimumToolParam.json new file mode 100644 index 00000000..c37c754f --- /dev/null +++ b/payloads/transforms/responses_to_google/exclusiveMinimumToolParam.json @@ -0,0 +1,32 @@ +{ + "candidates": [ + { + "content": { + "parts": [ + { + "text": "Please provide the `max_tokens` parameter to configure the LLM.", + "thoughtSignature": "CokCAQw51scOFKbgK5bO/X1BLBLzjBzYmtbGvqe5Ct/W+tkUSoFxY91vKIVqLrFNMUtuHyaDjXlS84QEz6Am96xV1F6QtwJrjiCLO+05PK6UiWtG4KTWTbtRXXQBeMxPkcTjGf0WBVz/DYCz3gSBMRgTYjmIfxscIVM0dhNtluqp8p1rnc3kPcl6Futaav7id4d06hgqOff/8Usvfa0XNGYE5GHaIw8yzNuoSJC40zLw51nCGcAW5HJne1jE4Ch4aT4g9H4rzQ0L7z7CL+rmTmlYaA6JbNdcnrcUEnPPNncfFLU94A3cMEDB/zbuOTuvjPEz8QWhkwAUmtex1g81jSba32o8vwqtMGQC8w==" + } + ], + "role": "model" + }, + "finishReason": "STOP", + "index": 0 + } + ], + "usageMetadata": { + "promptTokenCount": 53, + "candidatesTokenCount": 15, + "totalTokenCount": 127, + "promptTokensDetails": [ + { + "modality": "TEXT", + "tokenCount": 53 + } + ], + "thoughtsTokenCount": 59, + "serviceTier": "standard" + }, + "modelVersion": "gemini-2.5-flash", + "responseId": "c_sBavy5HIOc_uMP7reJoAc" +} \ No newline at end of file