Skip to content
Draft
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 7 additions & 7 deletions libs/partners/openai/langchain_openai/chat_models/base.py
Original file line number Diff line number Diff line change
Expand Up @@ -3987,7 +3987,7 @@ def _construct_lc_result_from_responses_api(
{"type": "refusal", "refusal": content.refusal, "id": output.id}
)
elif output.type == "function_call":
content_blocks.append(output.model_dump(exclude_none=True, mode="json"))
content_blocks.append(output.model_dump(exclude_none=False, mode="json"))
try:
args = json.loads(output.arguments, strict=False)
error = None
Expand All @@ -4012,7 +4012,7 @@ def _construct_lc_result_from_responses_api(
}
invalid_tool_calls.append(tool_call)
elif output.type == "custom_tool_call":
content_blocks.append(output.model_dump(exclude_none=True, mode="json"))
content_blocks.append(output.model_dump(exclude_none=False, mode="json"))
tool_call = {
"type": "tool_call",
"name": output.name,
Expand All @@ -4031,7 +4031,7 @@ def _construct_lc_result_from_responses_api(
"mcp_approval_request",
"image_generation_call",
):
content_blocks.append(output.model_dump(exclude_none=True, mode="json"))
content_blocks.append(output.model_dump(exclude_none=False, mode="json"))

# Workaround for parsing structured output in the streaming case.
# from openai import OpenAI
Expand Down Expand Up @@ -4160,7 +4160,7 @@ def _advance(output_idx: int, sub_idx: int | None = None) -> None:
# Appears to be a breaking change in openai==1.82.0
annotation = chunk.annotation
else:
annotation = chunk.annotation.model_dump(exclude_none=True, mode="json")
annotation = chunk.annotation.model_dump(exclude_none=False, mode="json")

content.append(
{
Expand Down Expand Up @@ -4232,15 +4232,15 @@ def _advance(output_idx: int, sub_idx: int | None = None) -> None:
"image_generation_call",
):
_advance(chunk.output_index)
tool_output = chunk.item.model_dump(exclude_none=True, mode="json")
tool_output = chunk.item.model_dump(exclude_none=False, mode="json")
tool_output["index"] = current_index
content.append(tool_output)
elif (
chunk.type == "response.output_item.done"
and chunk.item.type == "custom_tool_call"
):
_advance(chunk.output_index)
tool_output = chunk.item.model_dump(exclude_none=True, mode="json")
tool_output = chunk.item.model_dump(exclude_none=False, mode="json")
tool_output["index"] = current_index
content.append(tool_output)
tool_call_chunks.append(
Expand All @@ -4265,7 +4265,7 @@ def _advance(output_idx: int, sub_idx: int | None = None) -> None:
elif chunk.type == "response.output_item.added" and chunk.item.type == "reasoning":
_advance(chunk.output_index)
current_sub_index = 0
reasoning = chunk.item.model_dump(exclude_none=True, mode="json")
reasoning = chunk.item.model_dump(exclude_none=False, mode="json")
reasoning["index"] = current_index
content.append(reasoning)
elif chunk.type == "response.reasoning_summary_part.added":
Expand Down
Loading