mirror of
https://github.com/hwchase17/langchain.git
synced 2025-08-13 22:59:05 +00:00
fix(openai): fix test on standard outputs branch (#32329)
This commit is contained in:
parent
a0abb79f6d
commit
3afcd6132e
@ -1,8 +1,8 @@
|
|||||||
from typing import Any, Literal, Optional
|
from typing import Any, Optional
|
||||||
from unittest.mock import MagicMock, patch
|
from unittest.mock import MagicMock, patch
|
||||||
|
|
||||||
import pytest
|
|
||||||
from langchain_core.messages import AIMessageChunk, BaseMessageChunk
|
from langchain_core.messages import AIMessageChunk, BaseMessageChunk
|
||||||
|
from langchain_core.messages.v1 import AIMessageChunk as AIMessageChunkV1
|
||||||
from openai.types.responses import (
|
from openai.types.responses import (
|
||||||
ResponseCompletedEvent,
|
ResponseCompletedEvent,
|
||||||
ResponseContentPartAddedEvent,
|
ResponseContentPartAddedEvent,
|
||||||
@ -37,7 +37,7 @@ from openai.types.responses.response_usage import (
|
|||||||
from openai.types.shared.reasoning import Reasoning
|
from openai.types.shared.reasoning import Reasoning
|
||||||
from openai.types.shared.response_format_text import ResponseFormatText
|
from openai.types.shared.response_format_text import ResponseFormatText
|
||||||
|
|
||||||
from langchain_openai import ChatOpenAI
|
from langchain_openai import ChatOpenAI, ChatOpenAIV1
|
||||||
from tests.unit_tests.chat_models.test_base import MockSyncContextManager
|
from tests.unit_tests.chat_models.test_base import MockSyncContextManager
|
||||||
|
|
||||||
responses_stream = [
|
responses_stream = [
|
||||||
@ -621,98 +621,9 @@ def _strip_none(obj: Any) -> Any:
|
|||||||
return obj
|
return obj
|
||||||
|
|
||||||
|
|
||||||
@pytest.mark.parametrize(
|
def test_responses_stream() -> None:
|
||||||
"output_version, expected_content",
|
|
||||||
[
|
|
||||||
(
|
|
||||||
"responses/v1",
|
|
||||||
[
|
|
||||||
{
|
|
||||||
"id": "rs_123",
|
|
||||||
"summary": [
|
|
||||||
{
|
|
||||||
"index": 0,
|
|
||||||
"type": "summary_text",
|
|
||||||
"text": "reasoning block one",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"index": 1,
|
|
||||||
"type": "summary_text",
|
|
||||||
"text": "another reasoning block",
|
|
||||||
},
|
|
||||||
],
|
|
||||||
"type": "reasoning",
|
|
||||||
"index": 0,
|
|
||||||
},
|
|
||||||
{"type": "text", "text": "text block one", "index": 1, "id": "msg_123"},
|
|
||||||
{
|
|
||||||
"type": "text",
|
|
||||||
"text": "another text block",
|
|
||||||
"index": 2,
|
|
||||||
"id": "msg_123",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"id": "rs_234",
|
|
||||||
"summary": [
|
|
||||||
{"index": 0, "type": "summary_text", "text": "more reasoning"},
|
|
||||||
{
|
|
||||||
"index": 1,
|
|
||||||
"type": "summary_text",
|
|
||||||
"text": "still more reasoning",
|
|
||||||
},
|
|
||||||
],
|
|
||||||
"type": "reasoning",
|
|
||||||
"index": 3,
|
|
||||||
},
|
|
||||||
{"type": "text", "text": "more", "index": 4, "id": "msg_234"},
|
|
||||||
{"type": "text", "text": "text", "index": 5, "id": "msg_234"},
|
|
||||||
],
|
|
||||||
),
|
|
||||||
(
|
|
||||||
"v1",
|
|
||||||
[
|
|
||||||
{
|
|
||||||
"type": "reasoning",
|
|
||||||
"reasoning": "reasoning block one",
|
|
||||||
"id": "rs_123",
|
|
||||||
"index": 0,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "reasoning",
|
|
||||||
"reasoning": "another reasoning block",
|
|
||||||
"id": "rs_123",
|
|
||||||
"index": 1,
|
|
||||||
},
|
|
||||||
{"type": "text", "text": "text block one", "index": 2, "id": "msg_123"},
|
|
||||||
{
|
|
||||||
"type": "text",
|
|
||||||
"text": "another text block",
|
|
||||||
"index": 3,
|
|
||||||
"id": "msg_123",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "reasoning",
|
|
||||||
"reasoning": "more reasoning",
|
|
||||||
"id": "rs_234",
|
|
||||||
"index": 4,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "reasoning",
|
|
||||||
"reasoning": "still more reasoning",
|
|
||||||
"id": "rs_234",
|
|
||||||
"index": 5,
|
|
||||||
},
|
|
||||||
{"type": "text", "text": "more", "index": 6, "id": "msg_234"},
|
|
||||||
{"type": "text", "text": "text", "index": 7, "id": "msg_234"},
|
|
||||||
],
|
|
||||||
),
|
|
||||||
],
|
|
||||||
)
|
|
||||||
def test_responses_stream(
|
|
||||||
output_version: Literal["v0", "responses/v1"], expected_content: list[dict]
|
|
||||||
) -> None:
|
|
||||||
llm = ChatOpenAI(
|
llm = ChatOpenAI(
|
||||||
model="o4-mini", use_responses_api=True, output_version=output_version
|
model="o4-mini", use_responses_api=True, output_version="responses/v1"
|
||||||
)
|
)
|
||||||
mock_client = MagicMock()
|
mock_client = MagicMock()
|
||||||
|
|
||||||
@ -728,8 +639,32 @@ def test_responses_stream(
|
|||||||
assert isinstance(chunk, AIMessageChunk)
|
assert isinstance(chunk, AIMessageChunk)
|
||||||
full = chunk if full is None else full + chunk
|
full = chunk if full is None else full + chunk
|
||||||
chunks.append(chunk)
|
chunks.append(chunk)
|
||||||
|
|
||||||
assert isinstance(full, AIMessageChunk)
|
assert isinstance(full, AIMessageChunk)
|
||||||
|
|
||||||
|
expected_content = [
|
||||||
|
{
|
||||||
|
"id": "rs_123",
|
||||||
|
"summary": [
|
||||||
|
{"index": 0, "type": "summary_text", "text": "reasoning block one"},
|
||||||
|
{"index": 1, "type": "summary_text", "text": "another reasoning block"},
|
||||||
|
],
|
||||||
|
"type": "reasoning",
|
||||||
|
"index": 0,
|
||||||
|
},
|
||||||
|
{"type": "text", "text": "text block one", "index": 1, "id": "msg_123"},
|
||||||
|
{"type": "text", "text": "another text block", "index": 2, "id": "msg_123"},
|
||||||
|
{
|
||||||
|
"id": "rs_234",
|
||||||
|
"summary": [
|
||||||
|
{"index": 0, "type": "summary_text", "text": "more reasoning"},
|
||||||
|
{"index": 1, "type": "summary_text", "text": "still more reasoning"},
|
||||||
|
],
|
||||||
|
"type": "reasoning",
|
||||||
|
"index": 3,
|
||||||
|
},
|
||||||
|
{"type": "text", "text": "more", "index": 4, "id": "msg_234"},
|
||||||
|
{"type": "text", "text": "text", "index": 5, "id": "msg_234"},
|
||||||
|
]
|
||||||
assert full.content == expected_content
|
assert full.content == expected_content
|
||||||
assert full.additional_kwargs == {}
|
assert full.additional_kwargs == {}
|
||||||
assert full.id == "resp_123"
|
assert full.id == "resp_123"
|
||||||
@ -749,3 +684,71 @@ def test_responses_stream(
|
|||||||
dumped = _strip_none(item.model_dump())
|
dumped = _strip_none(item.model_dump())
|
||||||
_ = dumped.pop("status", None)
|
_ = dumped.pop("status", None)
|
||||||
assert dumped == payload["input"][idx]
|
assert dumped == payload["input"][idx]
|
||||||
|
|
||||||
|
|
||||||
|
def test_responses_stream_v1() -> None:
|
||||||
|
llm = ChatOpenAIV1(model="o4-mini", use_responses_api=True)
|
||||||
|
mock_client = MagicMock()
|
||||||
|
|
||||||
|
def mock_create(*args: Any, **kwargs: Any) -> MockSyncContextManager:
|
||||||
|
return MockSyncContextManager(responses_stream)
|
||||||
|
|
||||||
|
mock_client.responses.create = mock_create
|
||||||
|
|
||||||
|
full: Optional[AIMessageChunkV1] = None
|
||||||
|
chunks = []
|
||||||
|
with patch.object(llm, "root_client", mock_client):
|
||||||
|
for chunk in llm.stream("test"):
|
||||||
|
assert isinstance(chunk, AIMessageChunkV1)
|
||||||
|
full = chunk if full is None else full + chunk
|
||||||
|
chunks.append(chunk)
|
||||||
|
assert isinstance(full, AIMessageChunkV1)
|
||||||
|
|
||||||
|
expected_content = [
|
||||||
|
{
|
||||||
|
"type": "reasoning",
|
||||||
|
"reasoning": "reasoning block one",
|
||||||
|
"id": "rs_123",
|
||||||
|
"index": 0,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "reasoning",
|
||||||
|
"reasoning": "another reasoning block",
|
||||||
|
"id": "rs_123",
|
||||||
|
"index": 1,
|
||||||
|
},
|
||||||
|
{"type": "text", "text": "text block one", "index": 2, "id": "msg_123"},
|
||||||
|
{"type": "text", "text": "another text block", "index": 3, "id": "msg_123"},
|
||||||
|
{
|
||||||
|
"type": "reasoning",
|
||||||
|
"reasoning": "more reasoning",
|
||||||
|
"id": "rs_234",
|
||||||
|
"index": 4,
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "reasoning",
|
||||||
|
"reasoning": "still more reasoning",
|
||||||
|
"id": "rs_234",
|
||||||
|
"index": 5,
|
||||||
|
},
|
||||||
|
{"type": "text", "text": "more", "index": 6, "id": "msg_234"},
|
||||||
|
{"type": "text", "text": "text", "index": 7, "id": "msg_234"},
|
||||||
|
]
|
||||||
|
assert full.content == expected_content
|
||||||
|
assert full.id == "resp_123"
|
||||||
|
|
||||||
|
# Test reconstruction
|
||||||
|
payload = llm._get_request_payload([full])
|
||||||
|
completed = [
|
||||||
|
item
|
||||||
|
for item in responses_stream
|
||||||
|
if item.type == "response.completed" # type: ignore[attr-defined]
|
||||||
|
]
|
||||||
|
assert len(completed) == 1
|
||||||
|
response = completed[0].response # type: ignore[attr-defined]
|
||||||
|
|
||||||
|
assert len(response.output) == len(payload["input"])
|
||||||
|
for idx, item in enumerate(response.output):
|
||||||
|
dumped = _strip_none(item.model_dump())
|
||||||
|
_ = dumped.pop("status", None)
|
||||||
|
assert dumped == payload["input"][idx]
|
||||||
|
Loading…
Reference in New Issue
Block a user