mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-16 23:13:31 +00:00
openai[patch]: propagate service_tier to response metadata (#31089)
This commit is contained in:
@@ -350,6 +350,7 @@ def test_response_metadata() -> None:
|
||||
"logprobs",
|
||||
"system_fingerprint",
|
||||
"finish_reason",
|
||||
"service_tier",
|
||||
)
|
||||
)
|
||||
assert "content" in result.response_metadata["logprobs"]
|
||||
@@ -367,6 +368,7 @@ async def test_async_response_metadata() -> None:
|
||||
"logprobs",
|
||||
"system_fingerprint",
|
||||
"finish_reason",
|
||||
"service_tier",
|
||||
)
|
||||
)
|
||||
assert "content" in result.response_metadata["logprobs"]
|
||||
@@ -380,7 +382,7 @@ def test_response_metadata_streaming() -> None:
|
||||
full = chunk if full is None else full + chunk
|
||||
assert all(
|
||||
k in cast(BaseMessageChunk, full).response_metadata
|
||||
for k in ("logprobs", "finish_reason")
|
||||
for k in ("logprobs", "finish_reason", "service_tier")
|
||||
)
|
||||
assert "content" in cast(BaseMessageChunk, full).response_metadata["logprobs"]
|
||||
|
||||
@@ -393,7 +395,7 @@ async def test_async_response_metadata_streaming() -> None:
|
||||
full = chunk if full is None else full + chunk
|
||||
assert all(
|
||||
k in cast(BaseMessageChunk, full).response_metadata
|
||||
for k in ("logprobs", "finish_reason")
|
||||
for k in ("logprobs", "finish_reason", "service_tier")
|
||||
)
|
||||
assert "content" in cast(BaseMessageChunk, full).response_metadata["logprobs"]
|
||||
|
||||
|
@@ -47,6 +47,7 @@ def _check_response(response: Optional[BaseMessage]) -> None:
|
||||
assert response.usage_metadata["output_tokens"] > 0
|
||||
assert response.usage_metadata["total_tokens"] > 0
|
||||
assert response.response_metadata["model_name"]
|
||||
assert response.response_metadata["service_tier"]
|
||||
for tool_output in response.additional_kwargs["tool_outputs"]:
|
||||
assert tool_output["id"]
|
||||
assert tool_output["status"]
|
||||
|
Reference in New Issue
Block a user