diff --git a/libs/partners/openai/langchain_openai/chat_models/base.py b/libs/partners/openai/langchain_openai/chat_models/base.py index a446916d4e0..c3c94c46676 100644 --- a/libs/partners/openai/langchain_openai/chat_models/base.py +++ b/libs/partners/openai/langchain_openai/chat_models/base.py @@ -2330,6 +2330,7 @@ class BaseChatOpenAI(BaseChatModel): # Batch API processing (50% cost savings, polling required) results = llm.batch(inputs, use_batch_api=True) """ + def _convert_input_to_messages( self, input_item: LanguageModelInput ) -> list[BaseMessage]: diff --git a/libs/partners/openai/tests/unit_tests/chat_models/test_batch.py b/libs/partners/openai/tests/unit_tests/chat_models/test_batch.py index 47f1b03a755..9c2dedf20c6 100644 --- a/libs/partners/openai/tests/unit_tests/chat_models/test_batch.py +++ b/libs/partners/openai/tests/unit_tests/chat_models/test_batch.py @@ -48,9 +48,7 @@ class TestOpenAIBatchClient: ] batch_id = self.batch_client.create_batch( - requests=batch_requests, - description="Test batch", - metadata={"test": "true"}, + requests=batch_requests, description="Test batch", metadata={"test": "true"} ) assert batch_id == "batch_123"