diff --git a/libs/langchain/tests/unit_tests/llms/fake_chat_model.py b/libs/langchain/tests/unit_tests/llms/fake_chat_model.py index 3a746720d82..ffc80bf683d 100644 --- a/libs/langchain/tests/unit_tests/llms/fake_chat_model.py +++ b/libs/langchain/tests/unit_tests/llms/fake_chat_model.py @@ -120,9 +120,9 @@ class GenericFakeChatModel(BaseChatModel): for token in content_chunks: chunk = ChatGenerationChunk(message=AIMessageChunk(content=token)) - yield chunk if run_manager: run_manager.on_llm_new_token(token, chunk=chunk) + yield chunk if message.additional_kwargs: for key, value in message.additional_kwargs.items(): @@ -142,12 +142,12 @@ class GenericFakeChatModel(BaseChatModel): }, ) ) - yield chunk if run_manager: run_manager.on_llm_new_token( "", chunk=chunk, # No token for function call ) + yield chunk else: chunk = ChatGenerationChunk( message=AIMessageChunk( @@ -155,24 +155,24 @@ class GenericFakeChatModel(BaseChatModel): additional_kwargs={"function_call": {fkey: fvalue}}, ) ) - yield chunk if run_manager: run_manager.on_llm_new_token( "", chunk=chunk, # No token for function call ) + yield chunk else: chunk = ChatGenerationChunk( message=AIMessageChunk( content="", additional_kwargs={key: value} ) ) - yield chunk if run_manager: run_manager.on_llm_new_token( "", chunk=chunk, # No token for function call ) + yield chunk async def _astream( self,