diff --git a/libs/community/langchain_community/llms/openai.py b/libs/community/langchain_community/llms/openai.py index ce35db32e34..56ca15dda74 100644 --- a/libs/community/langchain_community/llms/openai.py +++ b/libs/community/langchain_community/llms/openai.py @@ -391,7 +391,6 @@ class BaseOpenAI(BaseLLM): if not isinstance(stream_resp, dict): stream_resp = stream_resp.dict() chunk = _stream_response_to_generation_chunk(stream_resp) - yield chunk if run_manager: await run_manager.on_llm_new_token( chunk.text, @@ -401,6 +400,7 @@ class BaseOpenAI(BaseLLM): if chunk.generation_info else None, ) + yield chunk def _generate( self, @@ -1113,9 +1113,9 @@ class OpenAIChat(BaseLLM): stream_resp = stream_resp.dict() token = stream_resp["choices"][0]["delta"].get("content", "") chunk = GenerationChunk(text=token) - yield chunk if run_manager: run_manager.on_llm_new_token(token, chunk=chunk) + yield chunk async def _astream( self, @@ -1133,9 +1133,9 @@ class OpenAIChat(BaseLLM): stream_resp = stream_resp.dict() token = stream_resp["choices"][0]["delta"].get("content", "") chunk = GenerationChunk(text=token) - yield chunk if run_manager: await run_manager.on_llm_new_token(token, chunk=chunk) + yield chunk def _generate( self,