diff --git a/libs/partners/anthropic/tests/integration_tests/test_chat_models.py b/libs/partners/anthropic/tests/integration_tests/test_chat_models.py index 6ec4b26e385..2a057a587ef 100644 --- a/libs/partners/anthropic/tests/integration_tests/test_chat_models.py +++ b/libs/partners/anthropic/tests/integration_tests/test_chat_models.py @@ -5,7 +5,7 @@ from langchain_anthropic.chat_models import ChatAnthropicMessages def test_stream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from Anthropic.""" llm = ChatAnthropicMessages(model_name="claude-instant-1.2") for token in llm.stream("I'm Pickle Rick"): @@ -13,7 +13,7 @@ def test_stream() -> None: async def test_astream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from Anthropic.""" llm = ChatAnthropicMessages(model_name="claude-instant-1.2") async for token in llm.astream("I'm Pickle Rick"): diff --git a/libs/partners/google-genai/tests/integration_tests/test_chat_models.py b/libs/partners/google-genai/tests/integration_tests/test_chat_models.py index c1afcdd54bd..4e9c7c764ac 100644 --- a/libs/partners/google-genai/tests/integration_tests/test_chat_models.py +++ b/libs/partners/google-genai/tests/integration_tests/test_chat_models.py @@ -13,7 +13,7 @@ _B64_string = """iVBORw0KGgoAAAANSUhEUgAAABQAAAAUCAIAAAAC64paAAABhGlDQ1BJQ0MgUHJ def test_chat_google_genai_stream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from Gemini.""" llm = ChatGoogleGenerativeAI(model=_MODEL) for token in llm.stream("This is a test. Say 'foo'"): @@ -21,7 +21,7 @@ def test_chat_google_genai_stream() -> None: async def test_chat_google_genai_astream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from Gemini.""" llm = ChatGoogleGenerativeAI(model=_MODEL) async for token in llm.astream("This is a test. Say 'foo'"): diff --git a/libs/partners/nvidia-trt/tests/integration_tests/test_llms.py b/libs/partners/nvidia-trt/tests/integration_tests/test_llms.py index 72d4a8b9fd9..200577d7eea 100644 --- a/libs/partners/nvidia-trt/tests/integration_tests/test_llms.py +++ b/libs/partners/nvidia-trt/tests/integration_tests/test_llms.py @@ -8,7 +8,7 @@ _MODEL_NAME = "ensemble" @pytest.mark.skip(reason="Need a working Triton server") def test_stream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from NVIDIA TRT.""" llm = TritonTensorRTLLM(model_name=_MODEL_NAME) for token in llm.stream("I'm Pickle Rick"): @@ -17,7 +17,7 @@ def test_stream() -> None: @pytest.mark.skip(reason="Need a working Triton server") async def test_astream() -> None: - """Test streaming tokens from OpenAI.""" + """Test streaming tokens from NVIDIA TRT.""" llm = TritonTensorRTLLM(model_name=_MODEL_NAME) async for token in llm.astream("I'm Pickle Rick"):