From e0162baa3bd7da13f6c61637ddc3bb677b26f1f1 Mon Sep 17 00:00:00 2001 From: Bagatur <22008038+baskaryan@users.noreply.github.com> Date: Tue, 15 Aug 2023 09:40:33 -0700 Subject: [PATCH] add oai sched tests (#9257) --- .../integration_tests/chat_models/test_openai.py | 13 +++++++++++++ .../integration_tests/embeddings/test_openai.py | 8 ++++++++ 2 files changed, 21 insertions(+) diff --git a/libs/langchain/tests/integration_tests/chat_models/test_openai.py b/libs/langchain/tests/integration_tests/chat_models/test_openai.py index ff7332c88cb..19adbf1cd41 100644 --- a/libs/langchain/tests/integration_tests/chat_models/test_openai.py +++ b/libs/langchain/tests/integration_tests/chat_models/test_openai.py @@ -14,6 +14,7 @@ from langchain.schema.messages import BaseMessage, HumanMessage, SystemMessage from tests.unit_tests.callbacks.fake_callback_handler import FakeCallbackHandler +@pytest.mark.scheduled def test_chat_openai() -> None: """Test ChatOpenAI wrapper.""" chat = ChatOpenAI(max_tokens=10) @@ -41,6 +42,7 @@ def test_chat_openai_system_message() -> None: assert isinstance(response.content, str) +@pytest.mark.scheduled def test_chat_openai_generate() -> None: """Test ChatOpenAI wrapper with generate.""" chat = ChatOpenAI(max_tokens=10, n=2) @@ -56,6 +58,7 @@ def test_chat_openai_generate() -> None: assert generation.text == generation.message.content +@pytest.mark.scheduled def test_chat_openai_multiple_completions() -> None: """Test ChatOpenAI wrapper with multiple completions.""" chat = ChatOpenAI(max_tokens=10, n=5) @@ -68,6 +71,7 @@ def test_chat_openai_multiple_completions() -> None: assert isinstance(generation.message.content, str) +@pytest.mark.scheduled def test_chat_openai_streaming() -> None: """Test that streaming correctly invokes on_llm_new_token callback.""" callback_handler = FakeCallbackHandler() @@ -114,6 +118,7 @@ def test_chat_openai_invalid_streaming_params() -> None: ) +@pytest.mark.scheduled @pytest.mark.asyncio async def test_async_chat_openai() -> None: """Test async generation.""" @@ -130,6 +135,7 @@ async def test_async_chat_openai() -> None: assert generation.text == generation.message.content +@pytest.mark.scheduled @pytest.mark.asyncio async def test_async_chat_openai_streaming() -> None: """Test that streaming correctly invokes on_llm_new_token callback.""" @@ -179,6 +185,7 @@ def test_chat_openai_extra_kwargs() -> None: ChatOpenAI(model_kwargs={"model": "text-davinci-003"}) +@pytest.mark.scheduled def test_openai_streaming() -> None: """Test streaming tokens from OpenAI.""" llm = ChatOpenAI(max_tokens=10) @@ -187,6 +194,7 @@ def test_openai_streaming() -> None: assert isinstance(token.content, str) +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_astream() -> None: """Test streaming tokens from OpenAI.""" @@ -196,6 +204,7 @@ async def test_openai_astream() -> None: assert isinstance(token.content, str) +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_abatch() -> None: """Test streaming tokens from ChatOpenAI.""" @@ -206,6 +215,7 @@ async def test_openai_abatch() -> None: assert isinstance(token.content, str) +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_abatch_tags() -> None: """Test batch tokens from ChatOpenAI.""" @@ -218,6 +228,7 @@ async def test_openai_abatch_tags() -> None: assert isinstance(token.content, str) +@pytest.mark.scheduled def test_openai_batch() -> None: """Test batch tokens from ChatOpenAI.""" llm = ChatOpenAI(max_tokens=10) @@ -227,6 +238,7 @@ def test_openai_batch() -> None: assert isinstance(token.content, str) +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_ainvoke() -> None: """Test invoke tokens from ChatOpenAI.""" @@ -236,6 +248,7 @@ async def test_openai_ainvoke() -> None: assert isinstance(result.content, str) +@pytest.mark.scheduled def test_openai_invoke() -> None: """Test invoke tokens from ChatOpenAI.""" llm = ChatOpenAI(max_tokens=10) diff --git a/libs/langchain/tests/integration_tests/embeddings/test_openai.py b/libs/langchain/tests/integration_tests/embeddings/test_openai.py index b1f2ae8785e..1288bb463e5 100644 --- a/libs/langchain/tests/integration_tests/embeddings/test_openai.py +++ b/libs/langchain/tests/integration_tests/embeddings/test_openai.py @@ -8,6 +8,7 @@ import pytest from langchain.embeddings.openai import OpenAIEmbeddings +@pytest.mark.scheduled def test_openai_embedding_documents() -> None: """Test openai embeddings.""" documents = ["foo bar"] @@ -17,6 +18,7 @@ def test_openai_embedding_documents() -> None: assert len(output[0]) == 1536 +@pytest.mark.scheduled def test_openai_embedding_documents_multiple() -> None: """Test openai embeddings.""" documents = ["foo bar", "bar foo", "foo"] @@ -29,6 +31,7 @@ def test_openai_embedding_documents_multiple() -> None: assert len(output[2]) == 1536 +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_embedding_documents_async_multiple() -> None: """Test openai embeddings.""" @@ -42,6 +45,7 @@ async def test_openai_embedding_documents_async_multiple() -> None: assert len(output[2]) == 1536 +@pytest.mark.scheduled def test_openai_embedding_query() -> None: """Test openai embeddings.""" document = "foo bar" @@ -50,6 +54,7 @@ def test_openai_embedding_query() -> None: assert len(output) == 1536 +@pytest.mark.scheduled @pytest.mark.asyncio async def test_openai_embedding_async_query() -> None: """Test openai embeddings.""" @@ -59,6 +64,7 @@ async def test_openai_embedding_async_query() -> None: assert len(output) == 1536 +@pytest.mark.scheduled def test_openai_embedding_with_empty_string() -> None: """Test openai embeddings with empty string.""" document = ["", "abc"] @@ -73,11 +79,13 @@ def test_openai_embedding_with_empty_string() -> None: assert len(output[1]) == 1536 +@pytest.mark.scheduled def test_embed_documents_normalized() -> None: output = OpenAIEmbeddings().embed_documents(["foo walked to the market"]) assert np.isclose(np.linalg.norm(output[0]), 1.0) +@pytest.mark.scheduled def test_embed_query_normalized() -> None: output = OpenAIEmbeddings().embed_query("foo walked to the market") assert np.isclose(np.linalg.norm(output), 1.0)