mirror of
https://github.com/hwchase17/langchain.git
synced 2025-08-10 21:35:08 +00:00
BUGFIX: llm backwards compat imports (#13698)
This commit is contained in:
parent
ace9e64d62
commit
a21e84faf7
19
libs/core/tests/unit_tests/language_models/llms/test_base.py
Normal file
19
libs/core/tests/unit_tests/language_models/llms/test_base.py
Normal file
@ -0,0 +1,19 @@
|
||||
from tests.unit_tests.fake.llm import FakeListLLM
|
||||
|
||||
|
||||
def test_batch() -> None:
|
||||
llm = FakeListLLM(responses=["foo"] * 3)
|
||||
output = llm.batch(["foo", "bar", "foo"])
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
output = llm.batch(["foo", "bar", "foo"], config={"max_concurrency": 2})
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
|
||||
async def test_abatch() -> None:
|
||||
llm = FakeListLLM(responses=["foo"] * 3)
|
||||
output = await llm.abatch(["foo", "bar", "foo"])
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
output = await llm.abatch(["foo", "bar", "foo"], config={"max_concurrency": 2})
|
||||
assert output == ["foo"] * 3
|
@ -3,6 +3,7 @@ from langchain_core.language_models.chat_models import (
|
||||
SimpleChatModel,
|
||||
_agenerate_from_stream,
|
||||
_generate_from_stream,
|
||||
_get_verbosity,
|
||||
)
|
||||
|
||||
__all__ = [
|
||||
@ -10,4 +11,5 @@ __all__ = [
|
||||
"SimpleChatModel",
|
||||
"_generate_from_stream",
|
||||
"_agenerate_from_stream",
|
||||
"_get_verbosity",
|
||||
]
|
||||
|
@ -1,6 +1,9 @@
|
||||
# Backwards compatibility.
|
||||
from langchain_core.language_models import BaseLanguageModel
|
||||
from langchain_core.language_models.llms import (
|
||||
LLM,
|
||||
BaseLLM,
|
||||
_get_verbosity,
|
||||
create_base_retry_decorator,
|
||||
get_prompts,
|
||||
update_cache,
|
||||
@ -10,6 +13,8 @@ __all__ = [
|
||||
"create_base_retry_decorator",
|
||||
"get_prompts",
|
||||
"update_cache",
|
||||
"BaseLanguageModel",
|
||||
"_get_verbosity",
|
||||
"BaseLLM",
|
||||
"LLM",
|
||||
]
|
||||
|
13
libs/langchain/tests/unit_tests/chat_models/test_base.py
Normal file
13
libs/langchain/tests/unit_tests/chat_models/test_base.py
Normal file
@ -0,0 +1,13 @@
|
||||
from langchain.chat_models.base import __all__
|
||||
|
||||
EXPECTED_ALL = [
|
||||
"BaseChatModel",
|
||||
"SimpleChatModel",
|
||||
"_agenerate_from_stream",
|
||||
"_generate_from_stream",
|
||||
"_get_verbosity",
|
||||
]
|
||||
|
||||
|
||||
def test_all_imports() -> None:
|
||||
assert set(__all__) == set(EXPECTED_ALL)
|
@ -10,8 +10,23 @@ from langchain_core.outputs import Generation, LLMResult
|
||||
|
||||
from langchain.cache import InMemoryCache, SQLAlchemyCache
|
||||
from langchain.globals import get_llm_cache, set_llm_cache
|
||||
from langchain.llms.base import __all__
|
||||
from tests.unit_tests.llms.fake_llm import FakeLLM
|
||||
|
||||
EXPECTED_ALL = [
|
||||
"BaseLLM",
|
||||
"LLM",
|
||||
"_get_verbosity",
|
||||
"create_base_retry_decorator",
|
||||
"get_prompts",
|
||||
"update_cache",
|
||||
"BaseLanguageModel",
|
||||
]
|
||||
|
||||
|
||||
def test_all_imports() -> None:
|
||||
assert set(__all__) == set(EXPECTED_ALL)
|
||||
|
||||
|
||||
def test_caching() -> None:
|
||||
"""Test caching behavior."""
|
||||
@ -74,21 +89,3 @@ def test_custom_caching() -> None:
|
||||
llm_output=None,
|
||||
)
|
||||
assert output == expected_output
|
||||
|
||||
|
||||
def test_batch() -> None:
|
||||
llm = FakeLLM()
|
||||
output = llm.batch(["foo", "bar", "foo"])
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
output = llm.batch(["foo", "bar", "foo"], config={"max_concurrency": 2})
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
|
||||
async def test_abatch() -> None:
|
||||
llm = FakeLLM()
|
||||
output = await llm.abatch(["foo", "bar", "foo"])
|
||||
assert output == ["foo"] * 3
|
||||
|
||||
output = await llm.abatch(["foo", "bar", "foo"], config={"max_concurrency": 2})
|
||||
assert output == ["foo"] * 3
|
||||
|
Loading…
Reference in New Issue
Block a user