mirror of
https://github.com/hwchase17/langchain.git
synced 2025-05-02 05:45:47 +00:00
Signed-off-by: ChengZi <chen.zhang@zilliz.com> Co-authored-by: Eugene Yurtsev <eyurtsev@gmail.com> Co-authored-by: Bagatur <22008038+baskaryan@users.noreply.github.com> Co-authored-by: Dan O'Donovan <dan.odonovan@gmail.com> Co-authored-by: Tom Daniel Grande <tomdgrande@gmail.com> Co-authored-by: Grande <Tom.Daniel.Grande@statsbygg.no> Co-authored-by: Bagatur <baskaryan@gmail.com> Co-authored-by: ccurme <chester.curme@gmail.com> Co-authored-by: Harrison Chase <hw.chase.17@gmail.com> Co-authored-by: Tomaz Bratanic <bratanic.tomaz@gmail.com> Co-authored-by: ZhangShenao <15201440436@163.com> Co-authored-by: Friso H. Kingma <fhkingma@gmail.com> Co-authored-by: ChengZi <chen.zhang@zilliz.com> Co-authored-by: Nuno Campos <nuno@langchain.dev> Co-authored-by: Morgante Pell <morgantep@google.com>
60 lines
1.8 KiB
Python
60 lines
1.8 KiB
Python
"""Test LLM chain."""
|
|
|
|
from typing import Dict, List, Union
|
|
|
|
import pytest
|
|
from langchain.chains.llm import LLMChain
|
|
from langchain_core.output_parsers import BaseOutputParser
|
|
from langchain_core.prompts import PromptTemplate
|
|
|
|
from tests.unit_tests.llms.fake_llm import FakeLLM
|
|
|
|
|
|
class FakeOutputParser(BaseOutputParser):
|
|
"""Fake output parser class for testing."""
|
|
|
|
def parse(self, text: str) -> Union[str, List[str], Dict[str, str]]:
|
|
"""Parse by splitting."""
|
|
return text.split()
|
|
|
|
|
|
@pytest.fixture
|
|
def fake_llm_chain() -> LLMChain:
|
|
"""Fake LLM chain for testing purposes."""
|
|
prompt = PromptTemplate(input_variables=["bar"], template="This is a {bar}:")
|
|
return LLMChain(prompt=prompt, llm=FakeLLM(), output_key="text1")
|
|
|
|
|
|
def test_missing_inputs(fake_llm_chain: LLMChain) -> None:
|
|
"""Test error is raised if inputs are missing."""
|
|
with pytest.raises(ValueError):
|
|
fake_llm_chain({"foo": "bar"})
|
|
|
|
|
|
def test_valid_call(fake_llm_chain: LLMChain) -> None:
|
|
"""Test valid call of LLM chain."""
|
|
output = fake_llm_chain({"bar": "baz"})
|
|
assert output == {"bar": "baz", "text1": "foo"}
|
|
|
|
# Test with stop words.
|
|
output = fake_llm_chain({"bar": "baz", "stop": ["foo"]})
|
|
# Response should be `bar` now.
|
|
assert output == {"bar": "baz", "stop": ["foo"], "text1": "bar"}
|
|
|
|
|
|
def test_predict_method(fake_llm_chain: LLMChain) -> None:
|
|
"""Test predict method works."""
|
|
output = fake_llm_chain.predict(bar="baz")
|
|
assert output == "foo"
|
|
|
|
|
|
def test_predict_and_parse() -> None:
|
|
"""Test parsing ability."""
|
|
prompt = PromptTemplate(
|
|
input_variables=["foo"], template="{foo}", output_parser=FakeOutputParser()
|
|
)
|
|
llm = FakeLLM(queries={"foo": "foo bar"})
|
|
chain = LLMChain(prompt=prompt, llm=llm)
|
|
output = chain.predict_and_parse(foo="foo")
|
|
assert output == ["foo", "bar"]
|