support kwargs (#5990)

This commit is contained in:
Harrison Chase
2023-06-11 10:09:22 -07:00
committed by GitHub
parent b934677a81
commit 704d56e241
58 changed files with 289 additions and 88 deletions

View File

@@ -20,6 +20,7 @@ class FakeListLLM(LLM):
prompt: str,
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> str:
"""Increment counter, and then return response in that index."""
self.i += 1

View File

@@ -38,6 +38,7 @@ class FakeListLLM(LLM):
prompt: str,
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> str:
"""Increment counter, and then return response in that index."""
self.i += 1

View File

@@ -1,5 +1,5 @@
"""Test HyDE."""
from typing import List, Optional
from typing import Any, List, Optional
import numpy as np
@@ -36,6 +36,7 @@ class FakeLLM(BaseLLM):
prompts: List[str],
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> LLMResult:
return LLMResult(generations=[[Generation(text="foo") for _ in range(self.n)]])
@@ -44,6 +45,7 @@ class FakeLLM(BaseLLM):
prompts: List[str],
stop: Optional[List[str]] = None,
run_manager: Optional[AsyncCallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> LLMResult:
return LLMResult(generations=[[Generation(text="foo") for _ in range(self.n)]])

View File

@@ -15,6 +15,7 @@ class FakeLLM(LLM):
prompt: str,
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> str:
"""Return `foo` if longer than 10000 words, else `bar`."""
if len(prompt) > 10000:

View File

@@ -17,6 +17,7 @@ class FakeChatModel(SimpleChatModel):
messages: List[BaseMessage],
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> str:
return "fake response"
@@ -25,6 +26,7 @@ class FakeChatModel(SimpleChatModel):
messages: List[BaseMessage],
stop: Optional[List[str]] = None,
run_manager: Optional[AsyncCallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> ChatResult:
output_str = "fake response"
message = AIMessage(content=output_str)

View File

@@ -34,6 +34,7 @@ class FakeLLM(LLM):
prompt: str,
stop: Optional[List[str]] = None,
run_manager: Optional[CallbackManagerForLLMRun] = None,
**kwargs: Any,
) -> str:
if self.sequential_responses:
return self._get_next_response_in_sequence