mirror of
https://github.com/hwchase17/langchain.git
synced 2025-08-17 16:39:52 +00:00
explicitly enable test
This commit is contained in:
parent
7285b173b8
commit
272efba5b9
@ -41,6 +41,10 @@ class TestAnthropicStandard(ChatModelIntegrationTests):
|
|||||||
def supports_anthropic_inputs(self) -> bool:
|
def supports_anthropic_inputs(self) -> bool:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enable_vcr_tests(self) -> bool:
|
||||||
|
return True
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def supported_usage_metadata_details(
|
def supported_usage_metadata_details(
|
||||||
self,
|
self,
|
||||||
|
@ -38,6 +38,10 @@ class TestAzureOpenAIStandard(ChatModelIntegrationTests):
|
|||||||
def supports_json_mode(self) -> bool:
|
def supports_json_mode(self) -> bool:
|
||||||
return True
|
return True
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enable_vcr_tests(self) -> bool:
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
class TestAzureOpenAIStandardLegacy(ChatModelIntegrationTests):
|
class TestAzureOpenAIStandardLegacy(ChatModelIntegrationTests):
|
||||||
"""Test a legacy model."""
|
"""Test a legacy model."""
|
||||||
@ -58,3 +62,7 @@ class TestAzureOpenAIStandardLegacy(ChatModelIntegrationTests):
|
|||||||
@property
|
@property
|
||||||
def structured_output_kwargs(self) -> dict:
|
def structured_output_kwargs(self) -> dict:
|
||||||
return {"method": "function_calling"}
|
return {"method": "function_calling"}
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enable_vcr_tests(self) -> bool:
|
||||||
|
return True
|
||||||
|
@ -57,6 +57,10 @@ class TestOpenAIStandard(ChatModelIntegrationTests):
|
|||||||
]:
|
]:
|
||||||
return {"invoke": ["reasoning_output", "cache_read_input"], "stream": []}
|
return {"invoke": ["reasoning_output", "cache_read_input"], "stream": []}
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enable_vcr_tests(self) -> bool:
|
||||||
|
return True
|
||||||
|
|
||||||
def invoke_with_cache_read_input(self, *, stream: bool = False) -> AIMessage:
|
def invoke_with_cache_read_input(self, *, stream: bool = False) -> AIMessage:
|
||||||
with open(REPO_ROOT_DIR / "README.md") as f:
|
with open(REPO_ROOT_DIR / "README.md") as f:
|
||||||
readme = f.read()
|
readme = f.read()
|
||||||
|
@ -2681,6 +2681,7 @@ class ChatModelIntegrationTests(ChatModelTests):
|
|||||||
)
|
)
|
||||||
assert isinstance(response, AIMessage)
|
assert isinstance(response, AIMessage)
|
||||||
|
|
||||||
|
@pytest.mark.benchmark
|
||||||
def test_stream_time(
|
def test_stream_time(
|
||||||
self, model: BaseChatModel, benchmark: BenchmarkFixture, vcr: vcr.VCR
|
self, model: BaseChatModel, benchmark: BenchmarkFixture, vcr: vcr.VCR
|
||||||
) -> None:
|
) -> None:
|
||||||
|
@ -201,6 +201,10 @@ class ChatModelTests(BaseStandardTests):
|
|||||||
content."""
|
content."""
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
@property
|
||||||
|
def enable_vcr_tests(self) -> bool:
|
||||||
|
return False
|
||||||
|
|
||||||
@property
|
@property
|
||||||
def supported_usage_metadata_details(
|
def supported_usage_metadata_details(
|
||||||
self,
|
self,
|
||||||
|
Loading…
Reference in New Issue
Block a user