diff --git a/libs/partners/huggingface/tests/integration_tests/test_chat_models.py b/libs/partners/huggingface/tests/integration_tests/test_chat_models.py index 4a64f7f0b5a..2c546056a2a 100644 --- a/libs/partners/huggingface/tests/integration_tests/test_chat_models.py +++ b/libs/partners/huggingface/tests/integration_tests/test_chat_models.py @@ -8,7 +8,7 @@ def test_stream_usage() -> None: llm = HuggingFaceEndpoint( # type: ignore[call-arg] # (model is inferred in class) repo_id="google/gemma-3-27b-it", task="conversational", - provider="nebius", + provider="scaleway", ) model = ChatHuggingFace(llm=llm, stream_usage=True) diff --git a/libs/partners/huggingface/tests/integration_tests/test_standard.py b/libs/partners/huggingface/tests/integration_tests/test_standard.py index 892c0bc0d39..cba4d775392 100644 --- a/libs/partners/huggingface/tests/integration_tests/test_standard.py +++ b/libs/partners/huggingface/tests/integration_tests/test_standard.py @@ -18,9 +18,9 @@ class TestHuggingFaceEndpoint(ChatModelIntegrationTests): @property def chat_model_params(self) -> dict: llm = HuggingFaceEndpoint( # type: ignore[call-arg] - repo_id="meta-llama/Llama-4-Maverick-17B-128E-Instruct", + repo_id="meta-llama/Llama-3.3-70B-Instruct", task="conversational", - provider="fireworks-ai", + provider="sambanova", temperature=0, ) return {"llm": llm}