diff --git a/libs/text-splitters/tests/integration_tests/test_text_splitter.py b/libs/text-splitters/tests/integration_tests/test_text_splitter.py index b44bffbcb3a..78e613a2fc5 100644 --- a/libs/text-splitters/tests/integration_tests/test_text_splitter.py +++ b/libs/text-splitters/tests/integration_tests/test_text_splitter.py @@ -58,7 +58,7 @@ def test_token_text_splitter_from_tiktoken() -> None: assert expected_tokenizer == actual_tokenizer -def test_sentence_transformers_count_tokens(sentence_transformers: Any) -> None: +def test_sentence_transformers_count_tokens() -> None: splitter = SentenceTransformersTokenTextSplitter( model_name="sentence-transformers/paraphrase-albert-small-v2" ) @@ -114,3 +114,4 @@ def test_sentence_transformers_multiple_tokens(sentence_transformers: Any) -> No ) assert expected == actual +