mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-16 23:13:31 +00:00
infra: add print rule to ruff (#16221)
Added noqa for existing prints. Can slowly remove / will prevent more being intro'd
This commit is contained in:
@@ -29,7 +29,7 @@ def test_language_loader_for_python() -> None:
|
||||
assert (
|
||||
docs[0].page_content
|
||||
== """def main():
|
||||
print("Hello World!")
|
||||
print("Hello World!") # noqa: T201
|
||||
|
||||
return 0"""
|
||||
)
|
||||
@@ -67,7 +67,7 @@ def esprima_installed() -> bool:
|
||||
|
||||
return True
|
||||
except Exception as e:
|
||||
print(f"esprima not installed, skipping test {e}")
|
||||
print(f"esprima not installed, skipping test {e}") # noqa: T201
|
||||
return False
|
||||
|
||||
|
||||
|
@@ -19,8 +19,8 @@ def test_load_success() -> None:
|
||||
|
||||
docs = loader.load()
|
||||
assert len(docs) == 1
|
||||
print(docs[0].metadata)
|
||||
print(docs[0].page_content)
|
||||
print(docs[0].metadata) # noqa: T201
|
||||
print(docs[0].page_content) # noqa: T201
|
||||
assert_docs(docs)
|
||||
|
||||
|
||||
@@ -53,7 +53,7 @@ def test_load_returns_full_set_of_metadata() -> None:
|
||||
assert set(doc.metadata).issuperset(
|
||||
{"Published", "Title", "Authors", "Summary"}
|
||||
)
|
||||
print(doc.metadata)
|
||||
print(doc.metadata) # noqa: T201
|
||||
assert len(set(doc.metadata)) > 4
|
||||
|
||||
|
||||
|
@@ -21,7 +21,7 @@ def test_get_nfts_valid_contract() -> None:
|
||||
)
|
||||
result = BlockchainDocumentLoader(contract_address).load()
|
||||
|
||||
print("Tokens returned for valid contract: ", len(result))
|
||||
print("Tokens returned for valid contract: ", len(result)) # noqa: T201
|
||||
|
||||
assert len(result) == max_alchemy_tokens, (
|
||||
f"Wrong number of NFTs returned. "
|
||||
@@ -43,7 +43,7 @@ def test_get_nfts_with_pagination() -> None:
|
||||
startToken=startToken,
|
||||
).load()
|
||||
|
||||
print("Tokens returned for contract with offset: ", len(result))
|
||||
print("Tokens returned for contract with offset: ", len(result)) # noqa: T201
|
||||
|
||||
assert len(result) > 0, "No NFTs returned"
|
||||
|
||||
@@ -57,7 +57,7 @@ def test_get_nfts_polygon() -> None:
|
||||
contract_address, BlockchainType.POLYGON_MAINNET
|
||||
).load()
|
||||
|
||||
print("Tokens returned for contract on Polygon: ", len(result))
|
||||
print("Tokens returned for contract on Polygon: ", len(result)) # noqa: T201
|
||||
|
||||
assert len(result) > 0, "No NFTs returned"
|
||||
|
||||
@@ -94,7 +94,7 @@ def test_get_all() -> None:
|
||||
|
||||
end_time = time.time()
|
||||
|
||||
print(
|
||||
print( # noqa: T201
|
||||
f"Tokens returned for {contract_address} "
|
||||
f"contract: {len(result)} in {end_time - start_time} seconds"
|
||||
)
|
||||
@@ -121,4 +121,4 @@ def test_get_all_10sec_timeout() -> None:
|
||||
|
||||
end_time = time.time()
|
||||
|
||||
print("Execution took ", end_time - start_time, " seconds")
|
||||
print("Execution took ", end_time - start_time, " seconds") # noqa: T201
|
||||
|
@@ -31,11 +31,11 @@ class TestCouchbaseLoader(unittest.TestCase):
|
||||
metadata_fields=self.valid_metadata_fields,
|
||||
)
|
||||
docs = loader.load()
|
||||
print(docs)
|
||||
print(docs) # noqa: T201
|
||||
|
||||
assert len(docs) > 0 # assuming the query returns at least one document
|
||||
for doc in docs:
|
||||
print(doc)
|
||||
print(doc) # noqa: T201
|
||||
assert (
|
||||
doc.page_content != ""
|
||||
) # assuming that every document has page_content
|
||||
|
@@ -28,7 +28,7 @@ class TestGitbookLoader:
|
||||
loader = GitbookLoader(
|
||||
web_page, load_all_paths=load_all_paths, base_url=base_url
|
||||
)
|
||||
print(loader.__dict__)
|
||||
print(loader.__dict__) # noqa: T201
|
||||
assert (
|
||||
loader.base_url == (base_url or web_page)[:-1]
|
||||
if (base_url or web_page).endswith("/")
|
||||
@@ -52,5 +52,5 @@ class TestGitbookLoader:
|
||||
def test_load_multiple_pages(self, web_page: str) -> None:
|
||||
loader = GitbookLoader(web_page, load_all_paths=True)
|
||||
result = loader.load()
|
||||
print(len(result))
|
||||
print(len(result)) # noqa: T201
|
||||
assert len(result) > 10
|
||||
|
@@ -145,14 +145,14 @@ def test_mathpix_loader() -> None:
|
||||
docs = loader.load()
|
||||
|
||||
assert len(docs) == 1
|
||||
print(docs[0].page_content)
|
||||
print(docs[0].page_content) # noqa: T201
|
||||
|
||||
file_path = Path(__file__).parent.parent / "examples/layout-parser-paper.pdf"
|
||||
loader = MathpixPDFLoader(str(file_path))
|
||||
|
||||
docs = loader.load()
|
||||
assert len(docs) == 1
|
||||
print(docs[0].page_content)
|
||||
print(docs[0].page_content) # noqa: T201
|
||||
|
||||
|
||||
@pytest.mark.parametrize(
|
||||
@@ -230,7 +230,7 @@ def test_amazontextract_loader(
|
||||
else:
|
||||
loader = AmazonTextractPDFLoader(file_path, textract_features=features)
|
||||
docs = loader.load()
|
||||
print(docs)
|
||||
print(docs) # noqa: T201
|
||||
|
||||
assert len(docs) == docs_length
|
||||
|
||||
|
@@ -1,4 +1,5 @@
|
||||
"""Integration test for PubMed API Wrapper."""
|
||||
|
||||
from typing import List
|
||||
|
||||
import pytest
|
||||
@@ -13,7 +14,7 @@ def test_load_success() -> None:
|
||||
"""Test that returns the correct answer"""
|
||||
api_client = PubMedLoader(query="chatgpt")
|
||||
docs = api_client.load()
|
||||
print(docs)
|
||||
print(docs) # noqa: T201
|
||||
assert len(docs) == api_client.load_max_docs == 3
|
||||
assert_docs(docs)
|
||||
|
||||
@@ -22,7 +23,7 @@ def test_load_success_load_max_docs() -> None:
|
||||
"""Test that returns the correct answer"""
|
||||
api_client = PubMedLoader(query="chatgpt", load_max_docs=2)
|
||||
docs = api_client.load()
|
||||
print(docs)
|
||||
print(docs) # noqa: T201
|
||||
assert len(docs) == api_client.load_max_docs == 2
|
||||
assert_docs(docs)
|
||||
|
||||
@@ -38,7 +39,7 @@ def test_load_no_content() -> None:
|
||||
"""Returns a Document without content."""
|
||||
api_client = PubMedLoader(query="37548971")
|
||||
docs = api_client.load()
|
||||
print(docs)
|
||||
print(docs) # noqa: T201
|
||||
assert len(docs) > 0
|
||||
assert docs[0].page_content == ""
|
||||
|
||||
|
@@ -33,7 +33,7 @@ def test_telegram_channel_loader_parsing() -> None:
|
||||
docs = loader.load()
|
||||
|
||||
assert len(docs) == 1
|
||||
print(docs[0].page_content)
|
||||
print(docs[0].page_content) # noqa: T201
|
||||
assert docs[0].page_content == (
|
||||
"Hello, world!.\nLLMs are awesome! Langchain is great. Telegram is the best!."
|
||||
)
|
||||
|
Reference in New Issue
Block a user