mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-06 13:33:37 +00:00
@@ -389,7 +389,7 @@ class ArxivAPIWrapper(BaseModel):
|
||||
|
||||
Returns:
|
||||
List of ArxivPaper objects.
|
||||
""" # noqa: E501
|
||||
"""
|
||||
|
||||
def cut_authors(authors: list) -> list[str]:
|
||||
if len(authors) > 3:
|
||||
|
@@ -186,7 +186,7 @@ def replace_imports(file):
|
||||
data = code_block_re.sub(replacer, data)
|
||||
|
||||
# if all_imports:
|
||||
# print(f"Adding {len(all_imports)} links for imports in {file}") # noqa: T201
|
||||
# print(f"Adding {len(all_imports)} links for imports in {file}")
|
||||
with open(file, "w") as f:
|
||||
f.write(data)
|
||||
return all_imports
|
||||
|
@@ -116,7 +116,7 @@ Each LLM integration can optionally provide native implementations for async, st
|
||||
|
||||
{table}
|
||||
|
||||
""" # noqa: E501
|
||||
"""
|
||||
|
||||
CHAT_MODEL_TEMPLATE = """\
|
||||
---
|
||||
@@ -133,7 +133,7 @@ The following table shows all the chat models that support one or more advanced
|
||||
|
||||
{table}
|
||||
|
||||
""" # noqa: E501
|
||||
"""
|
||||
|
||||
|
||||
def get_llm_table():
|
||||
|
Reference in New Issue
Block a user