community[patch]: flattening imports 3 (#18939)

@eyurtsev
This commit is contained in:
Leonid Ganeline 2024-03-12 15:18:54 -07:00 committed by GitHub
parent af50f21765
commit 9c8523b529
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
14 changed files with 507 additions and 2717 deletions

View File

@ -15,48 +15,30 @@
Document Document
""" # noqa: E501 """ # noqa: E501
from langchain_community.document_transformers.beautiful_soup_transformer import ( import importlib
BeautifulSoupTransformer, from typing import Any
)
from langchain_community.document_transformers.doctran_text_extract import (
DoctranPropertyExtractor,
)
from langchain_community.document_transformers.doctran_text_qa import (
DoctranQATransformer,
)
from langchain_community.document_transformers.doctran_text_translate import (
DoctranTextTranslator,
)
from langchain_community.document_transformers.embeddings_redundant_filter import (
EmbeddingsClusteringFilter,
EmbeddingsRedundantFilter,
get_stateful_documents,
)
from langchain_community.document_transformers.google_translate import (
GoogleTranslateTransformer,
)
from langchain_community.document_transformers.html2text import Html2TextTransformer
from langchain_community.document_transformers.long_context_reorder import (
LongContextReorder,
)
from langchain_community.document_transformers.nuclia_text_transform import (
NucliaTextTransformer,
)
from langchain_community.document_transformers.openai_functions import (
OpenAIMetadataTagger,
)
__all__ = [ _module_lookup = {
"BeautifulSoupTransformer", "BeautifulSoupTransformer": "langchain_community.document_transformers.beautiful_soup_transformer", # noqa: E501
"DoctranQATransformer", "DoctranPropertyExtractor": "langchain_community.document_transformers.doctran_text_extract", # noqa: E501
"DoctranTextTranslator", "DoctranQATransformer": "langchain_community.document_transformers.doctran_text_qa",
"DoctranPropertyExtractor", "DoctranTextTranslator": "langchain_community.document_transformers.doctran_text_translate", # noqa: E501
"EmbeddingsClusteringFilter", "EmbeddingsClusteringFilter": "langchain_community.document_transformers.embeddings_redundant_filter", # noqa: E501
"EmbeddingsRedundantFilter", "EmbeddingsRedundantFilter": "langchain_community.document_transformers.embeddings_redundant_filter", # noqa: E501
"GoogleTranslateTransformer", "GoogleTranslateTransformer": "langchain_community.document_transformers.google_translate", # noqa: E501
"get_stateful_documents", "Html2TextTransformer": "langchain_community.document_transformers.html2text",
"LongContextReorder", "LongContextReorder": "langchain_community.document_transformers.long_context_reorder", # noqa: E501
"NucliaTextTransformer", "NucliaTextTransformer": "langchain_community.document_transformers.nuclia_text_transform", # noqa: E501
"OpenAIMetadataTagger", "OpenAIMetadataTagger": "langchain_community.document_transformers.openai_functions", # noqa: E501
"Html2TextTransformer", "get_stateful_documents": "langchain_community.document_transformers.embeddings_redundant_filter", # noqa: E501
] }
def __getattr__(name: str) -> Any:
if name in _module_lookup:
module = importlib.import_module(_module_lookup[name])
return getattr(module, name)
raise AttributeError(f"module {__name__} has no attribute {name}")
__all__ = list(_module_lookup.keys())

View File

@ -11,157 +11,88 @@ from different APIs and services.
""" """
import importlib
import logging import logging
from typing import Any from typing import Any
from langchain_community.embeddings.aleph_alpha import ( _module_lookup = {
AlephAlphaAsymmetricSemanticEmbedding, "AlephAlphaAsymmetricSemanticEmbedding": "langchain_community.embeddings.aleph_alpha", # noqa: E501
AlephAlphaSymmetricSemanticEmbedding, "AlephAlphaSymmetricSemanticEmbedding": "langchain_community.embeddings.aleph_alpha", # noqa: E501
) "AwaEmbeddings": "langchain_community.embeddings.awa",
from langchain_community.embeddings.awa import AwaEmbeddings "AzureOpenAIEmbeddings": "langchain_community.embeddings.azure_openai",
from langchain_community.embeddings.azure_openai import AzureOpenAIEmbeddings "BaichuanTextEmbeddings": "langchain_community.embeddings.baichuan",
from langchain_community.embeddings.baichuan import BaichuanTextEmbeddings "BedrockEmbeddings": "langchain_community.embeddings.bedrock",
from langchain_community.embeddings.baidu_qianfan_endpoint import ( "BookendEmbeddings": "langchain_community.embeddings.bookend",
QianfanEmbeddingsEndpoint, "ClarifaiEmbeddings": "langchain_community.embeddings.clarifai",
) "CohereEmbeddings": "langchain_community.embeddings.cohere",
from langchain_community.embeddings.bedrock import BedrockEmbeddings "DashScopeEmbeddings": "langchain_community.embeddings.dashscope",
from langchain_community.embeddings.bookend import BookendEmbeddings "DatabricksEmbeddings": "langchain_community.embeddings.databricks",
from langchain_community.embeddings.clarifai import ClarifaiEmbeddings "DeepInfraEmbeddings": "langchain_community.embeddings.deepinfra",
from langchain_community.embeddings.cohere import CohereEmbeddings "DeterministicFakeEmbedding": "langchain_community.embeddings.fake",
from langchain_community.embeddings.dashscope import DashScopeEmbeddings "EdenAiEmbeddings": "langchain_community.embeddings.edenai",
from langchain_community.embeddings.databricks import DatabricksEmbeddings "ElasticsearchEmbeddings": "langchain_community.embeddings.elasticsearch",
from langchain_community.embeddings.deepinfra import DeepInfraEmbeddings "EmbaasEmbeddings": "langchain_community.embeddings.embaas",
from langchain_community.embeddings.edenai import EdenAiEmbeddings "ErnieEmbeddings": "langchain_community.embeddings.ernie",
from langchain_community.embeddings.elasticsearch import ElasticsearchEmbeddings "FakeEmbeddings": "langchain_community.embeddings.fake",
from langchain_community.embeddings.embaas import EmbaasEmbeddings "FastEmbedEmbeddings": "langchain_community.embeddings.fastembed",
from langchain_community.embeddings.ernie import ErnieEmbeddings "GPT4AllEmbeddings": "langchain_community.embeddings.gpt4all",
from langchain_community.embeddings.fake import ( "GooglePalmEmbeddings": "langchain_community.embeddings.google_palm",
DeterministicFakeEmbedding, "GradientEmbeddings": "langchain_community.embeddings.gradient_ai",
FakeEmbeddings, "HuggingFaceBgeEmbeddings": "langchain_community.embeddings.huggingface",
) "HuggingFaceEmbeddings": "langchain_community.embeddings.huggingface",
from langchain_community.embeddings.fastembed import FastEmbedEmbeddings "HuggingFaceHubEmbeddings": "langchain_community.embeddings.huggingface_hub",
from langchain_community.embeddings.google_palm import GooglePalmEmbeddings "HuggingFaceInferenceAPIEmbeddings": "langchain_community.embeddings.huggingface",
from langchain_community.embeddings.gpt4all import GPT4AllEmbeddings "HuggingFaceInstructEmbeddings": "langchain_community.embeddings.huggingface",
from langchain_community.embeddings.gradient_ai import GradientEmbeddings "InfinityEmbeddings": "langchain_community.embeddings.infinity",
from langchain_community.embeddings.huggingface import ( "InfinityEmbeddingsLocal": "langchain_community.embeddings.infinity_local",
HuggingFaceBgeEmbeddings, "JavelinAIGatewayEmbeddings": "langchain_community.embeddings.javelin_ai_gateway",
HuggingFaceEmbeddings, "JinaEmbeddings": "langchain_community.embeddings.jina",
HuggingFaceInferenceAPIEmbeddings, "JohnSnowLabsEmbeddings": "langchain_community.embeddings.johnsnowlabs",
HuggingFaceInstructEmbeddings, "LLMRailsEmbeddings": "langchain_community.embeddings.llm_rails",
) "LaserEmbeddings": "langchain_community.embeddings.laser",
from langchain_community.embeddings.huggingface_hub import HuggingFaceHubEmbeddings "LlamaCppEmbeddings": "langchain_community.embeddings.llamacpp",
from langchain_community.embeddings.infinity import InfinityEmbeddings "LlamafileEmbeddings": "langchain_community.embeddings.llamafile",
from langchain_community.embeddings.infinity_local import InfinityEmbeddingsLocal "LocalAIEmbeddings": "langchain_community.embeddings.localai",
from langchain_community.embeddings.javelin_ai_gateway import JavelinAIGatewayEmbeddings "MiniMaxEmbeddings": "langchain_community.embeddings.minimax",
from langchain_community.embeddings.jina import JinaEmbeddings "MlflowAIGatewayEmbeddings": "langchain_community.embeddings.mlflow_gateway",
from langchain_community.embeddings.johnsnowlabs import JohnSnowLabsEmbeddings "MlflowCohereEmbeddings": "langchain_community.embeddings.mlflow",
from langchain_community.embeddings.laser import LaserEmbeddings "MlflowEmbeddings": "langchain_community.embeddings.mlflow",
from langchain_community.embeddings.llamacpp import LlamaCppEmbeddings "ModelScopeEmbeddings": "langchain_community.embeddings.modelscope_hub",
from langchain_community.embeddings.llamafile import LlamafileEmbeddings "MosaicMLInstructorEmbeddings": "langchain_community.embeddings.mosaicml",
from langchain_community.embeddings.llm_rails import LLMRailsEmbeddings "NLPCloudEmbeddings": "langchain_community.embeddings.nlpcloud",
from langchain_community.embeddings.localai import LocalAIEmbeddings "NeMoEmbeddings": "langchain_community.embeddings.nemo",
from langchain_community.embeddings.minimax import MiniMaxEmbeddings "OCIGenAIEmbeddings": "langchain_community.embeddings.oci_generative_ai",
from langchain_community.embeddings.mlflow import ( "OctoAIEmbeddings": "langchain_community.embeddings.octoai_embeddings",
MlflowCohereEmbeddings, "OllamaEmbeddings": "langchain_community.embeddings.ollama",
MlflowEmbeddings, "OpenAIEmbeddings": "langchain_community.embeddings.openai",
) "QianfanEmbeddingsEndpoint": "langchain_community.embeddings.baidu_qianfan_endpoint", # noqa: E501
from langchain_community.embeddings.mlflow_gateway import MlflowAIGatewayEmbeddings "QuantizedBiEncoderEmbeddings": "langchain_community.embeddings.optimum_intel",
from langchain_community.embeddings.modelscope_hub import ModelScopeEmbeddings "SagemakerEndpointEmbeddings": "langchain_community.embeddings.sagemaker_endpoint",
from langchain_community.embeddings.mosaicml import MosaicMLInstructorEmbeddings "SelfHostedEmbeddings": "langchain_community.embeddings.self_hosted",
from langchain_community.embeddings.nemo import NeMoEmbeddings "SelfHostedHuggingFaceEmbeddings": "langchain_community.embeddings.self_hosted_hugging_face", # noqa: E501
from langchain_community.embeddings.nlpcloud import NLPCloudEmbeddings "SelfHostedHuggingFaceInstructEmbeddings": "langchain_community.embeddings.self_hosted_hugging_face", # noqa: E501
from langchain_community.embeddings.oci_generative_ai import OCIGenAIEmbeddings "SentenceTransformerEmbeddings": "langchain_community.embeddings.sentence_transformer", # noqa: E501
from langchain_community.embeddings.octoai_embeddings import OctoAIEmbeddings "SpacyEmbeddings": "langchain_community.embeddings.spacy_embeddings",
from langchain_community.embeddings.ollama import OllamaEmbeddings "SparkLLMTextEmbeddings": "langchain_community.embeddings.sparkllm",
from langchain_community.embeddings.openai import OpenAIEmbeddings "TensorflowHubEmbeddings": "langchain_community.embeddings.tensorflow_hub",
from langchain_community.embeddings.optimum_intel import QuantizedBiEncoderEmbeddings "VertexAIEmbeddings": "langchain_community.embeddings.vertexai",
from langchain_community.embeddings.sagemaker_endpoint import ( "VolcanoEmbeddings": "langchain_community.embeddings.volcengine",
SagemakerEndpointEmbeddings, "VoyageEmbeddings": "langchain_community.embeddings.voyageai",
) "XinferenceEmbeddings": "langchain_community.embeddings.xinference",
from langchain_community.embeddings.self_hosted import SelfHostedEmbeddings }
from langchain_community.embeddings.self_hosted_hugging_face import (
SelfHostedHuggingFaceEmbeddings,
SelfHostedHuggingFaceInstructEmbeddings, def __getattr__(name: str) -> Any:
) if name in _module_lookup:
from langchain_community.embeddings.sentence_transformer import ( module = importlib.import_module(_module_lookup[name])
SentenceTransformerEmbeddings, return getattr(module, name)
) raise AttributeError(f"module {__name__} has no attribute {name}")
from langchain_community.embeddings.spacy_embeddings import SpacyEmbeddings
from langchain_community.embeddings.sparkllm import SparkLLMTextEmbeddings
from langchain_community.embeddings.tensorflow_hub import TensorflowHubEmbeddings __all__ = list(_module_lookup.keys())
from langchain_community.embeddings.vertexai import VertexAIEmbeddings
from langchain_community.embeddings.volcengine import VolcanoEmbeddings
from langchain_community.embeddings.voyageai import VoyageEmbeddings
from langchain_community.embeddings.xinference import XinferenceEmbeddings
logger = logging.getLogger(__name__) logger = logging.getLogger(__name__)
__all__ = [
"OpenAIEmbeddings",
"AzureOpenAIEmbeddings",
"BaichuanTextEmbeddings",
"ClarifaiEmbeddings",
"CohereEmbeddings",
"DatabricksEmbeddings",
"ElasticsearchEmbeddings",
"FastEmbedEmbeddings",
"HuggingFaceEmbeddings",
"HuggingFaceInferenceAPIEmbeddings",
"InfinityEmbeddings",
"InfinityEmbeddingsLocal",
"GradientEmbeddings",
"JinaEmbeddings",
"LaserEmbeddings",
"LlamaCppEmbeddings",
"LlamafileEmbeddings",
"LLMRailsEmbeddings",
"HuggingFaceHubEmbeddings",
"MlflowEmbeddings",
"MlflowCohereEmbeddings",
"MlflowAIGatewayEmbeddings",
"ModelScopeEmbeddings",
"TensorflowHubEmbeddings",
"SagemakerEndpointEmbeddings",
"HuggingFaceInstructEmbeddings",
"MosaicMLInstructorEmbeddings",
"SelfHostedEmbeddings",
"SelfHostedHuggingFaceEmbeddings",
"SelfHostedHuggingFaceInstructEmbeddings",
"FakeEmbeddings",
"DeterministicFakeEmbedding",
"AlephAlphaAsymmetricSemanticEmbedding",
"AlephAlphaSymmetricSemanticEmbedding",
"SentenceTransformerEmbeddings",
"GooglePalmEmbeddings",
"MiniMaxEmbeddings",
"VertexAIEmbeddings",
"BedrockEmbeddings",
"DeepInfraEmbeddings",
"EdenAiEmbeddings",
"DashScopeEmbeddings",
"EmbaasEmbeddings",
"OctoAIEmbeddings",
"SpacyEmbeddings",
"NLPCloudEmbeddings",
"GPT4AllEmbeddings",
"XinferenceEmbeddings",
"LocalAIEmbeddings",
"AwaEmbeddings",
"HuggingFaceBgeEmbeddings",
"ErnieEmbeddings",
"JavelinAIGatewayEmbeddings",
"OllamaEmbeddings",
"QianfanEmbeddingsEndpoint",
"JohnSnowLabsEmbeddings",
"VoyageEmbeddings",
"BookendEmbeddings",
"VolcanoEmbeddings",
"OCIGenAIEmbeddings",
"QuantizedBiEncoderEmbeddings",
"NeMoEmbeddings",
"SparkLLMTextEmbeddings",
]
# TODO: this is in here to maintain backwards compatibility # TODO: this is in here to maintain backwards compatibility
class HypotheticalDocumentEmbedder: class HypotheticalDocumentEmbedder:

View File

@ -1,33 +1,31 @@
"""**Graphs** provide a natural language interface to graph databases.""" """**Graphs** provide a natural language interface to graph databases."""
from langchain_community.graphs.arangodb_graph import ArangoGraph import importlib
from langchain_community.graphs.falkordb_graph import FalkorDBGraph from typing import Any
from langchain_community.graphs.gremlin_graph import GremlinGraph
from langchain_community.graphs.hugegraph import HugeGraph
from langchain_community.graphs.kuzu_graph import KuzuGraph
from langchain_community.graphs.memgraph_graph import MemgraphGraph
from langchain_community.graphs.nebula_graph import NebulaGraph
from langchain_community.graphs.neo4j_graph import Neo4jGraph
from langchain_community.graphs.neptune_graph import NeptuneGraph
from langchain_community.graphs.neptune_rdf_graph import NeptuneRdfGraph
from langchain_community.graphs.networkx_graph import NetworkxEntityGraph
from langchain_community.graphs.ontotext_graphdb_graph import OntotextGraphDBGraph
from langchain_community.graphs.rdf_graph import RdfGraph
from langchain_community.graphs.tigergraph_graph import TigerGraph
__all__ = [ _module_lookup = {
"MemgraphGraph", "ArangoGraph": "langchain_community.graphs.arangodb_graph",
"NetworkxEntityGraph", "FalkorDBGraph": "langchain_community.graphs.falkordb_graph",
"Neo4jGraph", "GremlinGraph": "langchain_community.graphs.gremlin_graph",
"NebulaGraph", "HugeGraph": "langchain_community.graphs.hugegraph",
"NeptuneGraph", "KuzuGraph": "langchain_community.graphs.kuzu_graph",
"NeptuneRdfGraph", "MemgraphGraph": "langchain_community.graphs.memgraph_graph",
"KuzuGraph", "NebulaGraph": "langchain_community.graphs.nebula_graph",
"HugeGraph", "Neo4jGraph": "langchain_community.graphs.neo4j_graph",
"RdfGraph", "NeptuneGraph": "langchain_community.graphs.neptune_graph",
"ArangoGraph", "NeptuneRdfGraph": "langchain_community.graphs.neptune_rdf_graph",
"FalkorDBGraph", "NetworkxEntityGraph": "langchain_community.graphs.networkx_graph",
"TigerGraph", "OntotextGraphDBGraph": "langchain_community.graphs.ontotext_graphdb_graph",
"OntotextGraphDBGraph", "RdfGraph": "langchain_community.graphs.rdf_graph",
"GremlinGraph", "TigerGraph": "langchain_community.graphs.tigergraph_graph",
] }
def __getattr__(name: str) -> Any:
if name in _module_lookup:
module = importlib.import_module(_module_lookup[name])
return getattr(module, name)
raise AttributeError(f"module {__name__} has no attribute {name}")
__all__ = list(_module_lookup.keys())

View File

@ -18,98 +18,55 @@ the backbone of a retriever, but there are other types of retrievers as well.
CallbackManagerForRetrieverRun, AsyncCallbackManagerForRetrieverRun CallbackManagerForRetrieverRun, AsyncCallbackManagerForRetrieverRun
""" """
from langchain_community.retrievers.arcee import ArceeRetriever import importlib
from langchain_community.retrievers.arxiv import ArxivRetriever from typing import Any
from langchain_community.retrievers.azure_cognitive_search import (
AzureCognitiveSearchRetriever,
)
from langchain_community.retrievers.bedrock import AmazonKnowledgeBasesRetriever
from langchain_community.retrievers.bm25 import BM25Retriever
from langchain_community.retrievers.breebs import BreebsRetriever
from langchain_community.retrievers.chaindesk import ChaindeskRetriever
from langchain_community.retrievers.chatgpt_plugin_retriever import (
ChatGPTPluginRetriever,
)
from langchain_community.retrievers.cohere_rag_retriever import CohereRagRetriever
from langchain_community.retrievers.docarray import DocArrayRetriever
from langchain_community.retrievers.elastic_search_bm25 import (
ElasticSearchBM25Retriever,
)
from langchain_community.retrievers.embedchain import EmbedchainRetriever
from langchain_community.retrievers.google_cloud_documentai_warehouse import (
GoogleDocumentAIWarehouseRetriever,
)
from langchain_community.retrievers.google_vertex_ai_search import (
GoogleCloudEnterpriseSearchRetriever,
GoogleVertexAIMultiTurnSearchRetriever,
GoogleVertexAISearchRetriever,
)
from langchain_community.retrievers.kay import KayAiRetriever
from langchain_community.retrievers.kendra import AmazonKendraRetriever
from langchain_community.retrievers.knn import KNNRetriever
from langchain_community.retrievers.llama_index import (
LlamaIndexGraphRetriever,
LlamaIndexRetriever,
)
from langchain_community.retrievers.metal import MetalRetriever
from langchain_community.retrievers.milvus import MilvusRetriever
from langchain_community.retrievers.outline import OutlineRetriever
from langchain_community.retrievers.pinecone_hybrid_search import (
PineconeHybridSearchRetriever,
)
from langchain_community.retrievers.pubmed import PubMedRetriever
from langchain_community.retrievers.qdrant_sparse_vector_retriever import (
QdrantSparseVectorRetriever,
)
from langchain_community.retrievers.remote_retriever import RemoteLangChainRetriever
from langchain_community.retrievers.svm import SVMRetriever
from langchain_community.retrievers.tavily_search_api import TavilySearchAPIRetriever
from langchain_community.retrievers.tfidf import TFIDFRetriever
from langchain_community.retrievers.vespa_retriever import VespaRetriever
from langchain_community.retrievers.weaviate_hybrid_search import (
WeaviateHybridSearchRetriever,
)
from langchain_community.retrievers.wikipedia import WikipediaRetriever
from langchain_community.retrievers.you import YouRetriever
from langchain_community.retrievers.zep import ZepRetriever
from langchain_community.retrievers.zilliz import ZillizRetriever
__all__ = [ _module_lookup = {
"AmazonKendraRetriever", "AmazonKendraRetriever": "langchain_community.retrievers.kendra",
"AmazonKnowledgeBasesRetriever", "AmazonKnowledgeBasesRetriever": "langchain_community.retrievers.bedrock",
"ArceeRetriever", "ArceeRetriever": "langchain_community.retrievers.arcee",
"ArxivRetriever", "ArxivRetriever": "langchain_community.retrievers.arxiv",
"AzureCognitiveSearchRetriever", "AzureCognitiveSearchRetriever": "langchain_community.retrievers.azure_cognitive_search", # noqa: E501
"BM25Retriever", "BM25Retriever": "langchain_community.retrievers.bm25",
"BreebsRetriever", "BreebsRetriever": "langchain_community.retrievers.breebs",
"ChatGPTPluginRetriever", "ChaindeskRetriever": "langchain_community.retrievers.chaindesk",
"ChaindeskRetriever", "ChatGPTPluginRetriever": "langchain_community.retrievers.chatgpt_plugin_retriever",
"CohereRagRetriever", "CohereRagRetriever": "langchain_community.retrievers.cohere_rag_retriever",
"ElasticSearchBM25Retriever", "DocArrayRetriever": "langchain_community.retrievers.docarray",
"EmbedchainRetriever", "ElasticSearchBM25Retriever": "langchain_community.retrievers.elastic_search_bm25",
"GoogleDocumentAIWarehouseRetriever", "EmbedchainRetriever": "langchain_community.retrievers.embedchain",
"GoogleCloudEnterpriseSearchRetriever", "GoogleCloudEnterpriseSearchRetriever": "langchain_community.retrievers.google_vertex_ai_search", # noqa: E501
"GoogleVertexAIMultiTurnSearchRetriever", "GoogleDocumentAIWarehouseRetriever": "langchain_community.retrievers.google_cloud_documentai_warehouse", # noqa: E501
"GoogleVertexAISearchRetriever", "GoogleVertexAIMultiTurnSearchRetriever": "langchain_community.retrievers.google_vertex_ai_search", # noqa: E501
"KayAiRetriever", "GoogleVertexAISearchRetriever": "langchain_community.retrievers.google_vertex_ai_search", # noqa: E501
"KNNRetriever", "KNNRetriever": "langchain_community.retrievers.knn",
"LlamaIndexGraphRetriever", "KayAiRetriever": "langchain_community.retrievers.kay",
"LlamaIndexRetriever", "LlamaIndexGraphRetriever": "langchain_community.retrievers.llama_index",
"MetalRetriever", "LlamaIndexRetriever": "langchain_community.retrievers.llama_index",
"MilvusRetriever", "MetalRetriever": "langchain_community.retrievers.metal",
"OutlineRetriever", "MilvusRetriever": "langchain_community.retrievers.milvus",
"PineconeHybridSearchRetriever", "OutlineRetriever": "langchain_community.retrievers.outline",
"PubMedRetriever", "PineconeHybridSearchRetriever": "langchain_community.retrievers.pinecone_hybrid_search", # noqa: E501
"QdrantSparseVectorRetriever", "PubMedRetriever": "langchain_community.retrievers.pubmed",
"RemoteLangChainRetriever", "QdrantSparseVectorRetriever": "langchain_community.retrievers.qdrant_sparse_vector_retriever", # noqa: E501
"SVMRetriever", "RemoteLangChainRetriever": "langchain_community.retrievers.remote_retriever",
"TavilySearchAPIRetriever", "SVMRetriever": "langchain_community.retrievers.svm",
"TFIDFRetriever", "TFIDFRetriever": "langchain_community.retrievers.tfidf",
"VespaRetriever", "TavilySearchAPIRetriever": "langchain_community.retrievers.tavily_search_api",
"WeaviateHybridSearchRetriever", "VespaRetriever": "langchain_community.retrievers.vespa_retriever",
"WikipediaRetriever", "WeaviateHybridSearchRetriever": "langchain_community.retrievers.weaviate_hybrid_search", # noqa: E501
"YouRetriever", "WikipediaRetriever": "langchain_community.retrievers.wikipedia",
"ZepRetriever", "YouRetriever": "langchain_community.retrievers.you",
"ZillizRetriever", "ZepRetriever": "langchain_community.retrievers.zep",
"DocArrayRetriever", "ZillizRetriever": "langchain_community.retrievers.zilliz",
] }
def __getattr__(name: str) -> Any:
if name in _module_lookup:
module = importlib.import_module(_module_lookup[name])
return getattr(module, name)
raise AttributeError(f"module {__name__} has no attribute {name}")
__all__ = list(_module_lookup.keys())

View File

@ -14,22 +14,24 @@ The primary goal of these storages is to support caching.
""" # noqa: E501 """ # noqa: E501
from langchain_community.storage.astradb import ( import importlib
AstraDBByteStore, from typing import Any
AstraDBStore,
)
from langchain_community.storage.mongodb import MongoDBStore
from langchain_community.storage.redis import RedisStore
from langchain_community.storage.upstash_redis import (
UpstashRedisByteStore,
UpstashRedisStore,
)
__all__ = [ _module_lookup = {
"AstraDBStore", "AstraDBByteStore": "langchain_community.storage.astradb",
"AstraDBByteStore", "AstraDBStore": "langchain_community.storage.astradb",
"MongoDBStore", "MongoDBStore": "langchain_community.storage.mongodb",
"RedisStore", "RedisStore": "langchain_community.storage.redis",
"UpstashRedisByteStore", "UpstashRedisByteStore": "langchain_community.storage.upstash_redis",
"UpstashRedisStore", "UpstashRedisStore": "langchain_community.storage.upstash_redis",
] }
def __getattr__(name: str) -> Any:
if name in _module_lookup:
module = importlib.import_module(_module_lookup[name])
return getattr(module, name)
raise AttributeError(f"module {__name__} has no attribute {name}")
__all__ = list(_module_lookup.keys())

File diff suppressed because it is too large Load Diff

View File

@ -3,456 +3,72 @@
Other LangChain classes use **Utilities** to interact with third-part systems Other LangChain classes use **Utilities** to interact with third-part systems
and packages. and packages.
""" """
import importlib
from typing import Any from typing import Any
from langchain_community.utilities.requests import ( _module_lookup = {
Requests, "AlphaVantageAPIWrapper": "langchain_community.utilities.alpha_vantage",
RequestsWrapper, "ApifyWrapper": "langchain_community.utilities.apify",
TextRequestsWrapper, "ArceeWrapper": "langchain_community.utilities.arcee",
) "ArxivAPIWrapper": "langchain_community.utilities.arxiv",
"AudioStream": "langchain_community.utilities.nvidia_riva",
"BibtexparserWrapper": "langchain_community.utilities.bibtex",
def _import_alpha_vantage() -> Any: "BingSearchAPIWrapper": "langchain_community.utilities.bing_search",
from langchain_community.utilities.alpha_vantage import AlphaVantageAPIWrapper "BraveSearchWrapper": "langchain_community.utilities.brave_search",
"DuckDuckGoSearchAPIWrapper": "langchain_community.utilities.duckduckgo_search",
return AlphaVantageAPIWrapper "GoldenQueryAPIWrapper": "langchain_community.utilities.golden_query",
"GoogleFinanceAPIWrapper": "langchain_community.utilities.google_finance",
"GoogleJobsAPIWrapper": "langchain_community.utilities.google_jobs",
def _import_apify() -> Any: "GoogleLensAPIWrapper": "langchain_community.utilities.google_lens",
from langchain_community.utilities.apify import ApifyWrapper "GooglePlacesAPIWrapper": "langchain_community.utilities.google_places_api",
"GoogleScholarAPIWrapper": "langchain_community.utilities.google_scholar",
return ApifyWrapper "GoogleSearchAPIWrapper": "langchain_community.utilities.google_search",
"GoogleSerperAPIWrapper": "langchain_community.utilities.google_serper",
"GoogleTrendsAPIWrapper": "langchain_community.utilities.google_trends",
def _import_arcee() -> Any: "GraphQLAPIWrapper": "langchain_community.utilities.graphql",
from langchain_community.utilities.arcee import ArceeWrapper "JiraAPIWrapper": "langchain_community.utilities.jira",
"LambdaWrapper": "langchain_community.utilities.awslambda",
return ArceeWrapper "MaxComputeAPIWrapper": "langchain_community.utilities.max_compute",
"MerriamWebsterAPIWrapper": "langchain_community.utilities.merriam_webster",
"MetaphorSearchAPIWrapper": "langchain_community.utilities.metaphor_search",
def _import_arxiv() -> Any: "NVIDIARivaASR": "langchain_community.utilities.nvidia_riva",
from langchain_community.utilities.arxiv import ArxivAPIWrapper "NVIDIARivaStream": "langchain_community.utilities.nvidia_riva",
"NVIDIARivaTTS": "langchain_community.utilities.nvidia_riva",
return ArxivAPIWrapper "NasaAPIWrapper": "langchain_community.utilities.nasa",
"NutritionAIAPI": "langchain_community.utilities.passio_nutrition_ai",
"OpenWeatherMapAPIWrapper": "langchain_community.utilities.openweathermap",
def _import_awslambda() -> Any: "OutlineAPIWrapper": "langchain_community.utilities.outline",
from langchain_community.utilities.awslambda import LambdaWrapper "Portkey": "langchain_community.utilities.portkey",
"PowerBIDataset": "langchain_community.utilities.powerbi",
return LambdaWrapper "PubMedAPIWrapper": "langchain_community.utilities.pubmed",
"PythonREPL": "langchain_community.utilities.python",
"Requests": "langchain_community.utilities.requests",
def _import_bibtex() -> Any: "RequestsWrapper": "langchain_community.utilities.requests",
from langchain_community.utilities.bibtex import BibtexparserWrapper "RivaASR": "langchain_community.utilities.nvidia_riva",
"RivaTTS": "langchain_community.utilities.nvidia_riva",
return BibtexparserWrapper "SQLDatabase": "langchain_community.utilities.sql_database",
"SceneXplainAPIWrapper": "langchain_community.utilities.scenexplain",
"SearchApiAPIWrapper": "langchain_community.utilities.searchapi",
def _import_bing_search() -> Any: "SearxSearchWrapper": "langchain_community.utilities.searx_search",
from langchain_community.utilities.bing_search import BingSearchAPIWrapper "SerpAPIWrapper": "langchain_community.utilities.serpapi",
"SparkSQL": "langchain_community.utilities.spark_sql",
return BingSearchAPIWrapper "StackExchangeAPIWrapper": "langchain_community.utilities.stackexchange",
"SteamWebAPIWrapper": "langchain_community.utilities.steam",
"TensorflowDatasets": "langchain_community.utilities.tensorflow_datasets",
def _import_brave_search() -> Any: "TextRequestsWrapper": "langchain_community.utilities.requests",
from langchain_community.utilities.brave_search import BraveSearchWrapper "TwilioAPIWrapper": "langchain_community.utilities.twilio",
"WikipediaAPIWrapper": "langchain_community.utilities.wikipedia",
return BraveSearchWrapper "WolframAlphaAPIWrapper": "langchain_community.utilities.wolfram_alpha",
"YouSearchAPIWrapper": "langchain_community.utilities.you",
"ZapierNLAWrapper": "langchain_community.utilities.zapier",
def _import_duckduckgo_search() -> Any: }
from langchain_community.utilities.duckduckgo_search import (
DuckDuckGoSearchAPIWrapper,
)
return DuckDuckGoSearchAPIWrapper
def _import_golden_query() -> Any:
from langchain_community.utilities.golden_query import GoldenQueryAPIWrapper
return GoldenQueryAPIWrapper
def _import_google_lens() -> Any:
from langchain_community.utilities.google_lens import GoogleLensAPIWrapper
return GoogleLensAPIWrapper
def _import_google_places_api() -> Any:
from langchain_community.utilities.google_places_api import GooglePlacesAPIWrapper
return GooglePlacesAPIWrapper
def _import_google_jobs() -> Any:
from langchain_community.utilities.google_jobs import GoogleJobsAPIWrapper
return GoogleJobsAPIWrapper
def _import_google_scholar() -> Any:
from langchain_community.utilities.google_scholar import GoogleScholarAPIWrapper
return GoogleScholarAPIWrapper
def _import_google_trends() -> Any:
from langchain_community.utilities.google_trends import GoogleTrendsAPIWrapper
return GoogleTrendsAPIWrapper
def _import_google_finance() -> Any:
from langchain_community.utilities.google_finance import GoogleFinanceAPIWrapper
return GoogleFinanceAPIWrapper
def _import_google_search() -> Any:
from langchain_community.utilities.google_search import GoogleSearchAPIWrapper
return GoogleSearchAPIWrapper
def _import_google_serper() -> Any:
from langchain_community.utilities.google_serper import GoogleSerperAPIWrapper
return GoogleSerperAPIWrapper
def _import_graphql() -> Any:
from langchain_community.utilities.graphql import GraphQLAPIWrapper
return GraphQLAPIWrapper
def _import_jira() -> Any:
from langchain_community.utilities.jira import JiraAPIWrapper
return JiraAPIWrapper
def _import_max_compute() -> Any:
from langchain_community.utilities.max_compute import MaxComputeAPIWrapper
return MaxComputeAPIWrapper
def _import_merriam_webster() -> Any:
from langchain_community.utilities.merriam_webster import MerriamWebsterAPIWrapper
return MerriamWebsterAPIWrapper
def _import_metaphor_search() -> Any:
from langchain_community.utilities.metaphor_search import MetaphorSearchAPIWrapper
return MetaphorSearchAPIWrapper
def _import_openweathermap() -> Any:
from langchain_community.utilities.openweathermap import OpenWeatherMapAPIWrapper
return OpenWeatherMapAPIWrapper
def _import_outline() -> Any:
from langchain_community.utilities.outline import OutlineAPIWrapper
return OutlineAPIWrapper
def _import_passio_nutrition_ai() -> Any:
from langchain_community.utilities.passio_nutrition_ai import NutritionAIAPI
return NutritionAIAPI
def _import_portkey() -> Any:
from langchain_community.utilities.portkey import Portkey
return Portkey
def _import_powerbi() -> Any:
from langchain_community.utilities.powerbi import PowerBIDataset
return PowerBIDataset
def _import_pubmed() -> Any:
from langchain_community.utilities.pubmed import PubMedAPIWrapper
return PubMedAPIWrapper
def _import_python() -> Any:
from langchain_community.utilities.python import PythonREPL
return PythonREPL
def _import_scenexplain() -> Any:
from langchain_community.utilities.scenexplain import SceneXplainAPIWrapper
return SceneXplainAPIWrapper
def _import_searchapi() -> Any:
from langchain_community.utilities.searchapi import SearchApiAPIWrapper
return SearchApiAPIWrapper
def _import_searx_search() -> Any:
from langchain_community.utilities.searx_search import SearxSearchWrapper
return SearxSearchWrapper
def _import_serpapi() -> Any:
from langchain_community.utilities.serpapi import SerpAPIWrapper
return SerpAPIWrapper
def _import_spark_sql() -> Any:
from langchain_community.utilities.spark_sql import SparkSQL
return SparkSQL
def _import_sql_database() -> Any:
from langchain_community.utilities.sql_database import SQLDatabase
return SQLDatabase
def _import_steam_webapi() -> Any:
from langchain_community.utilities.steam import SteamWebAPIWrapper
return SteamWebAPIWrapper
def _import_stackexchange() -> Any:
from langchain_community.utilities.stackexchange import StackExchangeAPIWrapper
return StackExchangeAPIWrapper
def _import_tensorflow_datasets() -> Any:
from langchain_community.utilities.tensorflow_datasets import TensorflowDatasets
return TensorflowDatasets
def _import_twilio() -> Any:
from langchain_community.utilities.twilio import TwilioAPIWrapper
return TwilioAPIWrapper
def _import_you() -> Any:
from langchain_community.utilities.you import YouSearchAPIWrapper
return YouSearchAPIWrapper
def _import_wikipedia() -> Any:
from langchain_community.utilities.wikipedia import WikipediaAPIWrapper
return WikipediaAPIWrapper
def _import_wolfram_alpha() -> Any:
from langchain_community.utilities.wolfram_alpha import WolframAlphaAPIWrapper
return WolframAlphaAPIWrapper
def _import_zapier() -> Any:
from langchain_community.utilities.zapier import ZapierNLAWrapper
return ZapierNLAWrapper
def _import_nasa() -> Any:
from langchain_community.utilities.nasa import NasaAPIWrapper
return NasaAPIWrapper
def _import_nvidia_riva_asr() -> Any:
from langchain_community.utilities.nvidia_riva import RivaASR
return RivaASR
def _import_nvidia_riva_tts() -> Any:
from langchain_community.utilities.nvidia_riva import RivaTTS
return RivaTTS
def _import_nvidia_riva_stream() -> Any:
from langchain_community.utilities.nvidia_riva import AudioStream
return AudioStream
def __getattr__(name: str) -> Any: def __getattr__(name: str) -> Any:
if name == "AlphaVantageAPIWrapper": if name in _module_lookup:
return _import_alpha_vantage() module = importlib.import_module(_module_lookup[name])
elif name == "ApifyWrapper": return getattr(module, name)
return _import_apify() raise AttributeError(f"module {__name__} has no attribute {name}")
elif name == "ArceeWrapper":
return _import_arcee()
elif name == "ArxivAPIWrapper":
return _import_arxiv()
elif name == "LambdaWrapper":
return _import_awslambda()
elif name == "BibtexparserWrapper":
return _import_bibtex()
elif name == "BingSearchAPIWrapper":
return _import_bing_search()
elif name == "BraveSearchWrapper":
return _import_brave_search()
elif name == "DuckDuckGoSearchAPIWrapper":
return _import_duckduckgo_search()
elif name == "GoogleLensAPIWrapper":
return _import_google_lens()
elif name == "GoldenQueryAPIWrapper":
return _import_golden_query()
elif name == "GoogleJobsAPIWrapper":
return _import_google_jobs()
elif name == "GoogleScholarAPIWrapper":
return _import_google_scholar()
elif name == "GoogleFinanceAPIWrapper":
return _import_google_finance()
elif name == "GoogleTrendsAPIWrapper":
return _import_google_trends()
elif name == "GooglePlacesAPIWrapper":
return _import_google_places_api()
elif name == "GoogleSearchAPIWrapper":
return _import_google_search()
elif name == "GoogleSerperAPIWrapper":
return _import_google_serper()
elif name == "GraphQLAPIWrapper":
return _import_graphql()
elif name == "JiraAPIWrapper":
return _import_jira()
elif name == "MaxComputeAPIWrapper":
return _import_max_compute()
elif name == "MerriamWebsterAPIWrapper":
return _import_merriam_webster()
elif name == "MetaphorSearchAPIWrapper":
return _import_metaphor_search()
elif name == "NasaAPIWrapper":
return _import_nasa()
elif name == "NVIDIARivaASR":
return _import_nvidia_riva_asr()
elif name == "NVIDIARivaStream":
return _import_nvidia_riva_stream()
elif name == "NVIDIARivaTTS":
return _import_nvidia_riva_tts()
elif name == "OpenWeatherMapAPIWrapper":
return _import_openweathermap()
elif name == "OutlineAPIWrapper":
return _import_outline()
elif name == "NutritionAIAPI":
return _import_passio_nutrition_ai()
elif name == "Portkey":
return _import_portkey()
elif name == "PowerBIDataset":
return _import_powerbi()
elif name == "PubMedAPIWrapper":
return _import_pubmed()
elif name == "PythonREPL":
return _import_python()
elif name == "SceneXplainAPIWrapper":
return _import_scenexplain()
elif name == "SearchApiAPIWrapper":
return _import_searchapi()
elif name == "SearxSearchWrapper":
return _import_searx_search()
elif name == "SerpAPIWrapper":
return _import_serpapi()
elif name == "SparkSQL":
return _import_spark_sql()
elif name == "StackExchangeAPIWrapper":
return _import_stackexchange()
elif name == "SQLDatabase":
return _import_sql_database()
elif name == "SteamWebAPIWrapper":
return _import_steam_webapi()
elif name == "TensorflowDatasets":
return _import_tensorflow_datasets()
elif name == "TwilioAPIWrapper":
return _import_twilio()
elif name == "YouSearchAPIWrapper":
return _import_you()
elif name == "WikipediaAPIWrapper":
return _import_wikipedia()
elif name == "WolframAlphaAPIWrapper":
return _import_wolfram_alpha()
elif name == "ZapierNLAWrapper":
return _import_zapier()
else:
raise AttributeError(f"Could not find: {name}")
__all__ = [ __all__ = list(_module_lookup.keys())
"AlphaVantageAPIWrapper",
"ApifyWrapper",
"ArceeWrapper",
"ArxivAPIWrapper",
"BibtexparserWrapper",
"BingSearchAPIWrapper",
"BraveSearchWrapper",
"DuckDuckGoSearchAPIWrapper",
"GoldenQueryAPIWrapper",
"GoogleFinanceAPIWrapper",
"GoogleLensAPIWrapper",
"GoogleJobsAPIWrapper",
"GooglePlacesAPIWrapper",
"GoogleScholarAPIWrapper",
"GoogleTrendsAPIWrapper",
"GoogleSearchAPIWrapper",
"GoogleSerperAPIWrapper",
"GraphQLAPIWrapper",
"JiraAPIWrapper",
"LambdaWrapper",
"MaxComputeAPIWrapper",
"MerriamWebsterAPIWrapper",
"MetaphorSearchAPIWrapper",
"NasaAPIWrapper",
"NVIDIARivaASR",
"NVIDIARivaStream",
"NVIDIARivaTTS",
"OpenWeatherMapAPIWrapper",
"OutlineAPIWrapper",
"NutritionAIAPI",
"Portkey",
"PowerBIDataset",
"PubMedAPIWrapper",
"PythonREPL",
"Requests",
"RequestsWrapper",
"SteamWebAPIWrapper",
"SQLDatabase",
"SceneXplainAPIWrapper",
"SearchApiAPIWrapper",
"SearxSearchWrapper",
"SerpAPIWrapper",
"SparkSQL",
"StackExchangeAPIWrapper",
"TensorflowDatasets",
"TextRequestsWrapper",
"TwilioAPIWrapper",
"YouSearchAPIWrapper",
"WikipediaAPIWrapper",
"WolframAlphaAPIWrapper",
"ZapierNLAWrapper",
]

View File

@ -665,3 +665,9 @@ class RivaTTS(
await producer await producer
await consumer await consumer
# Backwards compatibility:
NVIDIARivaASR = RivaASR
NVIDIARivaTTS = RivaTTS
NVIDIARivaStream = AudioStream

View File

@ -19,759 +19,100 @@ and retrieve the data that are 'most similar' to the embedded query.
Embeddings, Document Embeddings, Document
""" # noqa: E501 """ # noqa: E501
import importlib
from typing import Any from typing import Any
from langchain_core.vectorstores import VectorStore _module_lookup = {
"AlibabaCloudOpenSearch": "langchain_community.vectorstores.alibabacloud_opensearch", # noqa: E501
"AlibabaCloudOpenSearchSettings": "langchain_community.vectorstores.alibabacloud_opensearch", # noqa: E501
def _import_alibaba_cloud_open_search() -> Any: "AnalyticDB": "langchain_community.vectorstores.analyticdb",
from langchain_community.vectorstores.alibabacloud_opensearch import ( "Annoy": "langchain_community.vectorstores.annoy",
AlibabaCloudOpenSearch, "ApacheDoris": "langchain_community.vectorstores.apache_doris",
) "AstraDB": "langchain_community.vectorstores.astradb",
"AtlasDB": "langchain_community.vectorstores.atlas",
return AlibabaCloudOpenSearch "AwaDB": "langchain_community.vectorstores.awadb",
"AzureCosmosDBVectorSearch": "langchain_community.vectorstores.azure_cosmos_db",
"AzureSearch": "langchain_community.vectorstores.azuresearch",
def _import_alibaba_cloud_open_search_settings() -> Any: "BESVectorStore": "langchain_community.vectorstores.baiducloud_vector_search",
from langchain_community.vectorstores.alibabacloud_opensearch import ( "Bagel": "langchain_community.vectorstores.bageldb",
AlibabaCloudOpenSearchSettings, "BigQueryVectorSearch": "langchain_community.vectorstores.bigquery_vector_search",
) "Cassandra": "langchain_community.vectorstores.cassandra",
"Chroma": "langchain_community.vectorstores.chroma",
return AlibabaCloudOpenSearchSettings "Clarifai": "langchain_community.vectorstores.clarifai",
"Clickhouse": "langchain_community.vectorstores.clickhouse",
"ClickhouseSettings": "langchain_community.vectorstores.clickhouse",
def _import_azure_cosmos_db() -> Any: "DashVector": "langchain_community.vectorstores.dashvector",
from langchain_community.vectorstores.azure_cosmos_db import ( "DatabricksVectorSearch": "langchain_community.vectorstores.databricks_vector_search", # noqa: E501
AzureCosmosDBVectorSearch, "DeepLake": "langchain_community.vectorstores.deeplake",
) "Dingo": "langchain_community.vectorstores.dingo",
"DistanceStrategy": "langchain_community.vectorstores.kinetica",
return AzureCosmosDBVectorSearch "DocArrayHnswSearch": "langchain_community.vectorstores.docarray",
"DocArrayInMemorySearch": "langchain_community.vectorstores.docarray",
"DocumentDBVectorSearch": "langchain_community.vectorstores.documentdb",
def _import_elastic_knn_search() -> Any: "ElasticKnnSearch": "langchain_community.vectorstores.elastic_vector_search",
from langchain_community.vectorstores.elastic_vector_search import ElasticKnnSearch "ElasticVectorSearch": "langchain_community.vectorstores.elastic_vector_search",
"ElasticsearchStore": "langchain_community.vectorstores.elasticsearch",
return ElasticKnnSearch "Epsilla": "langchain_community.vectorstores.epsilla",
"FAISS": "langchain_community.vectorstores.faiss",
"HanaDB": "langchain_community.vectorstores.hanavector",
def _import_elastic_vector_search() -> Any: "Hologres": "langchain_community.vectorstores.hologres",
from langchain_community.vectorstores.elastic_vector_search import ( "InfinispanVS": "langchain_community.vectorstores.infinispanvs",
ElasticVectorSearch, "KDBAI": "langchain_community.vectorstores.kdbai",
) "Kinetica": "langchain_community.vectorstores.kinetica",
"KineticaSettings": "langchain_community.vectorstores.kinetica",
return ElasticVectorSearch "LLMRails": "langchain_community.vectorstores.llm_rails",
"LanceDB": "langchain_community.vectorstores.lancedb",
"Lantern": "langchain_community.vectorstores.lantern",
def _import_analyticdb() -> Any: "Marqo": "langchain_community.vectorstores.marqo",
from langchain_community.vectorstores.analyticdb import AnalyticDB "MatchingEngine": "langchain_community.vectorstores.matching_engine",
"Meilisearch": "langchain_community.vectorstores.meilisearch",
return AnalyticDB "Milvus": "langchain_community.vectorstores.milvus",
"MomentoVectorIndex": "langchain_community.vectorstores.momento_vector_index",
"MongoDBAtlasVectorSearch": "langchain_community.vectorstores.mongodb_atlas",
def _import_annoy() -> Any: "MyScale": "langchain_community.vectorstores.myscale",
from langchain_community.vectorstores.annoy import Annoy "MyScaleSettings": "langchain_community.vectorstores.myscale",
"Neo4jVector": "langchain_community.vectorstores.neo4j_vector",
return Annoy "NeuralDBVectorStore": "langchain_community.vectorstores.thirdai_neuraldb",
"OpenSearchVectorSearch": "langchain_community.vectorstores.opensearch_vector_search", # noqa: E501
"PGEmbedding": "langchain_community.vectorstores.pgembedding",
def _import_apache_doris() -> Any: "PGVector": "langchain_community.vectorstores.pgvector",
from langchain_community.vectorstores.apache_doris import ApacheDoris "Pinecone": "langchain_community.vectorstores.pinecone",
"Qdrant": "langchain_community.vectorstores.qdrant",
return ApacheDoris "Redis": "langchain_community.vectorstores.redis",
"Rockset": "langchain_community.vectorstores.rocksetdb",
"SKLearnVectorStore": "langchain_community.vectorstores.sklearn",
def _import_atlas() -> Any: "SQLiteVSS": "langchain_community.vectorstores.sqlitevss",
from langchain_community.vectorstores.atlas import AtlasDB "ScaNN": "langchain_community.vectorstores.scann",
"SemaDB": "langchain_community.vectorstores.semadb",
return AtlasDB "SingleStoreDB": "langchain_community.vectorstores.singlestoredb",
"StarRocks": "langchain_community.vectorstores.starrocks",
"SupabaseVectorStore": "langchain_community.vectorstores.supabase",
def _import_awadb() -> Any: "SurrealDBStore": "langchain_community.vectorstores.surrealdb",
from langchain_community.vectorstores.awadb import AwaDB "Tair": "langchain_community.vectorstores.tair",
"TencentVectorDB": "langchain_community.vectorstores.tencentvectordb",
return AwaDB "TiDBVectorStore": "langchain_community.vectorstores.tidb_vector",
"Tigris": "langchain_community.vectorstores.tigris",
"TileDB": "langchain_community.vectorstores.tiledb",
def _import_azuresearch() -> Any: "TimescaleVector": "langchain_community.vectorstores.timescalevector",
from langchain_community.vectorstores.azuresearch import AzureSearch "Typesense": "langchain_community.vectorstores.typesense",
"USearch": "langchain_community.vectorstores.usearch",
return AzureSearch "Vald": "langchain_community.vectorstores.vald",
"Vearch": "langchain_community.vectorstores.vearch",
"Vectara": "langchain_community.vectorstores.vectara",
def _import_bageldb() -> Any: "VectorStore": "langchain_core.vectorstores",
from langchain_community.vectorstores.bageldb import Bagel "VespaStore": "langchain_community.vectorstores.vespa",
"Weaviate": "langchain_community.vectorstores.weaviate",
return Bagel "Yellowbrick": "langchain_community.vectorstores.yellowbrick",
"ZepVectorStore": "langchain_community.vectorstores.zep",
"Zilliz": "langchain_community.vectorstores.zilliz",
def _import_baiducloud_vector_search() -> Any: }
from langchain_community.vectorstores.baiducloud_vector_search import BESVectorStore
return BESVectorStore
def _import_bigquery() -> Any:
from langchain_community.vectorstores.bigquery_vector_search import (
BigQueryVectorSearch,
)
return BigQueryVectorSearch
def _import_cassandra() -> Any:
from langchain_community.vectorstores.cassandra import Cassandra
return Cassandra
def _import_astradb() -> Any:
from langchain_community.vectorstores.astradb import AstraDB
return AstraDB
def _import_chroma() -> Any:
from langchain_community.vectorstores.chroma import Chroma
return Chroma
def _import_clarifai() -> Any:
from langchain_community.vectorstores.clarifai import Clarifai
return Clarifai
def _import_clickhouse() -> Any:
from langchain_community.vectorstores.clickhouse import Clickhouse
return Clickhouse
def _import_clickhouse_settings() -> Any:
from langchain_community.vectorstores.clickhouse import ClickhouseSettings
return ClickhouseSettings
def _import_dashvector() -> Any:
from langchain_community.vectorstores.dashvector import DashVector
return DashVector
def _import_databricks_vector_search() -> Any:
from langchain_community.vectorstores.databricks_vector_search import (
DatabricksVectorSearch,
)
return DatabricksVectorSearch
def _import_deeplake() -> Any:
from langchain_community.vectorstores.deeplake import DeepLake
return DeepLake
def _import_dingo() -> Any:
from langchain_community.vectorstores.dingo import Dingo
return Dingo
def _import_docarray_hnsw() -> Any:
from langchain_community.vectorstores.docarray import DocArrayHnswSearch
return DocArrayHnswSearch
def _import_docarray_inmemory() -> Any:
from langchain_community.vectorstores.docarray import DocArrayInMemorySearch
return DocArrayInMemorySearch
def _import_documentdb() -> Any:
from langchain_community.vectorstores.documentdb import (
DocumentDBVectorSearch,
)
return DocumentDBVectorSearch
def _import_elasticsearch() -> Any:
from langchain_community.vectorstores.elasticsearch import ElasticsearchStore
return ElasticsearchStore
def _import_epsilla() -> Any:
from langchain_community.vectorstores.epsilla import Epsilla
return Epsilla
def _import_faiss() -> Any:
from langchain_community.vectorstores.faiss import FAISS
return FAISS
def _import_hanavector() -> Any:
from langchain_community.vectorstores.hanavector import HanaDB
return HanaDB
def _import_kinetica() -> Any:
from langchain_community.vectorstores.kinetica import Kinetica
return Kinetica
def _import_kinetica_settings() -> Any:
from langchain_community.vectorstores.kinetica import KineticaSettings
return KineticaSettings
def _import_distance_strategy() -> Any:
from langchain_community.vectorstores.kinetica import DistanceStrategy
return DistanceStrategy
def _import_hologres() -> Any:
from langchain_community.vectorstores.hologres import Hologres
return Hologres
def _import_infinispanvs() -> Any:
from langchain_community.vectorstores.infinispanvs import InfinispanVS
return InfinispanVS
def _import_kdbai() -> Any:
from langchain_community.vectorstores.kdbai import KDBAI
return KDBAI
def _import_lancedb() -> Any:
from langchain_community.vectorstores.lancedb import LanceDB
return LanceDB
def _import_llm_rails() -> Any:
from langchain_community.vectorstores.llm_rails import LLMRails
return LLMRails
def _import_marqo() -> Any:
from langchain_community.vectorstores.marqo import Marqo
return Marqo
def _import_matching_engine() -> Any:
from langchain_community.vectorstores.matching_engine import MatchingEngine
return MatchingEngine
def _import_meilisearch() -> Any:
from langchain_community.vectorstores.meilisearch import Meilisearch
return Meilisearch
def _import_milvus() -> Any:
from langchain_community.vectorstores.milvus import Milvus
return Milvus
def _import_momento_vector_index() -> Any:
from langchain_community.vectorstores.momento_vector_index import MomentoVectorIndex
return MomentoVectorIndex
def _import_mongodb_atlas() -> Any:
from langchain_community.vectorstores.mongodb_atlas import MongoDBAtlasVectorSearch
return MongoDBAtlasVectorSearch
def _import_myscale() -> Any:
from langchain_community.vectorstores.myscale import MyScale
return MyScale
def _import_myscale_settings() -> Any:
from langchain_community.vectorstores.myscale import MyScaleSettings
return MyScaleSettings
def _import_neo4j_vector() -> Any:
from langchain_community.vectorstores.neo4j_vector import Neo4jVector
return Neo4jVector
def _import_opensearch_vector_search() -> Any:
from langchain_community.vectorstores.opensearch_vector_search import (
OpenSearchVectorSearch,
)
return OpenSearchVectorSearch
def _import_pgembedding() -> Any:
from langchain_community.vectorstores.pgembedding import PGEmbedding
return PGEmbedding
def _import_pgvector() -> Any:
from langchain_community.vectorstores.pgvector import PGVector
return PGVector
def _import_pinecone() -> Any:
from langchain_community.vectorstores.pinecone import Pinecone
return Pinecone
def _import_qdrant() -> Any:
from langchain_community.vectorstores.qdrant import Qdrant
return Qdrant
def _import_redis() -> Any:
from langchain_community.vectorstores.redis import Redis
return Redis
def _import_rocksetdb() -> Any:
from langchain_community.vectorstores.rocksetdb import Rockset
return Rockset
def _import_vespa() -> Any:
from langchain_community.vectorstores.vespa import VespaStore
return VespaStore
def _import_scann() -> Any:
from langchain_community.vectorstores.scann import ScaNN
return ScaNN
def _import_semadb() -> Any:
from langchain_community.vectorstores.semadb import SemaDB
return SemaDB
def _import_singlestoredb() -> Any:
from langchain_community.vectorstores.singlestoredb import SingleStoreDB
return SingleStoreDB
def _import_sklearn() -> Any:
from langchain_community.vectorstores.sklearn import SKLearnVectorStore
return SKLearnVectorStore
def _import_sqlitevss() -> Any:
from langchain_community.vectorstores.sqlitevss import SQLiteVSS
return SQLiteVSS
def _import_starrocks() -> Any:
from langchain_community.vectorstores.starrocks import StarRocks
return StarRocks
def _import_supabase() -> Any:
from langchain_community.vectorstores.supabase import SupabaseVectorStore
return SupabaseVectorStore
def _import_surrealdb() -> Any:
from langchain_community.vectorstores.surrealdb import SurrealDBStore
return SurrealDBStore
def _import_tair() -> Any:
from langchain_community.vectorstores.tair import Tair
return Tair
def _import_tencentvectordb() -> Any:
from langchain_community.vectorstores.tencentvectordb import TencentVectorDB
return TencentVectorDB
def _import_tidb_vectorstore() -> Any:
from langchain_community.vectorstores.tidb_vector import TiDBVectorStore
return TiDBVectorStore
def _import_tiledb() -> Any:
from langchain_community.vectorstores.tiledb import TileDB
return TileDB
def _import_tigris() -> Any:
from langchain_community.vectorstores.tigris import Tigris
return Tigris
def _import_timescalevector() -> Any:
from langchain_community.vectorstores.timescalevector import TimescaleVector
return TimescaleVector
def _import_typesense() -> Any:
from langchain_community.vectorstores.typesense import Typesense
return Typesense
def _import_usearch() -> Any:
from langchain_community.vectorstores.usearch import USearch
return USearch
def _import_vald() -> Any:
from langchain_community.vectorstores.vald import Vald
return Vald
def _import_vearch() -> Any:
from langchain_community.vectorstores.vearch import Vearch
return Vearch
def _import_vectara() -> Any:
from langchain_community.vectorstores.vectara import Vectara
return Vectara
def _import_weaviate() -> Any:
from langchain_community.vectorstores.weaviate import Weaviate
return Weaviate
def _import_yellowbrick() -> Any:
from langchain_community.vectorstores.yellowbrick import Yellowbrick
return Yellowbrick
def _import_zep() -> Any:
from langchain_community.vectorstores.zep import ZepVectorStore
return ZepVectorStore
def _import_zilliz() -> Any:
from langchain_community.vectorstores.zilliz import Zilliz
return Zilliz
def _import_neuraldb() -> Any:
from langchain_community.vectorstores.thirdai_neuraldb import NeuralDBVectorStore
return NeuralDBVectorStore
def _import_lantern() -> Any:
from langchain_community.vectorstores.lantern import Lantern
return Lantern
def __getattr__(name: str) -> Any: def __getattr__(name: str) -> Any:
if name == "AnalyticDB": if name in _module_lookup:
return _import_analyticdb() module = importlib.import_module(_module_lookup[name])
elif name == "AlibabaCloudOpenSearch": return getattr(module, name)
return _import_alibaba_cloud_open_search() raise AttributeError(f"module {__name__} has no attribute {name}")
elif name == "AlibabaCloudOpenSearchSettings":
return _import_alibaba_cloud_open_search_settings()
elif name == "AzureCosmosDBVectorSearch":
return _import_azure_cosmos_db()
elif name == "ElasticKnnSearch":
return _import_elastic_knn_search()
elif name == "ElasticVectorSearch":
return _import_elastic_vector_search()
elif name == "Annoy":
return _import_annoy()
elif name == "ApacheDoris":
return _import_apache_doris()
elif name == "AtlasDB":
return _import_atlas()
elif name == "AwaDB":
return _import_awadb()
elif name == "AzureSearch":
return _import_azuresearch()
elif name == "Bagel":
return _import_bageldb()
elif name == "BigQueryVectorSearch":
return _import_bigquery()
elif name == "BESVectorStore":
return _import_baiducloud_vector_search()
elif name == "Cassandra":
return _import_cassandra()
elif name == "AstraDB":
return _import_astradb()
elif name == "Chroma":
return _import_chroma()
elif name == "Clarifai":
return _import_clarifai()
elif name == "ClickhouseSettings":
return _import_clickhouse_settings()
elif name == "Clickhouse":
return _import_clickhouse()
elif name == "DashVector":
return _import_dashvector()
elif name == "DatabricksVectorSearch":
return _import_databricks_vector_search()
elif name == "DeepLake":
return _import_deeplake()
elif name == "Dingo":
return _import_dingo()
elif name == "DocArrayInMemorySearch":
return _import_docarray_inmemory()
elif name == "DocumentDBVectorSearch":
return _import_documentdb()
elif name == "DocArrayHnswSearch":
return _import_docarray_hnsw()
elif name == "ElasticsearchStore":
return _import_elasticsearch()
elif name == "Epsilla":
return _import_epsilla()
elif name == "FAISS":
return _import_faiss()
elif name == "HanaDB":
return _import_hanavector()
elif name == "Hologres":
return _import_hologres()
elif name == "InfinispanVS":
return _import_infinispanvs()
elif name == "KDBAI":
return _import_kdbai()
elif name == "DistanceStrategy":
return _import_distance_strategy()
elif name == "KineticaSettings":
return _import_kinetica_settings()
elif name == "Kinetica":
return _import_kinetica()
elif name == "LanceDB":
return _import_lancedb()
elif name == "LLMRails":
return _import_llm_rails()
elif name == "Marqo":
return _import_marqo()
elif name == "MatchingEngine":
return _import_matching_engine()
elif name == "Meilisearch":
return _import_meilisearch()
elif name == "Milvus":
return _import_milvus()
elif name == "MomentoVectorIndex":
return _import_momento_vector_index()
elif name == "MongoDBAtlasVectorSearch":
return _import_mongodb_atlas()
elif name == "MyScaleSettings":
return _import_myscale_settings()
elif name == "MyScale":
return _import_myscale()
elif name == "Neo4jVector":
return _import_neo4j_vector()
elif name == "OpenSearchVectorSearch":
return _import_opensearch_vector_search()
elif name == "PGEmbedding":
return _import_pgembedding()
elif name == "PGVector":
return _import_pgvector()
elif name == "Pinecone":
return _import_pinecone()
elif name == "Qdrant":
return _import_qdrant()
elif name == "Redis":
return _import_redis()
elif name == "Rockset":
return _import_rocksetdb()
elif name == "ScaNN":
return _import_scann()
elif name == "SemaDB":
return _import_semadb()
elif name == "SingleStoreDB":
return _import_singlestoredb()
elif name == "SKLearnVectorStore":
return _import_sklearn()
elif name == "SQLiteVSS":
return _import_sqlitevss()
elif name == "StarRocks":
return _import_starrocks()
elif name == "SupabaseVectorStore":
return _import_supabase()
elif name == "SurrealDBStore":
return _import_surrealdb()
elif name == "Tair":
return _import_tair()
elif name == "TencentVectorDB":
return _import_tencentvectordb()
elif name == "TiDBVectorStore":
return _import_tidb_vectorstore()
elif name == "TileDB":
return _import_tiledb()
elif name == "Tigris":
return _import_tigris()
elif name == "TimescaleVector":
return _import_timescalevector()
elif name == "Typesense":
return _import_typesense()
elif name == "USearch":
return _import_usearch()
elif name == "Vald":
return _import_vald()
elif name == "Vearch":
return _import_vearch()
elif name == "Vectara":
return _import_vectara()
elif name == "Weaviate":
return _import_weaviate()
elif name == "Yellowbrick":
return _import_yellowbrick()
elif name == "ZepVectorStore":
return _import_zep()
elif name == "Zilliz":
return _import_zilliz()
elif name == "VespaStore":
return _import_vespa()
elif name == "NeuralDBVectorStore":
return _import_neuraldb()
elif name == "Lantern":
return _import_lantern()
else:
raise AttributeError(f"Could not find: {name}")
__all__ = [ __all__ = list(_module_lookup.keys())
"AlibabaCloudOpenSearch",
"AlibabaCloudOpenSearchSettings",
"AnalyticDB",
"Annoy",
"ApacheDoris",
"AtlasDB",
"AwaDB",
"AzureSearch",
"Bagel",
"Cassandra",
"AstraDB",
"Chroma",
"Clarifai",
"Clickhouse",
"ClickhouseSettings",
"DashVector",
"DatabricksVectorSearch",
"DeepLake",
"Dingo",
"DocArrayHnswSearch",
"DocArrayInMemorySearch",
"ElasticKnnSearch",
"ElasticVectorSearch",
"ElasticsearchStore",
"Epsilla",
"FAISS",
"HanaDB",
"Hologres",
"InfinispanVS",
"KDBAI",
"DistanceStrategy",
"Kinetica",
"KineticaSettings",
"LanceDB",
"LLMRails",
"Marqo",
"MatchingEngine",
"Meilisearch",
"Milvus",
"MomentoVectorIndex",
"MongoDBAtlasVectorSearch",
"MyScale",
"MyScaleSettings",
"Neo4jVector",
"OpenSearchVectorSearch",
"PGEmbedding",
"PGVector",
"Pinecone",
"Qdrant",
"Redis",
"Rockset",
"SKLearnVectorStore",
"ScaNN",
"SemaDB",
"SingleStoreDB",
"SQLiteVSS",
"StarRocks",
"SupabaseVectorStore",
"SurrealDBStore",
"Tair",
"TiDBVectorStore",
"TileDB",
"Tigris",
"TimescaleVector",
"Typesense",
"USearch",
"Vald",
"Vearch",
"Vectara",
"VespaStore",
"Weaviate",
"Yellowbrick",
"ZepVectorStore",
"Zilliz",
"TencentVectorDB",
"AzureCosmosDBVectorSearch",
"VectorStore",
"NeuralDBVectorStore",
"Lantern",
"DocumentDBVectorSearch",
]

View File

@ -89,6 +89,7 @@ EXPECTED_ALL = [
"PolygonLastQuote", "PolygonLastQuote",
"PolygonTickerNews", "PolygonTickerNews",
"RedditSearchRun", "RedditSearchRun",
"RedditSearchSchema",
"QueryCheckerTool", "QueryCheckerTool",
"QueryPowerBITool", "QueryPowerBITool",
"QuerySQLCheckerTool", "QuerySQLCheckerTool",

View File

@ -91,6 +91,7 @@ _EXPECTED = [
"PolygonLastQuote", "PolygonLastQuote",
"PolygonTickerNews", "PolygonTickerNews",
"RedditSearchRun", "RedditSearchRun",
"RedditSearchSchema",
"QueryCheckerTool", "QueryCheckerTool",
"QueryPowerBITool", "QueryPowerBITool",
"QuerySQLCheckerTool", "QuerySQLCheckerTool",

View File

@ -24,6 +24,9 @@ EXPECTED_ALL = [
"MaxComputeAPIWrapper", "MaxComputeAPIWrapper",
"MetaphorSearchAPIWrapper", "MetaphorSearchAPIWrapper",
"NasaAPIWrapper", "NasaAPIWrapper",
"RivaASR",
"RivaTTS",
"AudioStream",
"NVIDIARivaASR", "NVIDIARivaASR",
"NVIDIARivaTTS", "NVIDIARivaTTS",
"NVIDIARivaStream", "NVIDIARivaStream",

View File

@ -51,6 +51,8 @@ def test_compatible_vectorstore_documentation() -> None:
"AzureCosmosDBVectorSearch", "AzureCosmosDBVectorSearch",
"AwaDB", "AwaDB",
"Bagel", "Bagel",
"BESVectorStore",
"BigQueryVectorSearch",
"Cassandra", "Cassandra",
"Chroma", "Chroma",
"DashVector", "DashVector",

View File

@ -11,6 +11,8 @@ _EXPECTED = [
"AwaDB", "AwaDB",
"AzureSearch", "AzureSearch",
"Bagel", "Bagel",
"BESVectorStore",
"BigQueryVectorSearch",
"Cassandra", "Cassandra",
"AstraDB", "AstraDB",
"Chroma", "Chroma",