mirror of
https://github.com/hwchase17/langchain.git
synced 2025-05-02 05:45:47 +00:00
community[patch]: deprecate langchain_community Chroma in favor of langchain_chroma (#24474)
This commit is contained in:
parent
0f7569ddbc
commit
dcba7df2fe
cookbook
Multi_modal_RAG.ipynbMulti_modal_RAG_google.ipynbRAPTOR.ipynbSemi_Structured_RAG.ipynbSemi_structured_and_multi_modal_RAG.ipynbSemi_structured_multi_modal_RAG_LLaMA2.ipynbadvanced_rag_eval.ipynbagent_vectorstore.ipynbairbyte_github.ipynbdocugami_xml_kg_rag.ipynbfireworks_rag.ipynbhypothetical_document_embeddings.ipynblanggraph_agentic_rag.ipynblanggraph_crag.ipynblanggraph_self_rag.ipynbmulti_modal_RAG_chroma.ipynbnomic_embedding_rag.ipynbnomic_multimodal_rag.ipynbopenai_functions_retrieval_qa.ipynboptimization.ipynbrag_with_quantized_embeddings.ipynbsql_db_qa.mdxtogether_ai.ipynb
docs/docs/integrations/document_loaders
libs
community
langchain_community/vectorstores
tests
langchain/langchain
templates
cohere-librarian
hyde
intel-rag-xeon
propositional-retrieval
rag-chroma-multi-modal-multi-vector
rag-chroma-multi-modal
rag-chroma-private
rag-chroma
rag-codellama-fireworks
rag-gemini-multi-modal
rag-gpt-crawler
rag-multi-modal-local
rag-multi-modal-mv-local
rag-ollama-multi-query
rag-semi-structured
@ -64,7 +64,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)"
|
||||
"! pip install -U langchain openai langchain-chroma langchain-experimental # (newest versions required for multi-modal)"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -355,7 +355,7 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
|
@ -37,7 +37,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"%pip install -U --quiet langchain langchain_community openai chromadb langchain-experimental\n",
|
||||
"%pip install -U --quiet langchain langchain-chroma langchain-community openai langchain-experimental\n",
|
||||
"%pip install --quiet \"unstructured[all-docs]\" pypdf pillow pydantic lxml pillow matplotlib chromadb tiktoken"
|
||||
]
|
||||
},
|
||||
@ -344,8 +344,8 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.embeddings import VertexAIEmbeddings\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"\n",
|
||||
"\n",
|
||||
|
@ -7,7 +7,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"pip install -U langchain umap-learn scikit-learn langchain_community tiktoken langchain-openai langchainhub chromadb langchain-anthropic"
|
||||
"pip install -U langchain umap-learn scikit-learn langchain_community tiktoken langchain-openai langchainhub langchain-chroma langchain-anthropic"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -645,7 +645,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"\n",
|
||||
"# Initialize all_texts with leaf_texts\n",
|
||||
"all_texts = leaf_texts.copy()\n",
|
||||
|
@ -39,7 +39,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain unstructured[all-docs] pydantic lxml langchainhub"
|
||||
"! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml langchainhub"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -320,7 +320,7 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
|
@ -59,7 +59,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain unstructured[all-docs] pydantic lxml"
|
||||
"! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -375,7 +375,7 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
|
@ -59,7 +59,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain unstructured[all-docs] pydantic lxml"
|
||||
"! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -378,8 +378,8 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.embeddings import GPT4AllEmbeddings\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"\n",
|
||||
"# The vectorstore to use to index the child chunks\n",
|
||||
|
@ -19,7 +19,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)"
|
||||
"! pip install -U langchain openai langchain_chroma langchain-experimental # (newest versions required for multi-modal)"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -132,7 +132,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"baseline = Chroma.from_texts(\n",
|
||||
|
@ -28,7 +28,7 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.chains import RetrievalQA\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAI, OpenAIEmbeddings\n",
|
||||
"from langchain_text_splitters import CharacterTextSplitter\n",
|
||||
"\n",
|
||||
|
@ -14,7 +14,7 @@
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"%pip install -qU langchain-airbyte"
|
||||
"%pip install -qU langchain-airbyte langchain_chroma"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -123,7 +123,7 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"import tiktoken\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"enc = tiktoken.get_encoding(\"cl100k_base\")\n",
|
||||
|
@ -39,7 +39,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain docugami==0.0.8 dgml-utils==0.3.0 pydantic langchainhub chromadb hnswlib --upgrade --quiet"
|
||||
"! pip install langchain docugami==0.0.8 dgml-utils==0.3.0 pydantic langchainhub langchain-chroma hnswlib --upgrade --quiet"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -547,7 +547,7 @@
|
||||
"\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_community.vectorstores.chroma import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.documents import Document\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
|
@ -84,7 +84,7 @@
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"%pip install --quiet pypdf chromadb tiktoken openai \n",
|
||||
"%pip install --quiet pypdf langchain-chroma tiktoken openai \n",
|
||||
"%pip uninstall -y langchain-fireworks\n",
|
||||
"%pip install --editable /mnt/disks/data/langchain/libs/partners/fireworks"
|
||||
]
|
||||
@ -138,7 +138,7 @@
|
||||
"all_splits = text_splitter.split_documents(data)\n",
|
||||
"\n",
|
||||
"# Add to vectorDB\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_fireworks.embeddings import FireworksEmbeddings\n",
|
||||
"\n",
|
||||
"vectorstore = Chroma.from_documents(\n",
|
||||
|
@ -170,7 +170,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_text_splitters import CharacterTextSplitter\n",
|
||||
"\n",
|
||||
"with open(\"../../state_of_the_union.txt\") as f:\n",
|
||||
|
@ -7,7 +7,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph"
|
||||
"! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -30,8 +30,8 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.document_loaders import WebBaseLoader\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"urls = [\n",
|
||||
|
@ -7,7 +7,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph tavily-python"
|
||||
"! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph tavily-python"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -77,8 +77,8 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.document_loaders import WebBaseLoader\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"urls = [\n",
|
||||
@ -180,8 +180,8 @@
|
||||
"from langchain.output_parsers.openai_tools import PydanticToolsParser\n",
|
||||
"from langchain.prompts import PromptTemplate\n",
|
||||
"from langchain.schema import Document\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.tools.tavily_search import TavilySearchResults\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_core.messages import BaseMessage, FunctionMessage\n",
|
||||
"from langchain_core.output_parsers import StrOutputParser\n",
|
||||
"from langchain_core.pydantic_v1 import BaseModel, Field\n",
|
||||
|
@ -7,7 +7,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph"
|
||||
"! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -86,8 +86,8 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.document_loaders import WebBaseLoader\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"urls = [\n",
|
||||
@ -188,7 +188,7 @@
|
||||
"from langchain.output_parsers import PydanticOutputParser\n",
|
||||
"from langchain.output_parsers.openai_tools import PydanticToolsParser\n",
|
||||
"from langchain.prompts import PromptTemplate\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.messages import BaseMessage, FunctionMessage\n",
|
||||
"from langchain_core.output_parsers import StrOutputParser\n",
|
||||
"from langchain_core.pydantic_v1 import BaseModel, Field\n",
|
||||
|
@ -58,7 +58,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)"
|
||||
"! pip install -U langchain openai langchain-chroma langchain-experimental # (newest versions required for multi-modal)"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -187,7 +187,7 @@
|
||||
"\n",
|
||||
"import chromadb\n",
|
||||
"import numpy as np\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_experimental.open_clip import OpenCLIPEmbeddings\n",
|
||||
"from PIL import Image as _PILImage\n",
|
||||
"\n",
|
||||
|
@ -58,7 +58,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install -U langchain-nomic langchain_community tiktoken langchain-openai chromadb langchain"
|
||||
"! pip install -U langchain-nomic langchain-chroma langchain-community tiktoken langchain-openai langchain"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -167,7 +167,7 @@
|
||||
"source": [
|
||||
"import os\n",
|
||||
"\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_core.output_parsers import StrOutputParser\n",
|
||||
"from langchain_core.runnables import RunnableLambda, RunnablePassthrough\n",
|
||||
"from langchain_nomic import NomicEmbeddings\n",
|
||||
|
@ -56,7 +56,7 @@
|
||||
},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install -U langchain-nomic langchain_community tiktoken langchain-openai chromadb langchain # (newest versions required for multi-modal)"
|
||||
"! pip install -U langchain-nomic langchain-chroma langchain-community tiktoken langchain-openai langchain # (newest versions required for multi-modal)"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -194,7 +194,7 @@
|
||||
"\n",
|
||||
"import chromadb\n",
|
||||
"import numpy as np\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_nomic import NomicEmbeddings\n",
|
||||
"from PIL import Image as _PILImage\n",
|
||||
"\n",
|
||||
|
@ -20,8 +20,8 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.chains import RetrievalQA\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.document_loaders import TextLoader\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"from langchain_text_splitters import CharacterTextSplitter"
|
||||
]
|
||||
|
@ -80,7 +80,7 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.schema import Document\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"embeddings = OpenAIEmbeddings()"
|
||||
|
@ -36,10 +36,10 @@
|
||||
"from bs4 import BeautifulSoup as Soup\n",
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
|
||||
"from langchain.storage import InMemoryByteStore, LocalFileStore\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.document_loaders.recursive_url_loader import (\n",
|
||||
" RecursiveUrlLoader,\n",
|
||||
")\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"\n",
|
||||
"# For our example, we'll load docs from the web\n",
|
||||
"from langchain_text_splitters import RecursiveCharacterTextSplitter\n",
|
||||
|
@ -740,7 +740,7 @@ Even this relatively large model will most likely fail to generate more complica
|
||||
|
||||
|
||||
```bash
|
||||
poetry run pip install pyyaml chromadb
|
||||
poetry run pip install pyyaml langchain_chroma
|
||||
import yaml
|
||||
```
|
||||
|
||||
@ -994,7 +994,7 @@ from langchain.prompts import FewShotPromptTemplate, PromptTemplate
|
||||
from langchain.chains.sql_database.prompt import _sqlite_prompt, PROMPT_SUFFIX
|
||||
from langchain_huggingface import HuggingFaceEmbeddings
|
||||
from langchain.prompts.example_selector.semantic_similarity import SemanticSimilarityExampleSelector
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
|
||||
example_prompt = PromptTemplate(
|
||||
input_variables=["table_info", "input", "sql_cmd", "sql_result", "answer"],
|
||||
|
@ -22,7 +22,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"! pip install --quiet pypdf chromadb tiktoken openai langchain-together"
|
||||
"! pip install --quiet pypdf tiktoken openai langchain-chroma langchain-together"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -45,8 +45,8 @@
|
||||
"all_splits = text_splitter.split_documents(data)\n",
|
||||
"\n",
|
||||
"# Add to vectorDB\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_community.embeddings import OpenAIEmbeddings\n",
|
||||
"from langchain_community.vectorstores import Chroma\n",
|
||||
"\n",
|
||||
"\"\"\"\n",
|
||||
"from langchain_together.embeddings import TogetherEmbeddings\n",
|
||||
|
@ -162,7 +162,7 @@
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"!poetry run pip install --upgrade langchain-openai tiktoken chromadb hnswlib"
|
||||
"!poetry run pip install --upgrade langchain-openai tiktoken langchain-chroma hnswlib"
|
||||
]
|
||||
},
|
||||
{
|
||||
@ -211,7 +211,7 @@
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"from langchain.chains import RetrievalQA\n",
|
||||
"from langchain_community.vectorstores.chroma import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAI, OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"embedding = OpenAIEmbeddings()\n",
|
||||
@ -365,7 +365,7 @@
|
||||
"source": [
|
||||
"from langchain.chains.query_constructor.schema import AttributeInfo\n",
|
||||
"from langchain.retrievers.self_query.base import SelfQueryRetriever\n",
|
||||
"from langchain_community.vectorstores.chroma import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"\n",
|
||||
"EXCLUDE_KEYS = [\"id\", \"xpath\", \"structure\"]\n",
|
||||
"metadata_field_info = [\n",
|
||||
@ -540,7 +540,7 @@
|
||||
"source": [
|
||||
"from langchain.retrievers.multi_vector import MultiVectorRetriever, SearchType\n",
|
||||
"from langchain.storage import InMemoryStore\n",
|
||||
"from langchain_community.vectorstores.chroma import Chroma\n",
|
||||
"from langchain_chroma import Chroma\n",
|
||||
"from langchain_openai import OpenAIEmbeddings\n",
|
||||
"\n",
|
||||
"# The vectorstore to use to index the child chunks\n",
|
||||
|
@ -50,6 +50,7 @@ def _results_to_docs_and_scores(results: Any) -> List[Tuple[Document, float]]:
|
||||
]
|
||||
|
||||
|
||||
@deprecated(since="0.2.9", removal="0.4", alternative_import="langchain_chroma.Chroma")
|
||||
class Chroma(VectorStore):
|
||||
"""`ChromaDB` vector store.
|
||||
|
||||
|
@ -1,10 +1,11 @@
|
||||
# flake8: noqa
|
||||
"""Test sentence_transformer embeddings."""
|
||||
|
||||
from langchain_core.vectorstores import InMemoryVectorStore
|
||||
|
||||
from langchain_community.embeddings.sentence_transformer import (
|
||||
SentenceTransformerEmbeddings,
|
||||
)
|
||||
from langchain_community.vectorstores import Chroma
|
||||
|
||||
|
||||
def test_sentence_transformer_embedding_documents() -> None:
|
||||
@ -34,7 +35,7 @@ def test_sentence_transformer_db_query() -> None:
|
||||
query = "what the foo is a bar?"
|
||||
query_vector = embedding.embed_query(query)
|
||||
assert len(query_vector) == 384
|
||||
db = Chroma(embedding_function=embedding)
|
||||
db = InMemoryVectorStore(embedding=embedding)
|
||||
db.add_texts(texts)
|
||||
docs = db.similarity_search_by_vector(query_vector, k=2)
|
||||
assert docs[0].page_content == "we will foo your bar until you can't foo any more"
|
||||
|
@ -1,7 +1,7 @@
|
||||
from langchain.retrievers.merger_retriever import MergerRetriever
|
||||
from langchain_core.vectorstores import InMemoryVectorStore
|
||||
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
|
||||
|
||||
def test_merger_retriever_get_relevant_docs() -> None:
|
||||
@ -17,12 +17,12 @@ def test_merger_retriever_get_relevant_docs() -> None:
|
||||
"Real stupidity beats artificial intelligence every time. TP",
|
||||
]
|
||||
embeddings = OpenAIEmbeddings()
|
||||
retriever_a = Chroma.from_texts(texts_group_a, embedding=embeddings).as_retriever(
|
||||
search_kwargs={"k": 1}
|
||||
)
|
||||
retriever_b = Chroma.from_texts(texts_group_b, embedding=embeddings).as_retriever(
|
||||
search_kwargs={"k": 1}
|
||||
)
|
||||
retriever_a = InMemoryVectorStore.from_texts(
|
||||
texts_group_a, embedding=embeddings
|
||||
).as_retriever(search_kwargs={"k": 1})
|
||||
retriever_b = InMemoryVectorStore.from_texts(
|
||||
texts_group_b, embedding=embeddings
|
||||
).as_retriever(search_kwargs={"k": 1})
|
||||
|
||||
# The Lord of the Retrievers.
|
||||
lotr = MergerRetriever(retrievers=[retriever_a, retriever_b])
|
||||
|
@ -1,10 +1,11 @@
|
||||
"""Integration test for doc reordering."""
|
||||
|
||||
from langchain_core.vectorstores import InMemoryVectorStore
|
||||
|
||||
from langchain_community.document_transformers.long_context_reorder import (
|
||||
LongContextReorder,
|
||||
)
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
|
||||
|
||||
def test_long_context_reorder() -> None:
|
||||
@ -22,9 +23,9 @@ def test_long_context_reorder() -> None:
|
||||
"Larry Bird was an iconic NBA player.",
|
||||
]
|
||||
embeddings = OpenAIEmbeddings()
|
||||
retriever = Chroma.from_texts(texts, embedding=embeddings).as_retriever(
|
||||
search_kwargs={"k": 10}
|
||||
)
|
||||
retriever = InMemoryVectorStore.from_texts(
|
||||
texts, embedding=embeddings
|
||||
).as_retriever(search_kwargs={"k": 10})
|
||||
reordering = LongContextReorder()
|
||||
docs = retriever.invoke("Tell me about the Celtics")
|
||||
actual = reordering.transform_documents(docs)
|
||||
|
@ -11,7 +11,11 @@ from langchain_core.callbacks import (
|
||||
CallbackManagerForRetrieverRun,
|
||||
)
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.vectorstores import VectorStore, VectorStoreRetriever
|
||||
from langchain_core.vectorstores import (
|
||||
InMemoryVectorStore,
|
||||
VectorStore,
|
||||
VectorStoreRetriever,
|
||||
)
|
||||
|
||||
from langchain_community.chains import PebbloRetrievalQA
|
||||
from langchain_community.chains.pebblo_retrieval.models import (
|
||||
@ -19,7 +23,6 @@ from langchain_community.chains.pebblo_retrieval.models import (
|
||||
ChainInput,
|
||||
SemanticContext,
|
||||
)
|
||||
from langchain_community.vectorstores.chroma import Chroma
|
||||
from langchain_community.vectorstores.pinecone import Pinecone
|
||||
from tests.unit_tests.llms.fake_llm import FakeLLM
|
||||
|
||||
@ -49,8 +52,8 @@ def unsupported_retriever() -> FakeRetriever:
|
||||
"""
|
||||
retriever = FakeRetriever()
|
||||
retriever.search_kwargs = {}
|
||||
# Set the class of vectorstore to Chroma
|
||||
retriever.vectorstore.__class__ = Chroma
|
||||
# Set the class of vectorstore
|
||||
retriever.vectorstore.__class__ = InMemoryVectorStore
|
||||
return retriever
|
||||
|
||||
|
||||
|
@ -67,7 +67,7 @@ class ConversationVectorStoreTokenBufferMemory(ConversationTokenBufferMemory):
|
||||
from langchain.memory.token_buffer_vectorstore_memory import (
|
||||
ConversationVectorStoreTokenBufferMemory
|
||||
)
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import HuggingFaceInstructEmbeddings
|
||||
from langchain_openai import OpenAI
|
||||
|
||||
|
@ -31,8 +31,8 @@ class ParentDocumentRetriever(MultiVectorRetriever):
|
||||
|
||||
.. code-block:: python
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_text_splitters import RecursiveCharacterTextSplitter
|
||||
from langchain.storage import InMemoryStore
|
||||
|
||||
|
@ -1,8 +1,8 @@
|
||||
import csv
|
||||
|
||||
from langchain.chains.question_answering import load_qa_chain
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import CohereEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.prompts import PromptTemplate
|
||||
|
||||
from .chat import chat
|
||||
|
@ -9,7 +9,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
cohere = "^4.37"
|
||||
chromadb = "^0.4.18"
|
||||
langchain-chroma = "^0.1.2"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
langchain-cli = ">=0.0.21"
|
||||
|
@ -1,6 +1,6 @@
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -9,7 +9,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
chromadb = "^0.4.15"
|
||||
langchain-chroma = "^0.1.2"
|
||||
tiktoken = "^0.5.1"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@ -1,9 +1,9 @@
|
||||
import os
|
||||
|
||||
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.document_loaders import UnstructuredFileLoader
|
||||
from langchain_community.embeddings import HuggingFaceEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
|
||||
|
||||
|
@ -1,7 +1,7 @@
|
||||
from langchain.callbacks import streaming_stdout
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import HuggingFaceEmbeddings
|
||||
from langchain_community.llms import HuggingFaceEndpoint
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -14,7 +14,7 @@ fastapi = "^0.104.0"
|
||||
sse-starlette = "^1.6.5"
|
||||
sentence-transformers = "2.2.2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
beautifulsoup4 = ">=4.12.2"
|
||||
|
||||
[tool.poetry.dependencies.unstructured]
|
||||
|
@ -3,8 +3,8 @@ from pathlib import Path
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
bs4 = "^0.0.1"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@ -8,9 +8,9 @@ from pathlib import Path
|
||||
import pypdfium2 as pdfium
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore, UpstashRedisByteStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from PIL import Image
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
upstash-redis = ">=1.0.0"
|
||||
|
@ -6,9 +6,9 @@ from pathlib import Path
|
||||
from langchain.pydantic_v1 import BaseModel
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore, UpstashRedisByteStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@ -2,7 +2,7 @@ import os
|
||||
from pathlib import Path
|
||||
|
||||
import pypdfium2 as pdfium
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_experimental.open_clip import OpenCLIPEmbeddings
|
||||
|
||||
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
|
@ -2,8 +2,8 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@ -11,7 +11,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
gpt4all = ">=1.0.8"
|
||||
beautifulsoup4 = ">=4.12.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
@ -1,8 +1,8 @@
|
||||
# Load
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.document_loaders import WebBaseLoader
|
||||
from langchain_community.embeddings import GPT4AllEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@ -1,6 +1,6 @@
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
|
||||
langchain = "^0.1"
|
||||
gpt4all = ">=1.0.8"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
fireworks-ai = ">=0.6.0"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@ -1,11 +1,11 @@
|
||||
import os
|
||||
|
||||
from git import Repo
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.document_loaders.generic import GenericLoader
|
||||
from langchain_community.document_loaders.parsers import LanguageParser
|
||||
from langchain_community.embeddings import GPT4AllEmbeddings
|
||||
from langchain_community.llms.fireworks import Fireworks
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -2,7 +2,7 @@ import os
|
||||
from pathlib import Path
|
||||
|
||||
import pypdfium2 as pdfium
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_experimental.open_clip import OpenCLIPEmbeddings
|
||||
|
||||
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
|
@ -2,7 +2,7 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@ -1,9 +1,9 @@
|
||||
import json
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
|
@ -1,7 +1,7 @@
|
||||
import os
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_nomic import NomicMultimodalEmbeddings
|
||||
|
||||
# Load images
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
|
@ -2,8 +2,8 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@ -7,9 +7,9 @@ from pathlib import Path
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.embeddings import OllamaEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from PIL import Image
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
pillow = ">=10.1.0"
|
||||
|
@ -5,9 +5,9 @@ from pathlib import Path
|
||||
from langchain.pydantic_v1 import BaseModel
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.embeddings import OllamaEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@ -1,8 +1,8 @@
|
||||
from langchain.retrievers.multi_query import MultiQueryRetriever
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama, ChatOpenAI
|
||||
from langchain_community.document_loaders import WebBaseLoader
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate, PromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@ -11,7 +11,7 @@ readme = "README.md"
|
||||
python = ">=3.9,<4"
|
||||
langchain = "^0.1"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
openai = "<2"
|
||||
unstructured = ">=0.10.19"
|
||||
pdf2image = ">=1.16.3"
|
||||
|
@ -3,9 +3,9 @@ import uuid
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import InMemoryStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
|
Loading…
Reference in New Issue
Block a user