From 20f5cd7c9583431689144ad17cd419d1a63b5ded Mon Sep 17 00:00:00 2001 From: Erick Friis Date: Fri, 12 Apr 2024 11:17:05 -0700 Subject: [PATCH] docs: langchain-chroma package (#20394) --- docs/docs/integrations/callbacks/confident.ipynb | 4 ++-- docs/docs/integrations/document_loaders/psychic.ipynb | 4 ++-- docs/docs/integrations/providers/chroma.mdx | 4 ++-- docs/docs/integrations/retrievers/merger_retriever.ipynb | 2 +- docs/docs/integrations/retrievers/re_phrase.ipynb | 2 +- .../retrievers/self_query/chroma_self_query.ipynb | 6 +++--- docs/docs/integrations/vectorstores/chroma.ipynb | 4 ++-- docs/docs/modules/agents/how_to/agent_structured.ipynb | 4 ++-- .../data_connection/retrievers/MultiQueryRetriever.ipynb | 2 +- .../data_connection/retrievers/long_context_reorder.ipynb | 4 ++-- .../modules/data_connection/retrievers/multi_vector.ipynb | 2 +- .../retrievers/parent_document_retriever.ipynb | 2 +- .../modules/data_connection/retrievers/self_query.ipynb | 4 ++-- docs/docs/modules/data_connection/vectorstores/index.mdx | 4 ++-- .../memory/adding_memory_chain_multiple_inputs.ipynb | 2 +- .../model_io/prompts/example_selectors/similarity.ipynb | 2 +- docs/docs/modules/model_io/prompts/few_shot_examples.ipynb | 2 +- .../modules/model_io/prompts/few_shot_examples_chat.ipynb | 2 +- docs/docs/use_cases/chatbots/quickstart.ipynb | 6 +++--- docs/docs/use_cases/chatbots/retrieval.ipynb | 4 ++-- docs/docs/use_cases/code_understanding.ipynb | 4 ++-- .../use_cases/query_analysis/how_to/high_cardinality.ipynb | 4 ++-- .../use_cases/query_analysis/how_to/multiple_queries.ipynb | 4 ++-- .../query_analysis/how_to/multiple_retrievers.ipynb | 4 ++-- docs/docs/use_cases/query_analysis/how_to/no_queries.ipynb | 4 ++-- docs/docs/use_cases/query_analysis/quickstart.ipynb | 4 ++-- docs/docs/use_cases/question_answering/chat_history.ipynb | 6 +++--- .../use_cases/question_answering/local_retrieval_qa.ipynb | 4 ++-- docs/docs/use_cases/question_answering/quickstart.mdx | 6 +++--- docs/docs/use_cases/question_answering/sources.ipynb | 4 ++-- docs/docs/use_cases/question_answering/streaming.ipynb | 4 ++-- docs/docs/use_cases/web_scraping.ipynb | 2 +- 32 files changed, 58 insertions(+), 58 deletions(-) diff --git a/docs/docs/integrations/callbacks/confident.ipynb b/docs/docs/integrations/callbacks/confident.ipynb index 0ff4b0e04f5..a5110206291 100644 --- a/docs/docs/integrations/callbacks/confident.ipynb +++ b/docs/docs/integrations/callbacks/confident.ipynb @@ -42,7 +42,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain langchain-openai deepeval" + "%pip install --upgrade --quiet langchain langchain-openai deepeval langchain-chroma" ] }, { @@ -215,8 +215,8 @@ "source": [ "import requests\n", "from langchain.chains import RetrievalQA\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import TextLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAI, OpenAIEmbeddings\n", "from langchain_text_splitters import CharacterTextSplitter\n", "\n", diff --git a/docs/docs/integrations/document_loaders/psychic.ipynb b/docs/docs/integrations/document_loaders/psychic.ipynb index 30a91493454..fec1921cb07 100644 --- a/docs/docs/integrations/document_loaders/psychic.ipynb +++ b/docs/docs/integrations/document_loaders/psychic.ipynb @@ -39,7 +39,7 @@ ], "source": [ "# Uncomment this to install psychicapi if you don't already have it installed\n", - "!poetry run pip -q install psychicapi" + "!poetry run pip -q install psychicapi langchain-chroma" ] }, { @@ -78,7 +78,7 @@ "outputs": [], "source": [ "from langchain.chains import RetrievalQAWithSourcesChain\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAI, OpenAIEmbeddings\n", "from langchain_text_splitters import CharacterTextSplitter" ] diff --git a/docs/docs/integrations/providers/chroma.mdx b/docs/docs/integrations/providers/chroma.mdx index ab7af6029bd..d5436c9dc2a 100644 --- a/docs/docs/integrations/providers/chroma.mdx +++ b/docs/docs/integrations/providers/chroma.mdx @@ -5,7 +5,7 @@ ## Installation and Setup ```bash -pip install chromadb +pip install langchain-chroma ``` @@ -15,7 +15,7 @@ There exists a wrapper around Chroma vector databases, allowing you to use it as whether for semantic search or example selection. ```python -from langchain_community.vectorstores import Chroma +from langchain_chroma import Chroma ``` For a more detailed walkthrough of the Chroma wrapper, see [this notebook](/docs/integrations/vectorstores/chroma) diff --git a/docs/docs/integrations/retrievers/merger_retriever.ipynb b/docs/docs/integrations/retrievers/merger_retriever.ipynb index 22d29481769..cc6dc2cb45b 100644 --- a/docs/docs/integrations/retrievers/merger_retriever.ipynb +++ b/docs/docs/integrations/retrievers/merger_retriever.ipynb @@ -28,12 +28,12 @@ " DocumentCompressorPipeline,\n", " MergerRetriever,\n", ")\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_transformers import (\n", " EmbeddingsClusteringFilter,\n", " EmbeddingsRedundantFilter,\n", ")\n", "from langchain_community.embeddings import HuggingFaceEmbeddings\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "# Get 3 diff embeddings.\n", diff --git a/docs/docs/integrations/retrievers/re_phrase.ipynb b/docs/docs/integrations/retrievers/re_phrase.ipynb index c99be7db796..dd4dcd5f7fe 100644 --- a/docs/docs/integrations/retrievers/re_phrase.ipynb +++ b/docs/docs/integrations/retrievers/re_phrase.ipynb @@ -28,8 +28,8 @@ "import logging\n", "\n", "from langchain.retrievers import RePhraseQueryRetriever\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter" ] diff --git a/docs/docs/integrations/retrievers/self_query/chroma_self_query.ipynb b/docs/docs/integrations/retrievers/self_query/chroma_self_query.ipynb index 8e316f146d9..e04509af54b 100644 --- a/docs/docs/integrations/retrievers/self_query/chroma_self_query.ipynb +++ b/docs/docs/integrations/retrievers/self_query/chroma_self_query.ipynb @@ -20,7 +20,7 @@ "## Creating a Chroma vector store\n", "First we'll want to create a Chroma vector store and seed it with some data. We've created a small demo set of documents that contain summaries of movies.\n", "\n", - "**Note:** The self-query retriever requires you to have `lark` installed (`pip install lark`). We also need the `chromadb` package." + "**Note:** The self-query retriever requires you to have `lark` installed (`pip install lark`). We also need the `langchain-chroma` package." ] }, { @@ -44,7 +44,7 @@ }, "outputs": [], "source": [ - "%pip install --upgrade --quiet chromadb" + "%pip install --upgrade --quiet langchain-chroma" ] }, { @@ -87,7 +87,7 @@ }, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_core.documents import Document\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", diff --git a/docs/docs/integrations/vectorstores/chroma.ipynb b/docs/docs/integrations/vectorstores/chroma.ipynb index 1b43327c15e..773db6c9dce 100644 --- a/docs/docs/integrations/vectorstores/chroma.ipynb +++ b/docs/docs/integrations/vectorstores/chroma.ipynb @@ -13,7 +13,7 @@ "Install Chroma with:\n", "\n", "```sh\n", - "pip install chromadb\n", + "pip install langchain-chroma\n", "```\n", "\n", "Chroma runs in various modes. See below for examples of each integrated with LangChain.\n", @@ -65,11 +65,11 @@ ], "source": [ "# import\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.embeddings.sentence_transformer import (\n", " SentenceTransformerEmbeddings,\n", ")\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_text_splitters import CharacterTextSplitter\n", "\n", "# load the document and split it into chunks\n", diff --git a/docs/docs/modules/agents/how_to/agent_structured.ipynb b/docs/docs/modules/agents/how_to/agent_structured.ipynb index 0c1d550bbbf..b7b38d24e52 100644 --- a/docs/docs/modules/agents/how_to/agent_structured.ipynb +++ b/docs/docs/modules/agents/how_to/agent_structured.ipynb @@ -43,7 +43,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install -qU chromadb langchain langchain-community langchain-openai" + "%pip install -qU langchain langchain-community langchain-openai langchain-chroma" ] }, { @@ -53,8 +53,8 @@ "metadata": {}, "outputs": [], "source": [ + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import TextLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter" ] diff --git a/docs/docs/modules/data_connection/retrievers/MultiQueryRetriever.ipynb b/docs/docs/modules/data_connection/retrievers/MultiQueryRetriever.ipynb index 1b38be518e0..4f5d543cef6 100644 --- a/docs/docs/modules/data_connection/retrievers/MultiQueryRetriever.ipynb +++ b/docs/docs/modules/data_connection/retrievers/MultiQueryRetriever.ipynb @@ -20,8 +20,8 @@ "outputs": [], "source": [ "# Build a sample vectorDB\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter\n", "\n", diff --git a/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb b/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb index b95f8882023..f52015fb780 100644 --- a/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb +++ b/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb @@ -21,7 +21,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet sentence-transformers > /dev/null" + "%pip install --upgrade --quiet sentence-transformers langchain-chroma langchain langchain-openai > /dev/null" ] }, { @@ -53,11 +53,11 @@ "source": [ "from langchain.chains import LLMChain, StuffDocumentsChain\n", "from langchain.prompts import PromptTemplate\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_transformers import (\n", " LongContextReorder,\n", ")\n", "from langchain_community.embeddings import HuggingFaceEmbeddings\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAI\n", "\n", "# Get embeddings.\n", diff --git a/docs/docs/modules/data_connection/retrievers/multi_vector.ipynb b/docs/docs/modules/data_connection/retrievers/multi_vector.ipynb index 5952202d054..9e42a8f1e07 100644 --- a/docs/docs/modules/data_connection/retrievers/multi_vector.ipynb +++ b/docs/docs/modules/data_connection/retrievers/multi_vector.ipynb @@ -37,8 +37,8 @@ "outputs": [], "source": [ "from langchain.storage import InMemoryByteStore\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import TextLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter" ] diff --git a/docs/docs/modules/data_connection/retrievers/parent_document_retriever.ipynb b/docs/docs/modules/data_connection/retrievers/parent_document_retriever.ipynb index 7fde529167e..1653e3f558c 100644 --- a/docs/docs/modules/data_connection/retrievers/parent_document_retriever.ipynb +++ b/docs/docs/modules/data_connection/retrievers/parent_document_retriever.ipynb @@ -43,8 +43,8 @@ "outputs": [], "source": [ "from langchain.storage import InMemoryStore\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import TextLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter" ] diff --git a/docs/docs/modules/data_connection/retrievers/self_query.ipynb b/docs/docs/modules/data_connection/retrievers/self_query.ipynb index 973a55ad699..5584fae50b4 100644 --- a/docs/docs/modules/data_connection/retrievers/self_query.ipynb +++ b/docs/docs/modules/data_connection/retrievers/self_query.ipynb @@ -30,7 +30,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet lark chromadb" + "%pip install --upgrade --quiet lark langchain-chroma" ] }, { @@ -40,7 +40,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_core.documents import Document\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", diff --git a/docs/docs/modules/data_connection/vectorstores/index.mdx b/docs/docs/modules/data_connection/vectorstores/index.mdx index 532a26fd4e3..060df47026d 100644 --- a/docs/docs/modules/data_connection/vectorstores/index.mdx +++ b/docs/docs/modules/data_connection/vectorstores/index.mdx @@ -30,7 +30,7 @@ There are many great vector store options, here are a few that are free, open-so This walkthrough uses the `chroma` vector database, which runs on your local machine as a library. ```bash -pip install chromadb +pip install langchain-chroma ``` We want to use OpenAIEmbeddings so we have to get the OpenAI API Key. @@ -47,7 +47,7 @@ os.environ['OPENAI_API_KEY'] = getpass.getpass('OpenAI API Key:') from langchain_community.document_loaders import TextLoader from langchain_openai import OpenAIEmbeddings from langchain_text_splitters import CharacterTextSplitter -from langchain_community.vectorstores import Chroma +from langchain_chroma import Chroma # Load the document, split it into chunks, embed each chunk and load it into the vector store. raw_documents = TextLoader('../../../state_of_the_union.txt').load() diff --git a/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb b/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb index ddc4b5c4b8f..1f25eca6628 100644 --- a/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb +++ b/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb @@ -17,7 +17,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import CharacterTextSplitter" ] diff --git a/docs/docs/modules/model_io/prompts/example_selectors/similarity.ipynb b/docs/docs/modules/model_io/prompts/example_selectors/similarity.ipynb index 0d9be3b77b6..733e2ad74b4 100644 --- a/docs/docs/modules/model_io/prompts/example_selectors/similarity.ipynb +++ b/docs/docs/modules/model_io/prompts/example_selectors/similarity.ipynb @@ -19,7 +19,7 @@ "source": [ "from langchain.prompts import FewShotPromptTemplate, PromptTemplate\n", "from langchain.prompts.example_selector import SemanticSimilarityExampleSelector\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "example_prompt = PromptTemplate(\n", diff --git a/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb b/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb index 8b603315b3a..04a567ae9fa 100644 --- a/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb +++ b/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb @@ -254,7 +254,7 @@ ], "source": [ "from langchain.prompts.example_selector import SemanticSimilarityExampleSelector\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "example_selector = SemanticSimilarityExampleSelector.from_examples(\n", diff --git a/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb b/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb index 5cb2e99c91f..52cb0ea0b5f 100644 --- a/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb +++ b/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb @@ -202,7 +202,7 @@ "outputs": [], "source": [ "from langchain.prompts import SemanticSimilarityExampleSelector\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings" ] }, diff --git a/docs/docs/use_cases/chatbots/quickstart.ipynb b/docs/docs/use_cases/chatbots/quickstart.ipynb index 2875a094e24..f48f4d00775 100644 --- a/docs/docs/use_cases/chatbots/quickstart.ipynb +++ b/docs/docs/use_cases/chatbots/quickstart.ipynb @@ -64,7 +64,7 @@ } ], "source": [ - "%pip install --upgrade --quiet langchain langchain-openai\n", + "%pip install --upgrade --quiet langchain langchain-openai langchain-chroma\n", "\n", "# Set env var OPENAI_API_KEY or load from a .env file:\n", "import dotenv\n", @@ -391,7 +391,7 @@ } ], "source": [ - "%pip install --upgrade --quiet chromadb beautifulsoup4" + "%pip install --upgrade --quiet langchain-chroma beautifulsoup4" ] }, { @@ -445,7 +445,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "vectorstore = Chroma.from_documents(documents=all_splits, embedding=OpenAIEmbeddings())" diff --git a/docs/docs/use_cases/chatbots/retrieval.ipynb b/docs/docs/use_cases/chatbots/retrieval.ipynb index 726beafb0bc..dd474418afb 100644 --- a/docs/docs/use_cases/chatbots/retrieval.ipynb +++ b/docs/docs/use_cases/chatbots/retrieval.ipynb @@ -48,7 +48,7 @@ } ], "source": [ - "%pip install --upgrade --quiet langchain langchain-openai chromadb beautifulsoup4\n", + "%pip install --upgrade --quiet langchain langchain-openai langchain-chroma beautifulsoup4\n", "\n", "# Set env var OPENAI_API_KEY or load from a .env file:\n", "import dotenv\n", @@ -129,7 +129,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "vectorstore = Chroma.from_documents(documents=all_splits, embedding=OpenAIEmbeddings())" diff --git a/docs/docs/use_cases/code_understanding.ipynb b/docs/docs/use_cases/code_understanding.ipynb index 15ffdb29fbe..fa0668833d3 100644 --- a/docs/docs/use_cases/code_understanding.ipynb +++ b/docs/docs/use_cases/code_understanding.ipynb @@ -45,7 +45,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain-openai tiktoken chromadb langchain git\n", + "%pip install --upgrade --quiet langchain-openai tiktoken langchain-chroma langchain git\n", "\n", "# Set env var OPENAI_API_KEY or load from a .env file\n", "# import dotenv\n", @@ -201,7 +201,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "db = Chroma.from_documents(texts, OpenAIEmbeddings(disallowed_special=()))\n", diff --git a/docs/docs/use_cases/query_analysis/how_to/high_cardinality.ipynb b/docs/docs/use_cases/query_analysis/how_to/high_cardinality.ipynb index 6cd1ef21aca..67572e2ec35 100644 --- a/docs/docs/use_cases/query_analysis/how_to/high_cardinality.ipynb +++ b/docs/docs/use_cases/query_analysis/how_to/high_cardinality.ipynb @@ -38,7 +38,7 @@ "metadata": {}, "outputs": [], "source": [ - "# %pip install -qU langchain langchain-community langchain-openai faker" + "# %pip install -qU langchain langchain-community langchain-openai faker langchain-chroma" ] }, { @@ -394,7 +394,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "embeddings = OpenAIEmbeddings(model=\"text-embedding-3-small\")\n", diff --git a/docs/docs/use_cases/query_analysis/how_to/multiple_queries.ipynb b/docs/docs/use_cases/query_analysis/how_to/multiple_queries.ipynb index 2fe92b552a5..866a7aa8879 100644 --- a/docs/docs/use_cases/query_analysis/how_to/multiple_queries.ipynb +++ b/docs/docs/use_cases/query_analysis/how_to/multiple_queries.ipynb @@ -36,7 +36,7 @@ "metadata": {}, "outputs": [], "source": [ - "# %pip install -qU langchain langchain-community langchain-openai chromadb" + "# %pip install -qU langchain langchain-community langchain-openai langchain-chroma" ] }, { @@ -84,7 +84,7 @@ "outputs": [], "source": [ "from langchain.text_splitter import RecursiveCharacterTextSplitter\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "texts = [\"Harrison worked at Kensho\", \"Ankush worked at Facebook\"]\n", diff --git a/docs/docs/use_cases/query_analysis/how_to/multiple_retrievers.ipynb b/docs/docs/use_cases/query_analysis/how_to/multiple_retrievers.ipynb index d668515097f..7d3683e026f 100644 --- a/docs/docs/use_cases/query_analysis/how_to/multiple_retrievers.ipynb +++ b/docs/docs/use_cases/query_analysis/how_to/multiple_retrievers.ipynb @@ -36,7 +36,7 @@ "metadata": {}, "outputs": [], "source": [ - "# %pip install -qU langchain langchain-community langchain-openai chromadb" + "# %pip install -qU langchain langchain-community langchain-openai langchain-chroma" ] }, { @@ -84,7 +84,7 @@ "outputs": [], "source": [ "from langchain.text_splitter import RecursiveCharacterTextSplitter\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "texts = [\"Harrison worked at Kensho\"]\n", diff --git a/docs/docs/use_cases/query_analysis/how_to/no_queries.ipynb b/docs/docs/use_cases/query_analysis/how_to/no_queries.ipynb index 943bf3f73cb..4668eca73d0 100644 --- a/docs/docs/use_cases/query_analysis/how_to/no_queries.ipynb +++ b/docs/docs/use_cases/query_analysis/how_to/no_queries.ipynb @@ -38,7 +38,7 @@ "metadata": {}, "outputs": [], "source": [ - "# %pip install -qU langchain langchain-community langchain-openai chromadb" + "# %pip install -qU langchain langchain-community langchain-openai langchain-chroma" ] }, { @@ -86,7 +86,7 @@ "outputs": [], "source": [ "from langchain.text_splitter import RecursiveCharacterTextSplitter\n", - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "texts = [\"Harrison worked at Kensho\"]\n", diff --git a/docs/docs/use_cases/query_analysis/quickstart.ipynb b/docs/docs/use_cases/query_analysis/quickstart.ipynb index f5d383e1c14..2d627ceb4f1 100644 --- a/docs/docs/use_cases/query_analysis/quickstart.ipynb +++ b/docs/docs/use_cases/query_analysis/quickstart.ipynb @@ -38,7 +38,7 @@ "metadata": {}, "outputs": [], "source": [ - "# %pip install -qU langchain langchain-community langchain-openai youtube-transcript-api pytube chromadb" + "# %pip install -qU langchain langchain-community langchain-openai youtube-transcript-api pytube langchain-chroma" ] }, { @@ -249,7 +249,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain_community.vectorstores import Chroma\n", + "from langchain_chroma import Chroma\n", "from langchain_openai import OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter\n", "\n", diff --git a/docs/docs/use_cases/question_answering/chat_history.ipynb b/docs/docs/use_cases/question_answering/chat_history.ipynb index ecd093e523c..17b46504f27 100644 --- a/docs/docs/use_cases/question_answering/chat_history.ipynb +++ b/docs/docs/use_cases/question_answering/chat_history.ipynb @@ -48,7 +48,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai chromadb bs4" + "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai langchain-chroma bs4" ] }, { @@ -118,8 +118,8 @@ "source": [ "import bs4\n", "from langchain import hub\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.runnables import RunnablePassthrough\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", @@ -406,9 +406,9 @@ "from langchain import hub\n", "from langchain.chains import create_history_aware_retriever, create_retrieval_chain\n", "from langchain.chains.combine_documents import create_stuff_documents_chain\n", + "from langchain_chroma import Chroma\n", "from langchain_community.chat_message_histories import ChatMessageHistory\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_core.chat_history import BaseChatMessageHistory\n", "from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.prompts import ChatPromptTemplate, MessagesPlaceholder\n", diff --git a/docs/docs/use_cases/question_answering/local_retrieval_qa.ipynb b/docs/docs/use_cases/question_answering/local_retrieval_qa.ipynb index aff5c342c1f..d1e67ddfbbd 100644 --- a/docs/docs/use_cases/question_answering/local_retrieval_qa.ipynb +++ b/docs/docs/use_cases/question_answering/local_retrieval_qa.ipynb @@ -27,7 +27,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain langchain-community langchainhub gpt4all chromadb " + "%pip install --upgrade --quiet langchain langchain-community langchainhub gpt4all langchain-chroma " ] }, { @@ -72,8 +72,8 @@ "metadata": {}, "outputs": [], "source": [ + "from langchain_chroma import Chroma\n", "from langchain_community.embeddings import GPT4AllEmbeddings\n", - "from langchain_community.vectorstores import Chroma\n", "\n", "vectorstore = Chroma.from_documents(documents=all_splits, embedding=GPT4AllEmbeddings())" ] diff --git a/docs/docs/use_cases/question_answering/quickstart.mdx b/docs/docs/use_cases/question_answering/quickstart.mdx index 8ec6ac9068c..fd360cef701 100644 --- a/docs/docs/use_cases/question_answering/quickstart.mdx +++ b/docs/docs/use_cases/question_answering/quickstart.mdx @@ -72,7 +72,7 @@ in this walkthrough, but everything shown here works with any We’ll use the following packages: ```python -%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai chromadb bs4 +%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai langchain-chroma bs4 ``` We need to set environment variable `OPENAI_API_KEY` for the embeddings model, which can be done @@ -120,7 +120,7 @@ lines of code: import bs4 from langchain import hub from langchain_community.document_loaders import WebBaseLoader -from langchain_community.vectorstores import Chroma +from langchain_chroma import Chroma from langchain_core.output_parsers import StrOutputParser from langchain_core.runnables import RunnablePassthrough from langchain_openai import OpenAIEmbeddings @@ -350,7 +350,7 @@ vector store and model. ```python -from langchain_community.vectorstores import Chroma +from langchain_chroma import Chroma from langchain_openai import OpenAIEmbeddings vectorstore = Chroma.from_documents(documents=all_splits, embedding=OpenAIEmbeddings()) diff --git a/docs/docs/use_cases/question_answering/sources.ipynb b/docs/docs/use_cases/question_answering/sources.ipynb index 858181bd949..0bbe0759aed 100644 --- a/docs/docs/use_cases/question_answering/sources.ipynb +++ b/docs/docs/use_cases/question_answering/sources.ipynb @@ -43,7 +43,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai chromadb bs4" + "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai langchain-chroma bs4" ] }, { @@ -113,8 +113,8 @@ "source": [ "import bs4\n", "from langchain import hub\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.runnables import RunnablePassthrough\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", diff --git a/docs/docs/use_cases/question_answering/streaming.ipynb b/docs/docs/use_cases/question_answering/streaming.ipynb index 5867e75d537..975316bdc59 100644 --- a/docs/docs/use_cases/question_answering/streaming.ipynb +++ b/docs/docs/use_cases/question_answering/streaming.ipynb @@ -43,7 +43,7 @@ "metadata": {}, "outputs": [], "source": [ - "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai chromadb bs4" + "%pip install --upgrade --quiet langchain langchain-community langchainhub langchain-openai langchain-chroma bs4" ] }, { @@ -113,8 +113,8 @@ "source": [ "import bs4\n", "from langchain import hub\n", + "from langchain_chroma import Chroma\n", "from langchain_community.document_loaders import WebBaseLoader\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.runnables import RunnableParallel, RunnablePassthrough\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", diff --git a/docs/docs/use_cases/web_scraping.ipynb b/docs/docs/use_cases/web_scraping.ipynb index f62aaeea1f8..9735afa3dba 100644 --- a/docs/docs/use_cases/web_scraping.ipynb +++ b/docs/docs/use_cases/web_scraping.ipynb @@ -480,8 +480,8 @@ "outputs": [], "source": [ "from langchain.retrievers.web_research import WebResearchRetriever\n", + "from langchain_chroma import Chroma\n", "from langchain_community.utilities import GoogleSearchAPIWrapper\n", - "from langchain_community.vectorstores import Chroma\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings" ] },