community[patch]: deprecate langchain_community Chroma in favor of langchain_chroma (#24474)

This commit is contained in:
ccurme 2024-07-22 11:00:13 -04:00 committed by GitHub
parent 0f7569ddbc
commit dcba7df2fe
No known key found for this signature in database
GPG Key ID: B5690EEEBB952194
67 changed files with 108 additions and 102 deletions

View File

@ -64,7 +64,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)" "! pip install -U langchain openai langchain-chroma langchain-experimental # (newest versions required for multi-modal)"
] ]
}, },
{ {
@ -355,7 +355,7 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",

View File

@ -37,7 +37,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"%pip install -U --quiet langchain langchain_community openai chromadb langchain-experimental\n", "%pip install -U --quiet langchain langchain-chroma langchain-community openai langchain-experimental\n",
"%pip install --quiet \"unstructured[all-docs]\" pypdf pillow pydantic lxml pillow matplotlib chromadb tiktoken" "%pip install --quiet \"unstructured[all-docs]\" pypdf pillow pydantic lxml pillow matplotlib chromadb tiktoken"
] ]
}, },
@ -344,8 +344,8 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.embeddings import VertexAIEmbeddings\n", "from langchain_community.embeddings import VertexAIEmbeddings\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"\n", "\n",
"\n", "\n",

View File

@ -7,7 +7,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"pip install -U langchain umap-learn scikit-learn langchain_community tiktoken langchain-openai langchainhub chromadb langchain-anthropic" "pip install -U langchain umap-learn scikit-learn langchain_community tiktoken langchain-openai langchainhub langchain-chroma langchain-anthropic"
] ]
}, },
{ {
@ -645,7 +645,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"\n", "\n",
"# Initialize all_texts with leaf_texts\n", "# Initialize all_texts with leaf_texts\n",
"all_texts = leaf_texts.copy()\n", "all_texts = leaf_texts.copy()\n",

View File

@ -39,7 +39,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain unstructured[all-docs] pydantic lxml langchainhub" "! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml langchainhub"
] ]
}, },
{ {
@ -320,7 +320,7 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",

View File

@ -59,7 +59,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain unstructured[all-docs] pydantic lxml" "! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml"
] ]
}, },
{ {
@ -375,7 +375,7 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",

View File

@ -59,7 +59,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain unstructured[all-docs] pydantic lxml" "! pip install langchain langchain-chroma unstructured[all-docs] pydantic lxml"
] ]
}, },
{ {
@ -378,8 +378,8 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.embeddings import GPT4AllEmbeddings\n", "from langchain_community.embeddings import GPT4AllEmbeddings\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"\n", "\n",
"# The vectorstore to use to index the child chunks\n", "# The vectorstore to use to index the child chunks\n",

View File

@ -19,7 +19,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)" "! pip install -U langchain openai langchain_chroma langchain-experimental # (newest versions required for multi-modal)"
] ]
}, },
{ {
@ -132,7 +132,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"baseline = Chroma.from_texts(\n", "baseline = Chroma.from_texts(\n",

View File

@ -28,7 +28,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.chains import RetrievalQA\n", "from langchain.chains import RetrievalQA\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAI, OpenAIEmbeddings\n", "from langchain_openai import OpenAI, OpenAIEmbeddings\n",
"from langchain_text_splitters import CharacterTextSplitter\n", "from langchain_text_splitters import CharacterTextSplitter\n",
"\n", "\n",

View File

@ -14,7 +14,7 @@
} }
], ],
"source": [ "source": [
"%pip install -qU langchain-airbyte" "%pip install -qU langchain-airbyte langchain_chroma"
] ]
}, },
{ {
@ -123,7 +123,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"import tiktoken\n", "import tiktoken\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"enc = tiktoken.get_encoding(\"cl100k_base\")\n", "enc = tiktoken.get_encoding(\"cl100k_base\")\n",

View File

@ -39,7 +39,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain docugami==0.0.8 dgml-utils==0.3.0 pydantic langchainhub chromadb hnswlib --upgrade --quiet" "! pip install langchain docugami==0.0.8 dgml-utils==0.3.0 pydantic langchainhub langchain-chroma hnswlib --upgrade --quiet"
] ]
}, },
{ {
@ -547,7 +547,7 @@
"\n", "\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_community.vectorstores.chroma import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.documents import Document\n", "from langchain_core.documents import Document\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",

View File

@ -84,7 +84,7 @@
} }
], ],
"source": [ "source": [
"%pip install --quiet pypdf chromadb tiktoken openai \n", "%pip install --quiet pypdf langchain-chroma tiktoken openai \n",
"%pip uninstall -y langchain-fireworks\n", "%pip uninstall -y langchain-fireworks\n",
"%pip install --editable /mnt/disks/data/langchain/libs/partners/fireworks" "%pip install --editable /mnt/disks/data/langchain/libs/partners/fireworks"
] ]
@ -138,7 +138,7 @@
"all_splits = text_splitter.split_documents(data)\n", "all_splits = text_splitter.split_documents(data)\n",
"\n", "\n",
"# Add to vectorDB\n", "# Add to vectorDB\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_fireworks.embeddings import FireworksEmbeddings\n", "from langchain_fireworks.embeddings import FireworksEmbeddings\n",
"\n", "\n",
"vectorstore = Chroma.from_documents(\n", "vectorstore = Chroma.from_documents(\n",

View File

@ -170,7 +170,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_text_splitters import CharacterTextSplitter\n", "from langchain_text_splitters import CharacterTextSplitter\n",
"\n", "\n",
"with open(\"../../state_of_the_union.txt\") as f:\n", "with open(\"../../state_of_the_union.txt\") as f:\n",

View File

@ -7,7 +7,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph" "! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph"
] ]
}, },
{ {
@ -30,8 +30,8 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n", "from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.document_loaders import WebBaseLoader\n", "from langchain_community.document_loaders import WebBaseLoader\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"urls = [\n", "urls = [\n",

View File

@ -7,7 +7,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph tavily-python" "! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph tavily-python"
] ]
}, },
{ {
@ -77,8 +77,8 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n", "from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.document_loaders import WebBaseLoader\n", "from langchain_community.document_loaders import WebBaseLoader\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"urls = [\n", "urls = [\n",
@ -180,8 +180,8 @@
"from langchain.output_parsers.openai_tools import PydanticToolsParser\n", "from langchain.output_parsers.openai_tools import PydanticToolsParser\n",
"from langchain.prompts import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"from langchain.schema import Document\n", "from langchain.schema import Document\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.tools.tavily_search import TavilySearchResults\n", "from langchain_community.tools.tavily_search import TavilySearchResults\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_core.messages import BaseMessage, FunctionMessage\n", "from langchain_core.messages import BaseMessage, FunctionMessage\n",
"from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.output_parsers import StrOutputParser\n",
"from langchain_core.pydantic_v1 import BaseModel, Field\n", "from langchain_core.pydantic_v1 import BaseModel, Field\n",

View File

@ -7,7 +7,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install langchain_community tiktoken langchain-openai langchainhub chromadb langchain langgraph" "! pip install langchain-chroma langchain_community tiktoken langchain-openai langchainhub langchain langgraph"
] ]
}, },
{ {
@ -86,8 +86,8 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n", "from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.document_loaders import WebBaseLoader\n", "from langchain_community.document_loaders import WebBaseLoader\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"urls = [\n", "urls = [\n",
@ -188,7 +188,7 @@
"from langchain.output_parsers import PydanticOutputParser\n", "from langchain.output_parsers import PydanticOutputParser\n",
"from langchain.output_parsers.openai_tools import PydanticToolsParser\n", "from langchain.output_parsers.openai_tools import PydanticToolsParser\n",
"from langchain.prompts import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.messages import BaseMessage, FunctionMessage\n", "from langchain_core.messages import BaseMessage, FunctionMessage\n",
"from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.output_parsers import StrOutputParser\n",
"from langchain_core.pydantic_v1 import BaseModel, Field\n", "from langchain_core.pydantic_v1 import BaseModel, Field\n",

View File

@ -58,7 +58,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install -U langchain openai chromadb langchain-experimental # (newest versions required for multi-modal)" "! pip install -U langchain openai langchain-chroma langchain-experimental # (newest versions required for multi-modal)"
] ]
}, },
{ {
@ -187,7 +187,7 @@
"\n", "\n",
"import chromadb\n", "import chromadb\n",
"import numpy as np\n", "import numpy as np\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_experimental.open_clip import OpenCLIPEmbeddings\n", "from langchain_experimental.open_clip import OpenCLIPEmbeddings\n",
"from PIL import Image as _PILImage\n", "from PIL import Image as _PILImage\n",
"\n", "\n",

View File

@ -58,7 +58,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install -U langchain-nomic langchain_community tiktoken langchain-openai chromadb langchain" "! pip install -U langchain-nomic langchain-chroma langchain-community tiktoken langchain-openai langchain"
] ]
}, },
{ {
@ -167,7 +167,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"\n", "\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_core.output_parsers import StrOutputParser\n", "from langchain_core.output_parsers import StrOutputParser\n",
"from langchain_core.runnables import RunnableLambda, RunnablePassthrough\n", "from langchain_core.runnables import RunnableLambda, RunnablePassthrough\n",
"from langchain_nomic import NomicEmbeddings\n", "from langchain_nomic import NomicEmbeddings\n",

View File

@ -56,7 +56,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install -U langchain-nomic langchain_community tiktoken langchain-openai chromadb langchain # (newest versions required for multi-modal)" "! pip install -U langchain-nomic langchain-chroma langchain-community tiktoken langchain-openai langchain # (newest versions required for multi-modal)"
] ]
}, },
{ {
@ -194,7 +194,7 @@
"\n", "\n",
"import chromadb\n", "import chromadb\n",
"import numpy as np\n", "import numpy as np\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_nomic import NomicEmbeddings\n", "from langchain_nomic import NomicEmbeddings\n",
"from PIL import Image as _PILImage\n", "from PIL import Image as _PILImage\n",
"\n", "\n",

View File

@ -20,8 +20,8 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.chains import RetrievalQA\n", "from langchain.chains import RetrievalQA\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.document_loaders import TextLoader\n", "from langchain_community.document_loaders import TextLoader\n",
"from langchain_community.vectorstores import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"from langchain_text_splitters import CharacterTextSplitter" "from langchain_text_splitters import CharacterTextSplitter"
] ]

View File

@ -80,7 +80,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.schema import Document\n", "from langchain.schema import Document\n",
"from langchain_community.vectorstores import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"embeddings = OpenAIEmbeddings()" "embeddings = OpenAIEmbeddings()"

View File

@ -36,10 +36,10 @@
"from bs4 import BeautifulSoup as Soup\n", "from bs4 import BeautifulSoup as Soup\n",
"from langchain.retrievers.multi_vector import MultiVectorRetriever\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever\n",
"from langchain.storage import InMemoryByteStore, LocalFileStore\n", "from langchain.storage import InMemoryByteStore, LocalFileStore\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.document_loaders.recursive_url_loader import (\n", "from langchain_community.document_loaders.recursive_url_loader import (\n",
" RecursiveUrlLoader,\n", " RecursiveUrlLoader,\n",
")\n", ")\n",
"from langchain_community.vectorstores import Chroma\n",
"\n", "\n",
"# For our example, we'll load docs from the web\n", "# For our example, we'll load docs from the web\n",
"from langchain_text_splitters import RecursiveCharacterTextSplitter\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter\n",

View File

@ -740,7 +740,7 @@ Even this relatively large model will most likely fail to generate more complica
```bash ```bash
poetry run pip install pyyaml chromadb poetry run pip install pyyaml langchain_chroma
import yaml import yaml
``` ```
@ -994,7 +994,7 @@ from langchain.prompts import FewShotPromptTemplate, PromptTemplate
from langchain.chains.sql_database.prompt import _sqlite_prompt, PROMPT_SUFFIX from langchain.chains.sql_database.prompt import _sqlite_prompt, PROMPT_SUFFIX
from langchain_huggingface import HuggingFaceEmbeddings from langchain_huggingface import HuggingFaceEmbeddings
from langchain.prompts.example_selector.semantic_similarity import SemanticSimilarityExampleSelector from langchain.prompts.example_selector.semantic_similarity import SemanticSimilarityExampleSelector
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
example_prompt = PromptTemplate( example_prompt = PromptTemplate(
input_variables=["table_info", "input", "sql_cmd", "sql_result", "answer"], input_variables=["table_info", "input", "sql_cmd", "sql_result", "answer"],

View File

@ -22,7 +22,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"! pip install --quiet pypdf chromadb tiktoken openai langchain-together" "! pip install --quiet pypdf tiktoken openai langchain-chroma langchain-together"
] ]
}, },
{ {
@ -45,8 +45,8 @@
"all_splits = text_splitter.split_documents(data)\n", "all_splits = text_splitter.split_documents(data)\n",
"\n", "\n",
"# Add to vectorDB\n", "# Add to vectorDB\n",
"from langchain_chroma import Chroma\n",
"from langchain_community.embeddings import OpenAIEmbeddings\n", "from langchain_community.embeddings import OpenAIEmbeddings\n",
"from langchain_community.vectorstores import Chroma\n",
"\n", "\n",
"\"\"\"\n", "\"\"\"\n",
"from langchain_together.embeddings import TogetherEmbeddings\n", "from langchain_together.embeddings import TogetherEmbeddings\n",

View File

@ -162,7 +162,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"!poetry run pip install --upgrade langchain-openai tiktoken chromadb hnswlib" "!poetry run pip install --upgrade langchain-openai tiktoken langchain-chroma hnswlib"
] ]
}, },
{ {
@ -211,7 +211,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.chains import RetrievalQA\n", "from langchain.chains import RetrievalQA\n",
"from langchain_community.vectorstores.chroma import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAI, OpenAIEmbeddings\n", "from langchain_openai import OpenAI, OpenAIEmbeddings\n",
"\n", "\n",
"embedding = OpenAIEmbeddings()\n", "embedding = OpenAIEmbeddings()\n",
@ -365,7 +365,7 @@
"source": [ "source": [
"from langchain.chains.query_constructor.schema import AttributeInfo\n", "from langchain.chains.query_constructor.schema import AttributeInfo\n",
"from langchain.retrievers.self_query.base import SelfQueryRetriever\n", "from langchain.retrievers.self_query.base import SelfQueryRetriever\n",
"from langchain_community.vectorstores.chroma import Chroma\n", "from langchain_chroma import Chroma\n",
"\n", "\n",
"EXCLUDE_KEYS = [\"id\", \"xpath\", \"structure\"]\n", "EXCLUDE_KEYS = [\"id\", \"xpath\", \"structure\"]\n",
"metadata_field_info = [\n", "metadata_field_info = [\n",
@ -540,7 +540,7 @@
"source": [ "source": [
"from langchain.retrievers.multi_vector import MultiVectorRetriever, SearchType\n", "from langchain.retrievers.multi_vector import MultiVectorRetriever, SearchType\n",
"from langchain.storage import InMemoryStore\n", "from langchain.storage import InMemoryStore\n",
"from langchain_community.vectorstores.chroma import Chroma\n", "from langchain_chroma import Chroma\n",
"from langchain_openai import OpenAIEmbeddings\n", "from langchain_openai import OpenAIEmbeddings\n",
"\n", "\n",
"# The vectorstore to use to index the child chunks\n", "# The vectorstore to use to index the child chunks\n",

View File

@ -50,6 +50,7 @@ def _results_to_docs_and_scores(results: Any) -> List[Tuple[Document, float]]:
] ]
@deprecated(since="0.2.9", removal="0.4", alternative_import="langchain_chroma.Chroma")
class Chroma(VectorStore): class Chroma(VectorStore):
"""`ChromaDB` vector store. """`ChromaDB` vector store.

View File

@ -1,10 +1,11 @@
# flake8: noqa # flake8: noqa
"""Test sentence_transformer embeddings.""" """Test sentence_transformer embeddings."""
from langchain_core.vectorstores import InMemoryVectorStore
from langchain_community.embeddings.sentence_transformer import ( from langchain_community.embeddings.sentence_transformer import (
SentenceTransformerEmbeddings, SentenceTransformerEmbeddings,
) )
from langchain_community.vectorstores import Chroma
def test_sentence_transformer_embedding_documents() -> None: def test_sentence_transformer_embedding_documents() -> None:
@ -34,7 +35,7 @@ def test_sentence_transformer_db_query() -> None:
query = "what the foo is a bar?" query = "what the foo is a bar?"
query_vector = embedding.embed_query(query) query_vector = embedding.embed_query(query)
assert len(query_vector) == 384 assert len(query_vector) == 384
db = Chroma(embedding_function=embedding) db = InMemoryVectorStore(embedding=embedding)
db.add_texts(texts) db.add_texts(texts)
docs = db.similarity_search_by_vector(query_vector, k=2) docs = db.similarity_search_by_vector(query_vector, k=2)
assert docs[0].page_content == "we will foo your bar until you can't foo any more" assert docs[0].page_content == "we will foo your bar until you can't foo any more"

View File

@ -1,7 +1,7 @@
from langchain.retrievers.merger_retriever import MergerRetriever from langchain.retrievers.merger_retriever import MergerRetriever
from langchain_core.vectorstores import InMemoryVectorStore
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
def test_merger_retriever_get_relevant_docs() -> None: def test_merger_retriever_get_relevant_docs() -> None:
@ -17,12 +17,12 @@ def test_merger_retriever_get_relevant_docs() -> None:
"Real stupidity beats artificial intelligence every time. TP", "Real stupidity beats artificial intelligence every time. TP",
] ]
embeddings = OpenAIEmbeddings() embeddings = OpenAIEmbeddings()
retriever_a = Chroma.from_texts(texts_group_a, embedding=embeddings).as_retriever( retriever_a = InMemoryVectorStore.from_texts(
search_kwargs={"k": 1} texts_group_a, embedding=embeddings
) ).as_retriever(search_kwargs={"k": 1})
retriever_b = Chroma.from_texts(texts_group_b, embedding=embeddings).as_retriever( retriever_b = InMemoryVectorStore.from_texts(
search_kwargs={"k": 1} texts_group_b, embedding=embeddings
) ).as_retriever(search_kwargs={"k": 1})
# The Lord of the Retrievers. # The Lord of the Retrievers.
lotr = MergerRetriever(retrievers=[retriever_a, retriever_b]) lotr = MergerRetriever(retrievers=[retriever_a, retriever_b])

View File

@ -1,10 +1,11 @@
"""Integration test for doc reordering.""" """Integration test for doc reordering."""
from langchain_core.vectorstores import InMemoryVectorStore
from langchain_community.document_transformers.long_context_reorder import ( from langchain_community.document_transformers.long_context_reorder import (
LongContextReorder, LongContextReorder,
) )
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
def test_long_context_reorder() -> None: def test_long_context_reorder() -> None:
@ -22,9 +23,9 @@ def test_long_context_reorder() -> None:
"Larry Bird was an iconic NBA player.", "Larry Bird was an iconic NBA player.",
] ]
embeddings = OpenAIEmbeddings() embeddings = OpenAIEmbeddings()
retriever = Chroma.from_texts(texts, embedding=embeddings).as_retriever( retriever = InMemoryVectorStore.from_texts(
search_kwargs={"k": 10} texts, embedding=embeddings
) ).as_retriever(search_kwargs={"k": 10})
reordering = LongContextReorder() reordering = LongContextReorder()
docs = retriever.invoke("Tell me about the Celtics") docs = retriever.invoke("Tell me about the Celtics")
actual = reordering.transform_documents(docs) actual = reordering.transform_documents(docs)

View File

@ -11,7 +11,11 @@ from langchain_core.callbacks import (
CallbackManagerForRetrieverRun, CallbackManagerForRetrieverRun,
) )
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.vectorstores import VectorStore, VectorStoreRetriever from langchain_core.vectorstores import (
InMemoryVectorStore,
VectorStore,
VectorStoreRetriever,
)
from langchain_community.chains import PebbloRetrievalQA from langchain_community.chains import PebbloRetrievalQA
from langchain_community.chains.pebblo_retrieval.models import ( from langchain_community.chains.pebblo_retrieval.models import (
@ -19,7 +23,6 @@ from langchain_community.chains.pebblo_retrieval.models import (
ChainInput, ChainInput,
SemanticContext, SemanticContext,
) )
from langchain_community.vectorstores.chroma import Chroma
from langchain_community.vectorstores.pinecone import Pinecone from langchain_community.vectorstores.pinecone import Pinecone
from tests.unit_tests.llms.fake_llm import FakeLLM from tests.unit_tests.llms.fake_llm import FakeLLM
@ -49,8 +52,8 @@ def unsupported_retriever() -> FakeRetriever:
""" """
retriever = FakeRetriever() retriever = FakeRetriever()
retriever.search_kwargs = {} retriever.search_kwargs = {}
# Set the class of vectorstore to Chroma # Set the class of vectorstore
retriever.vectorstore.__class__ = Chroma retriever.vectorstore.__class__ = InMemoryVectorStore
return retriever return retriever

View File

@ -67,7 +67,7 @@ class ConversationVectorStoreTokenBufferMemory(ConversationTokenBufferMemory):
from langchain.memory.token_buffer_vectorstore_memory import ( from langchain.memory.token_buffer_vectorstore_memory import (
ConversationVectorStoreTokenBufferMemory ConversationVectorStoreTokenBufferMemory
) )
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
from langchain_community.embeddings import HuggingFaceInstructEmbeddings from langchain_community.embeddings import HuggingFaceInstructEmbeddings
from langchain_openai import OpenAI from langchain_openai import OpenAI

View File

@ -31,8 +31,8 @@ class ParentDocumentRetriever(MultiVectorRetriever):
.. code-block:: python .. code-block:: python
from langchain_chroma import Chroma
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_text_splitters import RecursiveCharacterTextSplitter from langchain_text_splitters import RecursiveCharacterTextSplitter
from langchain.storage import InMemoryStore from langchain.storage import InMemoryStore

View File

@ -1,8 +1,8 @@
import csv import csv
from langchain.chains.question_answering import load_qa_chain from langchain.chains.question_answering import load_qa_chain
from langchain_chroma import Chroma
from langchain_community.embeddings import CohereEmbeddings from langchain_community.embeddings import CohereEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.prompts import PromptTemplate from langchain_core.prompts import PromptTemplate
from .chat import chat from .chat import chat

View File

@ -9,7 +9,7 @@ readme = "README.md"
python = ">=3.8.1,<4.0" python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
cohere = "^4.37" cohere = "^4.37"
chromadb = "^0.4.18" langchain-chroma = "^0.1.2"
[tool.poetry.group.dev.dependencies] [tool.poetry.group.dev.dependencies]
langchain-cli = ">=0.0.21" langchain-cli = ">=0.0.21"

View File

@ -1,6 +1,6 @@
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -9,7 +9,7 @@ readme = "README.md"
python = ">=3.8.1,<4.0" python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
openai = "<2" openai = "<2"
chromadb = "^0.4.15" langchain-chroma = "^0.1.2"
tiktoken = "^0.5.1" tiktoken = "^0.5.1"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"

View File

@ -1,9 +1,9 @@
import os import os
from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain_chroma import Chroma
from langchain_community.document_loaders import UnstructuredFileLoader from langchain_community.document_loaders import UnstructuredFileLoader
from langchain_community.embeddings import HuggingFaceEmbeddings from langchain_community.embeddings import HuggingFaceEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document

View File

@ -1,7 +1,7 @@
from langchain.callbacks import streaming_stdout from langchain.callbacks import streaming_stdout
from langchain_chroma import Chroma
from langchain_community.embeddings import HuggingFaceEmbeddings from langchain_community.embeddings import HuggingFaceEmbeddings
from langchain_community.llms import HuggingFaceEndpoint from langchain_community.llms import HuggingFaceEndpoint
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -14,7 +14,7 @@ fastapi = "^0.104.0"
sse-starlette = "^1.6.5" sse-starlette = "^1.6.5"
sentence-transformers = "2.2.2" sentence-transformers = "2.2.2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
beautifulsoup4 = ">=4.12.2" beautifulsoup4 = ">=4.12.2"
[tool.poetry.dependencies.unstructured] [tool.poetry.dependencies.unstructured]

View File

@ -3,8 +3,8 @@ from pathlib import Path
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore from langchain.storage import LocalFileStore
from langchain_chroma import Chroma
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
logging.basicConfig(level=logging.INFO) logging.basicConfig(level=logging.INFO)

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
bs4 = "^0.0.1" bs4 = "^0.0.1"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"

View File

@ -8,9 +8,9 @@ from pathlib import Path
import pypdfium2 as pdfium import pypdfium2 as pdfium
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore, UpstashRedisByteStore from langchain.storage import LocalFileStore, UpstashRedisByteStore
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from PIL import Image from PIL import Image

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = ">=0.0.353,<0.2" langchain = ">=0.0.353,<0.2"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
pypdfium2 = ">=4.20.0" pypdfium2 = ">=4.20.0"
langchain-experimental = ">=0.0.43" langchain-experimental = ">=0.0.43"
upstash-redis = ">=1.0.0" upstash-redis = ">=1.0.0"

View File

@ -6,9 +6,9 @@ from pathlib import Path
from langchain.pydantic_v1 import BaseModel from langchain.pydantic_v1 import BaseModel
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore, UpstashRedisByteStore from langchain.storage import LocalFileStore, UpstashRedisByteStore
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser

View File

@ -2,7 +2,7 @@ import os
from pathlib import Path from pathlib import Path
import pypdfium2 as pdfium import pypdfium2 as pdfium
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
from langchain_experimental.open_clip import OpenCLIPEmbeddings from langchain_experimental.open_clip import OpenCLIPEmbeddings

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = ">=0.0.353,<0.2" langchain = ">=0.0.353,<0.2"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
open-clip-torch = ">=2.23.0" open-clip-torch = ">=2.23.0"
torch = ">=2.1.0" torch = ">=2.1.0"
pypdfium2 = ">=4.20.0" pypdfium2 = ">=4.20.0"

View File

@ -2,8 +2,8 @@ import base64
import io import io
from pathlib import Path from pathlib import Path
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser

View File

@ -11,7 +11,7 @@ readme = "README.md"
python = ">=3.8.1,<4.0" python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
gpt4all = ">=1.0.8" gpt4all = ">=1.0.8"
beautifulsoup4 = ">=4.12.2" beautifulsoup4 = ">=4.12.2"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"

View File

@ -1,8 +1,8 @@
# Load # Load
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOllama from langchain_community.chat_models import ChatOllama
from langchain_community.document_loaders import WebBaseLoader from langchain_community.document_loaders import WebBaseLoader
from langchain_community.embeddings import GPT4AllEmbeddings from langchain_community.embeddings import GPT4AllEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"
[tool.poetry.group.dev.dependencies] [tool.poetry.group.dev.dependencies]

View File

@ -1,6 +1,6 @@
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
langchain = "^0.1" langchain = "^0.1"
gpt4all = ">=1.0.8" gpt4all = ">=1.0.8"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
fireworks-ai = ">=0.6.0" fireworks-ai = ">=0.6.0"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"

View File

@ -1,11 +1,11 @@
import os import os
from git import Repo from git import Repo
from langchain_chroma import Chroma
from langchain_community.document_loaders.generic import GenericLoader from langchain_community.document_loaders.generic import GenericLoader
from langchain_community.document_loaders.parsers import LanguageParser from langchain_community.document_loaders.parsers import LanguageParser
from langchain_community.embeddings import GPT4AllEmbeddings from langchain_community.embeddings import GPT4AllEmbeddings
from langchain_community.llms.fireworks import Fireworks from langchain_community.llms.fireworks import Fireworks
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -2,7 +2,7 @@ import os
from pathlib import Path from pathlib import Path
import pypdfium2 as pdfium import pypdfium2 as pdfium
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
from langchain_experimental.open_clip import OpenCLIPEmbeddings from langchain_experimental.open_clip import OpenCLIPEmbeddings

View File

@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
langchain = ">=0.0.353,<0.2" langchain = ">=0.0.353,<0.2"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
open-clip-torch = ">=2.23.0" open-clip-torch = ">=2.23.0"
torch = ">=2.1.0" torch = ">=2.1.0"
pypdfium2 = ">=4.20.0" pypdfium2 = ">=4.20.0"

View File

@ -2,7 +2,7 @@ import base64
import io import io
from pathlib import Path from pathlib import Path
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"
[tool.poetry.group.dev.dependencies] [tool.poetry.group.dev.dependencies]

View File

@ -1,9 +1,9 @@
import json import json
from pathlib import Path from pathlib import Path
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate

View File

@ -1,7 +1,7 @@
import os import os
from pathlib import Path from pathlib import Path
from langchain_community.vectorstores import Chroma from langchain_chroma import Chroma
from langchain_nomic import NomicMultimodalEmbeddings from langchain_nomic import NomicMultimodalEmbeddings
# Load images # Load images

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = ">=0.0.353,<0.2" langchain = ">=0.0.353,<0.2"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
open-clip-torch = ">=2.23.0" open-clip-torch = ">=2.23.0"
torch = ">=2.1.0" torch = ">=2.1.0"
langchain-experimental = ">=0.0.43" langchain-experimental = ">=0.0.43"

View File

@ -2,8 +2,8 @@ import base64
import io import io
from pathlib import Path from pathlib import Path
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOllama from langchain_community.chat_models import ChatOllama
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser

View File

@ -7,9 +7,9 @@ from pathlib import Path
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore from langchain.storage import LocalFileStore
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOllama from langchain_community.chat_models import ChatOllama
from langchain_community.embeddings import OllamaEmbeddings from langchain_community.embeddings import OllamaEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from PIL import Image from PIL import Image

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = ">=0.0.353,<0.2" langchain = ">=0.0.353,<0.2"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
pypdfium2 = ">=4.20.0" pypdfium2 = ">=4.20.0"
langchain-experimental = ">=0.0.43" langchain-experimental = ">=0.0.43"
pillow = ">=10.1.0" pillow = ">=10.1.0"

View File

@ -5,9 +5,9 @@ from pathlib import Path
from langchain.pydantic_v1 import BaseModel from langchain.pydantic_v1 import BaseModel
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import LocalFileStore from langchain.storage import LocalFileStore
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOllama from langchain_community.chat_models import ChatOllama
from langchain_community.embeddings import OllamaEmbeddings from langchain_community.embeddings import OllamaEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.messages import HumanMessage from langchain_core.messages import HumanMessage
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser

View File

@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
langchain = "^0.1" langchain = "^0.1"
openai = "<2" openai = "<2"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
langchain-text-splitters = ">=0.0.1,<0.1" langchain-text-splitters = ">=0.0.1,<0.1"
[tool.poetry.group.dev.dependencies] [tool.poetry.group.dev.dependencies]

View File

@ -1,8 +1,8 @@
from langchain.retrievers.multi_query import MultiQueryRetriever from langchain.retrievers.multi_query import MultiQueryRetriever
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOllama, ChatOpenAI from langchain_community.chat_models import ChatOllama, ChatOpenAI
from langchain_community.document_loaders import WebBaseLoader from langchain_community.document_loaders import WebBaseLoader
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate, PromptTemplate from langchain_core.prompts import ChatPromptTemplate, PromptTemplate
from langchain_core.pydantic_v1 import BaseModel from langchain_core.pydantic_v1 import BaseModel

View File

@ -11,7 +11,7 @@ readme = "README.md"
python = ">=3.9,<4" python = ">=3.9,<4"
langchain = "^0.1" langchain = "^0.1"
tiktoken = ">=0.5.1" tiktoken = ">=0.5.1"
chromadb = ">=0.4.14" langchain-chroma = "^0.1.2"
openai = "<2" openai = "<2"
unstructured = ">=0.10.19" unstructured = ">=0.10.19"
pdf2image = ">=1.16.3" pdf2image = ">=1.16.3"

View File

@ -3,9 +3,9 @@ import uuid
from langchain.retrievers.multi_vector import MultiVectorRetriever from langchain.retrievers.multi_vector import MultiVectorRetriever
from langchain.storage import InMemoryStore from langchain.storage import InMemoryStore
from langchain_chroma import Chroma
from langchain_community.chat_models import ChatOpenAI from langchain_community.chat_models import ChatOpenAI
from langchain_community.embeddings import OpenAIEmbeddings from langchain_community.embeddings import OpenAIEmbeddings
from langchain_community.vectorstores import Chroma
from langchain_core.documents import Document from langchain_core.documents import Document
from langchain_core.output_parsers import StrOutputParser from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate from langchain_core.prompts import ChatPromptTemplate