mirror of
https://github.com/hwchase17/langchain.git
synced 2025-09-04 20:46:45 +00:00
community[patch]: deprecate langchain_community Chroma in favor of langchain_chroma (#24474)
This commit is contained in:
@@ -1,8 +1,8 @@
|
||||
import csv
|
||||
|
||||
from langchain.chains.question_answering import load_qa_chain
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import CohereEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.prompts import PromptTemplate
|
||||
|
||||
from .chat import chat
|
||||
|
@@ -9,7 +9,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
cohere = "^4.37"
|
||||
chromadb = "^0.4.18"
|
||||
langchain-chroma = "^0.1.2"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
langchain-cli = ">=0.0.21"
|
||||
|
@@ -1,6 +1,6 @@
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -9,7 +9,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
chromadb = "^0.4.15"
|
||||
langchain-chroma = "^0.1.2"
|
||||
tiktoken = "^0.5.1"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@@ -1,9 +1,9 @@
|
||||
import os
|
||||
|
||||
from langchain.text_splitter import RecursiveCharacterTextSplitter
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.document_loaders import UnstructuredFileLoader
|
||||
from langchain_community.embeddings import HuggingFaceEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
|
||||
|
||||
|
@@ -1,7 +1,7 @@
|
||||
from langchain.callbacks import streaming_stdout
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import HuggingFaceEmbeddings
|
||||
from langchain_community.llms import HuggingFaceEndpoint
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -14,7 +14,7 @@ fastapi = "^0.104.0"
|
||||
sse-starlette = "^1.6.5"
|
||||
sentence-transformers = "2.2.2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
beautifulsoup4 = ">=4.12.2"
|
||||
|
||||
[tool.poetry.dependencies.unstructured]
|
||||
|
@@ -3,8 +3,8 @@ from pathlib import Path
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
|
||||
logging.basicConfig(level=logging.INFO)
|
||||
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
bs4 = "^0.0.1"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@@ -8,9 +8,9 @@ from pathlib import Path
|
||||
import pypdfium2 as pdfium
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore, UpstashRedisByteStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from PIL import Image
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
upstash-redis = ">=1.0.0"
|
||||
|
@@ -6,9 +6,9 @@ from pathlib import Path
|
||||
from langchain.pydantic_v1 import BaseModel
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore, UpstashRedisByteStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@@ -2,7 +2,7 @@ import os
|
||||
from pathlib import Path
|
||||
|
||||
import pypdfium2 as pdfium
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_experimental.open_clip import OpenCLIPEmbeddings
|
||||
|
||||
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
|
@@ -2,8 +2,8 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@@ -11,7 +11,7 @@ readme = "README.md"
|
||||
python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
gpt4all = ">=1.0.8"
|
||||
beautifulsoup4 = ">=4.12.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
@@ -1,8 +1,8 @@
|
||||
# Load
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.document_loaders import WebBaseLoader
|
||||
from langchain_community.embeddings import GPT4AllEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@@ -1,6 +1,6 @@
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
|
||||
langchain = "^0.1"
|
||||
gpt4all = ">=1.0.8"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
fireworks-ai = ">=0.6.0"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
|
@@ -1,11 +1,11 @@
|
||||
import os
|
||||
|
||||
from git import Repo
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.document_loaders.generic import GenericLoader
|
||||
from langchain_community.document_loaders.parsers import LanguageParser
|
||||
from langchain_community.embeddings import GPT4AllEmbeddings
|
||||
from langchain_community.llms.fireworks import Fireworks
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -2,7 +2,7 @@ import os
|
||||
from pathlib import Path
|
||||
|
||||
import pypdfium2 as pdfium
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_experimental.open_clip import OpenCLIPEmbeddings
|
||||
|
||||
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.9,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
|
@@ -2,7 +2,7 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@@ -1,9 +1,9 @@
|
||||
import json
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
|
@@ -1,7 +1,7 @@
|
||||
import os
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_nomic import NomicMultimodalEmbeddings
|
||||
|
||||
# Load images
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
open-clip-torch = ">=2.23.0"
|
||||
torch = ">=2.1.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
|
@@ -2,8 +2,8 @@ import base64
|
||||
import io
|
||||
from pathlib import Path
|
||||
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@@ -7,9 +7,9 @@ from pathlib import Path
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.embeddings import OllamaEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from PIL import Image
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = ">=0.0.353,<0.2"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
pypdfium2 = ">=4.20.0"
|
||||
langchain-experimental = ">=0.0.43"
|
||||
pillow = ">=10.1.0"
|
||||
|
@@ -5,9 +5,9 @@ from pathlib import Path
|
||||
from langchain.pydantic_v1 import BaseModel
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import LocalFileStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama
|
||||
from langchain_community.embeddings import OllamaEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.messages import HumanMessage
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
|
@@ -12,7 +12,7 @@ python = ">=3.8.1,<4.0"
|
||||
langchain = "^0.1"
|
||||
openai = "<2"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
langchain-text-splitters = ">=0.0.1,<0.1"
|
||||
|
||||
[tool.poetry.group.dev.dependencies]
|
||||
|
@@ -1,8 +1,8 @@
|
||||
from langchain.retrievers.multi_query import MultiQueryRetriever
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOllama, ChatOpenAI
|
||||
from langchain_community.document_loaders import WebBaseLoader
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate, PromptTemplate
|
||||
from langchain_core.pydantic_v1 import BaseModel
|
||||
|
@@ -11,7 +11,7 @@ readme = "README.md"
|
||||
python = ">=3.9,<4"
|
||||
langchain = "^0.1"
|
||||
tiktoken = ">=0.5.1"
|
||||
chromadb = ">=0.4.14"
|
||||
langchain-chroma = "^0.1.2"
|
||||
openai = "<2"
|
||||
unstructured = ">=0.10.19"
|
||||
pdf2image = ">=1.16.3"
|
||||
|
@@ -3,9 +3,9 @@ import uuid
|
||||
|
||||
from langchain.retrievers.multi_vector import MultiVectorRetriever
|
||||
from langchain.storage import InMemoryStore
|
||||
from langchain_chroma import Chroma
|
||||
from langchain_community.chat_models import ChatOpenAI
|
||||
from langchain_community.embeddings import OpenAIEmbeddings
|
||||
from langchain_community.vectorstores import Chroma
|
||||
from langchain_core.documents import Document
|
||||
from langchain_core.output_parsers import StrOutputParser
|
||||
from langchain_core.prompts import ChatPromptTemplate
|
||||
|
Reference in New Issue
Block a user