diff --git a/docs/docs/expression_language/how_to/inspect.ipynb b/docs/docs/expression_language/how_to/inspect.ipynb index 5e7e7f7f7e7..fdf74a16cdc 100644 --- a/docs/docs/expression_language/how_to/inspect.ipynb +++ b/docs/docs/expression_language/how_to/inspect.ipynb @@ -29,9 +29,9 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts import ChatPromptTemplate\n", "from langchain_community.vectorstores import FAISS\n", "from langchain_core.output_parsers import StrOutputParser\n", + "from langchain_core.prompts import ChatPromptTemplate\n", "from langchain_core.runnables import RunnablePassthrough\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings" ] diff --git a/docs/docs/integrations/callbacks/argilla.ipynb b/docs/docs/integrations/callbacks/argilla.ipynb index fc656f9687a..03df4421060 100644 --- a/docs/docs/integrations/callbacks/argilla.ipynb +++ b/docs/docs/integrations/callbacks/argilla.ipynb @@ -278,8 +278,8 @@ ], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_core.callbacks.stdout import StdOutCallbackHandler\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "argilla_callback = ArgillaCallbackHandler(\n", diff --git a/docs/docs/integrations/callbacks/context.ipynb b/docs/docs/integrations/callbacks/context.ipynb index a053b2261c8..3f5bae3c773 100644 --- a/docs/docs/integrations/callbacks/context.ipynb +++ b/docs/docs/integrations/callbacks/context.ipynb @@ -170,8 +170,8 @@ "import os\n", "\n", "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", - "from langchain.prompts.chat import (\n", + "from langchain_core.prompts import PromptTemplate\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", ")\n", diff --git a/docs/docs/integrations/callbacks/sagemaker_tracking.ipynb b/docs/docs/integrations/callbacks/sagemaker_tracking.ipynb index 0457e6134c3..adaa7d15717 100644 --- a/docs/docs/integrations/callbacks/sagemaker_tracking.ipynb +++ b/docs/docs/integrations/callbacks/sagemaker_tracking.ipynb @@ -91,7 +91,7 @@ "source": [ "from langchain.agents import initialize_agent, load_tools\n", "from langchain.chains import LLMChain, SimpleSequentialChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "from sagemaker.analytics import ExperimentAnalytics\n", "from sagemaker.experiments.run import Run\n", diff --git a/docs/docs/integrations/chat/jinachat.ipynb b/docs/docs/integrations/chat/jinachat.ipynb index 08b247c7698..42fc8f6907d 100644 --- a/docs/docs/integrations/chat/jinachat.ipynb +++ b/docs/docs/integrations/chat/jinachat.ipynb @@ -19,13 +19,13 @@ }, "outputs": [], "source": [ - "from langchain.prompts.chat import (\n", + "from langchain_community.chat_models import JinaChat\n", + "from langchain_core.messages import HumanMessage, SystemMessage\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " SystemMessagePromptTemplate,\n", - ")\n", - "from langchain_community.chat_models import JinaChat\n", - "from langchain_core.messages import HumanMessage, SystemMessage" + ")" ] }, { diff --git a/docs/docs/integrations/chat/llama2_chat.ipynb b/docs/docs/integrations/chat/llama2_chat.ipynb index 9b4623bcc13..f3e6059fb49 100644 --- a/docs/docs/integrations/chat/llama2_chat.ipynb +++ b/docs/docs/integrations/chat/llama2_chat.ipynb @@ -49,12 +49,12 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts.chat import (\n", + "from langchain_core.messages import SystemMessage\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " MessagesPlaceholder,\n", ")\n", - "from langchain_core.messages import SystemMessage\n", "\n", "template_messages = [\n", " SystemMessage(content=\"You are a helpful assistant.\"),\n", diff --git a/docs/docs/integrations/chat/vllm.ipynb b/docs/docs/integrations/chat/vllm.ipynb index ef03e1d8d13..9d70a06a3f1 100644 --- a/docs/docs/integrations/chat/vllm.ipynb +++ b/docs/docs/integrations/chat/vllm.ipynb @@ -31,12 +31,12 @@ }, "outputs": [], "source": [ - "from langchain.prompts.chat import (\n", + "from langchain_core.messages import HumanMessage, SystemMessage\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " SystemMessagePromptTemplate,\n", ")\n", - "from langchain_core.messages import HumanMessage, SystemMessage\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/integrations/chat/yuan2.ipynb b/docs/docs/integrations/chat/yuan2.ipynb index c4a0ca4fc73..0e37110aebf 100644 --- a/docs/docs/integrations/chat/yuan2.ipynb +++ b/docs/docs/integrations/chat/yuan2.ipynb @@ -348,7 +348,7 @@ "outputs": [], "source": [ "async def ainvoke_with_prompt_template():\n", - " from langchain.prompts.chat import (\n", + " from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " )\n", "\n", diff --git a/docs/docs/integrations/chat_loaders/facebook.ipynb b/docs/docs/integrations/chat_loaders/facebook.ipynb index 1e682d90b66..fc3a346f16c 100644 --- a/docs/docs/integrations/chat_loaders/facebook.ipynb +++ b/docs/docs/integrations/chat_loaders/facebook.ipynb @@ -258,7 +258,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.adapters.openai import convert_messages_for_finetuning" + "from langchain_community.adapters.openai import convert_messages_for_finetuning" ] }, { diff --git a/docs/docs/integrations/chat_loaders/imessage.ipynb b/docs/docs/integrations/chat_loaders/imessage.ipynb index 1344b9af740..f5a3acc6a71 100644 --- a/docs/docs/integrations/chat_loaders/imessage.ipynb +++ b/docs/docs/integrations/chat_loaders/imessage.ipynb @@ -173,7 +173,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.adapters.openai import convert_messages_for_finetuning" + "from langchain_community.adapters.openai import convert_messages_for_finetuning" ] }, { diff --git a/docs/docs/integrations/chat_loaders/langsmith_dataset.ipynb b/docs/docs/integrations/chat_loaders/langsmith_dataset.ipynb index 85586b03711..162747309e1 100644 --- a/docs/docs/integrations/chat_loaders/langsmith_dataset.ipynb +++ b/docs/docs/integrations/chat_loaders/langsmith_dataset.ipynb @@ -150,7 +150,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.adapters.openai import convert_messages_for_finetuning\n", + "from langchain_community.adapters.openai import convert_messages_for_finetuning\n", "\n", "training_data = convert_messages_for_finetuning(chat_sessions)" ] diff --git a/docs/docs/integrations/chat_loaders/langsmith_llm_runs.ipynb b/docs/docs/integrations/chat_loaders/langsmith_llm_runs.ipynb index dd3db4f2304..b703be818c3 100644 --- a/docs/docs/integrations/chat_loaders/langsmith_llm_runs.ipynb +++ b/docs/docs/integrations/chat_loaders/langsmith_llm_runs.ipynb @@ -285,7 +285,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.adapters.openai import convert_messages_for_finetuning\n", + "from langchain_community.adapters.openai import convert_messages_for_finetuning\n", "\n", "training_data = convert_messages_for_finetuning(chat_sessions)" ] diff --git a/docs/docs/integrations/chat_loaders/twitter.ipynb b/docs/docs/integrations/chat_loaders/twitter.ipynb index e906af7e67b..4dab87219d0 100644 --- a/docs/docs/integrations/chat_loaders/twitter.ipynb +++ b/docs/docs/integrations/chat_loaders/twitter.ipynb @@ -21,7 +21,7 @@ "source": [ "import json\n", "\n", - "from langchain.adapters.openai import convert_message_to_dict\n", + "from langchain_community.adapters.openai import convert_message_to_dict\n", "from langchain_core.messages import AIMessage" ] }, diff --git a/docs/docs/integrations/document_loaders/airbyte_cdk.ipynb b/docs/docs/integrations/document_loaders/airbyte_cdk.ipynb index d3bd108db7e..b92098a0392 100644 --- a/docs/docs/integrations/document_loaders/airbyte_cdk.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_cdk.ipynb @@ -166,7 +166,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_gong.ipynb b/docs/docs/integrations/document_loaders/airbyte_gong.ipynb index d6f32e7a0ab..2d479be51cd 100644 --- a/docs/docs/integrations/document_loaders/airbyte_gong.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_gong.ipynb @@ -149,7 +149,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_hubspot.ipynb b/docs/docs/integrations/document_loaders/airbyte_hubspot.ipynb index 0f12481ec15..77b889f6db6 100644 --- a/docs/docs/integrations/document_loaders/airbyte_hubspot.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_hubspot.ipynb @@ -151,7 +151,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_salesforce.ipynb b/docs/docs/integrations/document_loaders/airbyte_salesforce.ipynb index 03ec39faab8..588b0eaef6b 100644 --- a/docs/docs/integrations/document_loaders/airbyte_salesforce.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_salesforce.ipynb @@ -156,7 +156,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_shopify.ipynb b/docs/docs/integrations/document_loaders/airbyte_shopify.ipynb index 33f8c5b6e25..de8733fdc6e 100644 --- a/docs/docs/integrations/document_loaders/airbyte_shopify.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_shopify.ipynb @@ -152,7 +152,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_stripe.ipynb b/docs/docs/integrations/document_loaders/airbyte_stripe.ipynb index 840002a1619..2d097d6e303 100644 --- a/docs/docs/integrations/document_loaders/airbyte_stripe.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_stripe.ipynb @@ -149,7 +149,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_typeform.ipynb b/docs/docs/integrations/document_loaders/airbyte_typeform.ipynb index bce494de745..b8ebdb9b372 100644 --- a/docs/docs/integrations/document_loaders/airbyte_typeform.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_typeform.ipynb @@ -152,7 +152,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/airbyte_zendesk_support.ipynb b/docs/docs/integrations/document_loaders/airbyte_zendesk_support.ipynb index 73bbc53863f..c5a5fd9c101 100644 --- a/docs/docs/integrations/document_loaders/airbyte_zendesk_support.ipynb +++ b/docs/docs/integrations/document_loaders/airbyte_zendesk_support.ipynb @@ -153,7 +153,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def handle_record(record, id):\n", diff --git a/docs/docs/integrations/document_loaders/apify_dataset.ipynb b/docs/docs/integrations/document_loaders/apify_dataset.ipynb index 70184639920..1afa3a4d311 100644 --- a/docs/docs/integrations/document_loaders/apify_dataset.ipynb +++ b/docs/docs/integrations/document_loaders/apify_dataset.ipynb @@ -100,8 +100,8 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", "from langchain.indexes import VectorstoreIndexCreator\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import ApifyDatasetLoader" ] }, diff --git a/docs/docs/integrations/document_loaders/copypaste.ipynb b/docs/docs/integrations/document_loaders/copypaste.ipynb index 1abc65b933b..0375a4813ed 100644 --- a/docs/docs/integrations/document_loaders/copypaste.ipynb +++ b/docs/docs/integrations/document_loaders/copypaste.ipynb @@ -17,7 +17,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document" + "from langchain_community.docstore.document import Document" ] }, { diff --git a/docs/docs/integrations/document_loaders/figma.ipynb b/docs/docs/integrations/document_loaders/figma.ipynb index a01dfbc1f09..c32739339e9 100644 --- a/docs/docs/integrations/document_loaders/figma.ipynb +++ b/docs/docs/integrations/document_loaders/figma.ipynb @@ -24,12 +24,12 @@ "import os\n", "\n", "from langchain.indexes import VectorstoreIndexCreator\n", - "from langchain.prompts.chat import (\n", + "from langchain_community.document_loaders.figma import FigmaFileLoader\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " SystemMessagePromptTemplate,\n", ")\n", - "from langchain_community.document_loaders.figma import FigmaFileLoader\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/integrations/document_loaders/google_drive.ipynb b/docs/docs/integrations/document_loaders/google_drive.ipynb index 2d11faedade..94a11bd7ccb 100644 --- a/docs/docs/integrations/document_loaders/google_drive.ipynb +++ b/docs/docs/integrations/document_loaders/google_drive.ipynb @@ -431,7 +431,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "loader = GoogleDriveLoader(\n", " folder_id=folder_id,\n", diff --git a/docs/docs/integrations/llms/huggingface_endpoint.ipynb b/docs/docs/integrations/llms/huggingface_endpoint.ipynb index a71a987bac1..cfd9db8d525 100644 --- a/docs/docs/integrations/llms/huggingface_endpoint.ipynb +++ b/docs/docs/integrations/llms/huggingface_endpoint.ipynb @@ -93,7 +93,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate" + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/llms/llm_caching.ipynb b/docs/docs/integrations/llms/llm_caching.ipynb index 54971621115..65e55ee1dad 100644 --- a/docs/docs/integrations/llms/llm_caching.ipynb +++ b/docs/docs/integrations/llms/llm_caching.ipynb @@ -1633,7 +1633,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "docs = [Document(page_content=t) for t in texts[:3]]\n", "from langchain.chains.summarize import load_summarize_chain" diff --git a/docs/docs/integrations/llms/sagemaker.ipynb b/docs/docs/integrations/llms/sagemaker.ipynb index 4f418039dc7..122339077f9 100644 --- a/docs/docs/integrations/llms/sagemaker.ipynb +++ b/docs/docs/integrations/llms/sagemaker.ipynb @@ -58,7 +58,7 @@ }, "outputs": [], "source": [ - "from langchain.docstore.document import Document" + "from langchain_community.docstore.document import Document" ] }, { diff --git a/docs/docs/integrations/llms/titan_takeoff.ipynb b/docs/docs/integrations/llms/titan_takeoff.ipynb index ff714a477d4..57a3d3679e3 100644 --- a/docs/docs/integrations/llms/titan_takeoff.ipynb +++ b/docs/docs/integrations/llms/titan_takeoff.ipynb @@ -31,10 +31,10 @@ "\n", "from langchain.callbacks.manager import CallbackManager\n", "from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n", - "from langchain.prompts import PromptTemplate\n", "\n", "# Note importing TitanTakeoffPro instead of TitanTakeoff will work as well both use same object under the hood\n", - "from langchain_community.llms import TitanTakeoff" + "from langchain_community.llms import TitanTakeoff\n", + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/memory/motorhead_memory.ipynb b/docs/docs/integrations/memory/motorhead_memory.ipynb index ae56d38fefe..60d36996e2a 100644 --- a/docs/docs/integrations/memory/motorhead_memory.ipynb +++ b/docs/docs/integrations/memory/motorhead_memory.ipynb @@ -36,7 +36,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "template = \"\"\"You are a chatbot having a conversation with a human.\n", diff --git a/docs/docs/integrations/providers/aim_tracking.ipynb b/docs/docs/integrations/providers/aim_tracking.ipynb index 055249c8ecc..cadf2ba1f85 100644 --- a/docs/docs/integrations/providers/aim_tracking.ipynb +++ b/docs/docs/integrations/providers/aim_tracking.ipynb @@ -174,7 +174,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate" + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/providers/cohere.mdx b/docs/docs/integrations/providers/cohere.mdx index e64231eb383..e6a1861547c 100644 --- a/docs/docs/integrations/providers/cohere.mdx +++ b/docs/docs/integrations/providers/cohere.mdx @@ -51,7 +51,7 @@ Usage of the Cohere (legacy) [LLM model](/docs/integrations/llms/cohere) ```python from langchain_community.tools.tavily_search import TavilySearchResults from langchain_cohere import ChatCohere, create_cohere_react_agent -from langchain.prompts import ChatPromptTemplate +from langchain_core.prompts import ChatPromptTemplate from langchain.agents import AgentExecutor llm = ChatCohere() diff --git a/docs/docs/integrations/providers/comet_tracking.ipynb b/docs/docs/integrations/providers/comet_tracking.ipynb index c102b833d84..67e427bd235 100644 --- a/docs/docs/integrations/providers/comet_tracking.ipynb +++ b/docs/docs/integrations/providers/comet_tracking.ipynb @@ -154,7 +154,7 @@ "source": [ "from langchain.callbacks import CometCallbackHandler, StdOutCallbackHandler\n", "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "comet_callback = CometCallbackHandler(\n", @@ -251,7 +251,7 @@ "source": [ "from langchain.callbacks import CometCallbackHandler, StdOutCallbackHandler\n", "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "from rouge_score import rouge_scorer\n", "\n", diff --git a/docs/docs/integrations/providers/flyte.mdx b/docs/docs/integrations/providers/flyte.mdx index 38966d5831a..c37fe96c65f 100644 --- a/docs/docs/integrations/providers/flyte.mdx +++ b/docs/docs/integrations/providers/flyte.mdx @@ -29,7 +29,7 @@ from langchain.agents import AgentType, initialize_agent, load_tools from langchain.callbacks import FlyteCallbackHandler from langchain.chains import LLMChain from langchain_openai import ChatOpenAI -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate from langchain_core.messages import HumanMessage ``` diff --git a/docs/docs/integrations/providers/javelin_ai_gateway.mdx b/docs/docs/integrations/providers/javelin_ai_gateway.mdx index 41e7f6fe45b..d678e34597e 100644 --- a/docs/docs/integrations/providers/javelin_ai_gateway.mdx +++ b/docs/docs/integrations/providers/javelin_ai_gateway.mdx @@ -31,7 +31,7 @@ export JAVELIN_API_KEY=... from langchain.chains import LLMChain from langchain_community.llms import JavelinAIGateway -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate route_completions = "eng_dept03" diff --git a/docs/docs/integrations/providers/mlflow_ai_gateway.mdx b/docs/docs/integrations/providers/mlflow_ai_gateway.mdx index e1fa804f032..912ea449eba 100644 --- a/docs/docs/integrations/providers/mlflow_ai_gateway.mdx +++ b/docs/docs/integrations/providers/mlflow_ai_gateway.mdx @@ -140,7 +140,7 @@ Please contact a Databricks representative to enroll in the preview. ```python from langchain.chains import LLMChain -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate from langchain_community.llms import MlflowAIGateway gateway = MlflowAIGateway( diff --git a/docs/docs/integrations/providers/mlflow_tracking.ipynb b/docs/docs/integrations/providers/mlflow_tracking.ipynb index 79aaed4aa9f..bc1ace7b690 100644 --- a/docs/docs/integrations/providers/mlflow_tracking.ipynb +++ b/docs/docs/integrations/providers/mlflow_tracking.ipynb @@ -123,7 +123,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate" + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/providers/predictionguard.mdx b/docs/docs/integrations/providers/predictionguard.mdx index fdb0f0a397c..5e01eeef14d 100644 --- a/docs/docs/integrations/providers/predictionguard.mdx +++ b/docs/docs/integrations/providers/predictionguard.mdx @@ -37,7 +37,7 @@ import os import predictionguard as pg from langchain_community.llms import PredictionGuard -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate from langchain.chains import LLMChain # Your Prediction Guard API key. Get one at predictionguard.com @@ -77,7 +77,7 @@ Basic LLM Chaining with the Prediction Guard wrapper: ```python import os -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate from langchain.chains import LLMChain from langchain_community.llms import PredictionGuard diff --git a/docs/docs/integrations/providers/ray_serve.ipynb b/docs/docs/integrations/providers/ray_serve.ipynb index b48e76710d0..20d3ac7489f 100644 --- a/docs/docs/integrations/providers/ray_serve.ipynb +++ b/docs/docs/integrations/providers/ray_serve.ipynb @@ -108,7 +108,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI" ] }, diff --git a/docs/docs/integrations/providers/rebuff.ipynb b/docs/docs/integrations/providers/rebuff.ipynb index 7b8c07fbf9c..259c2d22525 100644 --- a/docs/docs/integrations/providers/rebuff.ipynb +++ b/docs/docs/integrations/providers/rebuff.ipynb @@ -104,7 +104,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "# Set up the LangChain SDK with the environment variable\n", diff --git a/docs/docs/integrations/providers/shaleprotocol.md b/docs/docs/integrations/providers/shaleprotocol.md index dbdd3caa6cf..eaafb4a3fd2 100644 --- a/docs/docs/integrations/providers/shaleprotocol.md +++ b/docs/docs/integrations/providers/shaleprotocol.md @@ -20,7 +20,7 @@ As of June 2023, the API supports Vicuna-13B by default. We are going to support For example ```python from langchain_openai import OpenAI -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate from langchain.chains import LLMChain import os diff --git a/docs/docs/integrations/providers/wandb_tracking.ipynb b/docs/docs/integrations/providers/wandb_tracking.ipynb index 1229c84318d..b836d1af067 100644 --- a/docs/docs/integrations/providers/wandb_tracking.ipynb +++ b/docs/docs/integrations/providers/wandb_tracking.ipynb @@ -382,7 +382,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate" + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/retrievers/chatgpt-plugin.ipynb b/docs/docs/integrations/retrievers/chatgpt-plugin.ipynb index 356b13fde1d..1e0388606b2 100644 --- a/docs/docs/integrations/retrievers/chatgpt-plugin.ipynb +++ b/docs/docs/integrations/retrievers/chatgpt-plugin.ipynb @@ -45,7 +45,7 @@ "import json\n", "from typing import List\n", "\n", - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "\n", "def write_json(path: str, documents: List[Document]) -> None:\n", diff --git a/docs/docs/integrations/retrievers/google_drive.ipynb b/docs/docs/integrations/retrievers/google_drive.ipynb index a316fc864c2..f008627c3a9 100644 --- a/docs/docs/integrations/retrievers/google_drive.ipynb +++ b/docs/docs/integrations/retrievers/google_drive.ipynb @@ -171,7 +171,7 @@ }, "outputs": [], "source": [ - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "retriever = GoogleDriveRetriever(\n", " template=PromptTemplate(\n", diff --git a/docs/docs/integrations/retrievers/re_phrase.ipynb b/docs/docs/integrations/retrievers/re_phrase.ipynb index dd4dcd5f7fe..5cbf3c0f8c1 100644 --- a/docs/docs/integrations/retrievers/re_phrase.ipynb +++ b/docs/docs/integrations/retrievers/re_phrase.ipynb @@ -141,7 +141,7 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "QUERY_PROMPT = PromptTemplate(\n", " input_variables=[\"question\"],\n", diff --git a/docs/docs/integrations/text_embedding/clarifai.ipynb b/docs/docs/integrations/text_embedding/clarifai.ipynb index f10a9a463a6..f1f3d6ca14c 100644 --- a/docs/docs/integrations/text_embedding/clarifai.ipynb +++ b/docs/docs/integrations/text_embedding/clarifai.ipynb @@ -74,8 +74,8 @@ "source": [ "# Import the required modules\n", "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", - "from langchain_community.embeddings import ClarifaiEmbeddings" + "from langchain_community.embeddings import ClarifaiEmbeddings\n", + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/integrations/tools/dalle_image_generator.ipynb b/docs/docs/integrations/tools/dalle_image_generator.ipynb index 0256bd2c850..d6a6515110c 100644 --- a/docs/docs/integrations/tools/dalle_image_generator.ipynb +++ b/docs/docs/integrations/tools/dalle_image_generator.ipynb @@ -52,8 +52,8 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_community.utilities.dalle_image_generator import DallEAPIWrapper\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "llm = OpenAI(temperature=0.9)\n", diff --git a/docs/docs/integrations/tools/reddit_search.ipynb b/docs/docs/integrations/tools/reddit_search.ipynb index 004da7a6401..52ac17a1fcd 100644 --- a/docs/docs/integrations/tools/reddit_search.ipynb +++ b/docs/docs/integrations/tools/reddit_search.ipynb @@ -172,9 +172,9 @@ "from langchain.agents import AgentExecutor, StructuredChatAgent, Tool\n", "from langchain.chains import LLMChain\n", "from langchain.memory import ConversationBufferMemory, ReadOnlySharedMemory\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_community.tools.reddit_search.tool import RedditSearchRun\n", "from langchain_community.utilities.reddit_search import RedditSearchAPIWrapper\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import ChatOpenAI\n", "\n", "# Provide keys for Reddit\n", diff --git a/docs/docs/integrations/tools/zapier.ipynb b/docs/docs/integrations/tools/zapier.ipynb index 5ef000319fe..a6deab26308 100644 --- a/docs/docs/integrations/tools/zapier.ipynb +++ b/docs/docs/integrations/tools/zapier.ipynb @@ -161,9 +161,9 @@ "outputs": [], "source": [ "from langchain.chains import LLMChain, SimpleSequentialChain, TransformChain\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_community.tools.zapier.tool import ZapierNLARunAction\n", "from langchain_community.utilities.zapier import ZapierNLAWrapper\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI" ] }, diff --git a/docs/docs/integrations/vectorstores/annoy.ipynb b/docs/docs/integrations/vectorstores/annoy.ipynb index fb953031ea9..470b2a02865 100644 --- a/docs/docs/integrations/vectorstores/annoy.ipynb +++ b/docs/docs/integrations/vectorstores/annoy.ipynb @@ -497,8 +497,8 @@ "import uuid\n", "\n", "from annoy import AnnoyIndex\n", - "from langchain.docstore.document import Document\n", - "from langchain.docstore.in_memory import InMemoryDocstore\n", + "from langchain_community.docstore.document import Document\n", + "from langchain_community.docstore.in_memory import InMemoryDocstore\n", "\n", "metadatas = [{\"x\": \"food\"}, {\"x\": \"food\"}, {\"x\": \"stuff\"}, {\"x\": \"animal\"}]\n", "\n", diff --git a/docs/docs/integrations/vectorstores/kinetica.ipynb b/docs/docs/integrations/vectorstores/kinetica.ipynb index 5ff269ee445..c8a666f2497 100644 --- a/docs/docs/integrations/vectorstores/kinetica.ipynb +++ b/docs/docs/integrations/vectorstores/kinetica.ipynb @@ -113,7 +113,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.vectorstores import (\n", " DistanceStrategy,\n", diff --git a/docs/docs/integrations/vectorstores/milvus.ipynb b/docs/docs/integrations/vectorstores/milvus.ipynb index 8d44b2dab04..b246a569ede 100644 --- a/docs/docs/integrations/vectorstores/milvus.ipynb +++ b/docs/docs/integrations/vectorstores/milvus.ipynb @@ -340,7 +340,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "\n", "# Insert data sample\n", "docs = [\n", diff --git a/docs/docs/integrations/vectorstores/mongodb_atlas.ipynb b/docs/docs/integrations/vectorstores/mongodb_atlas.ipynb index 57c2ebf3361..24081245dbf 100644 --- a/docs/docs/integrations/vectorstores/mongodb_atlas.ipynb +++ b/docs/docs/integrations/vectorstores/mongodb_atlas.ipynb @@ -365,7 +365,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "prompt_template = \"\"\"Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.\n", "\n", diff --git a/docs/docs/integrations/vectorstores/neo4jvector.ipynb b/docs/docs/integrations/vectorstores/neo4jvector.ipynb index 1b5c75455ab..1f3cf2d5089 100644 --- a/docs/docs/integrations/vectorstores/neo4jvector.ipynb +++ b/docs/docs/integrations/vectorstores/neo4jvector.ipynb @@ -72,7 +72,7 @@ }, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.vectorstores import Neo4jVector\n", "from langchain_openai import OpenAIEmbeddings\n", diff --git a/docs/docs/integrations/vectorstores/pgembedding.ipynb b/docs/docs/integrations/vectorstores/pgembedding.ipynb index 89a4279bc36..c4128704a05 100644 --- a/docs/docs/integrations/vectorstores/pgembedding.ipynb +++ b/docs/docs/integrations/vectorstores/pgembedding.ipynb @@ -81,7 +81,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.vectorstores import PGEmbedding\n", "from langchain_openai import OpenAIEmbeddings\n", diff --git a/docs/docs/integrations/vectorstores/pgvecto_rs.ipynb b/docs/docs/integrations/vectorstores/pgvecto_rs.ipynb index 1f4c2647084..4f9e6155244 100644 --- a/docs/docs/integrations/vectorstores/pgvecto_rs.ipynb +++ b/docs/docs/integrations/vectorstores/pgvecto_rs.ipynb @@ -26,7 +26,7 @@ "source": [ "from typing import List\n", "\n", - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.embeddings.fake import FakeEmbeddings\n", "from langchain_community.vectorstores.pgvecto_rs import PGVecto_rs\n", diff --git a/docs/docs/integrations/vectorstores/sap_hanavector.ipynb b/docs/docs/integrations/vectorstores/sap_hanavector.ipynb index f416f6b622e..d9437fef1df 100644 --- a/docs/docs/integrations/vectorstores/sap_hanavector.ipynb +++ b/docs/docs/integrations/vectorstores/sap_hanavector.ipynb @@ -109,7 +109,7 @@ }, "outputs": [], "source": [ - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.vectorstores.hanavector import HanaDB\n", "from langchain_openai import OpenAIEmbeddings\n", diff --git a/docs/docs/integrations/vectorstores/timescalevector.ipynb b/docs/docs/integrations/vectorstores/timescalevector.ipynb index 067928f7cbf..5058b82645a 100644 --- a/docs/docs/integrations/vectorstores/timescalevector.ipynb +++ b/docs/docs/integrations/vectorstores/timescalevector.ipynb @@ -122,7 +122,7 @@ "source": [ "from datetime import datetime, timedelta\n", "\n", - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.document_loaders import TextLoader\n", "from langchain_community.document_loaders.json_loader import JSONLoader\n", "from langchain_community.vectorstores.timescalevector import TimescaleVector\n", diff --git a/docs/docs/integrations/vectorstores/yellowbrick.ipynb b/docs/docs/integrations/vectorstores/yellowbrick.ipynb index efd9e9bf9fb..367fc8ca58f 100644 --- a/docs/docs/integrations/vectorstores/yellowbrick.ipynb +++ b/docs/docs/integrations/vectorstores/yellowbrick.ipynb @@ -98,7 +98,7 @@ "import psycopg2\n", "from IPython.display import Markdown, display\n", "from langchain.chains import LLMChain, RetrievalQAWithSourcesChain\n", - "from langchain.docstore.document import Document\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.vectorstores import Yellowbrick\n", "from langchain_openai import ChatOpenAI, OpenAIEmbeddings\n", "from langchain_text_splitters import RecursiveCharacterTextSplitter\n", @@ -115,7 +115,7 @@ "# API Key for OpenAI. Signup at https://platform.openai.com\n", "os.environ[\"OPENAI_API_KEY\"] = OPENAI_API_KEY\n", "\n", - "from langchain.prompts.chat import (\n", + "from langchain_core.prompts.chat import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " SystemMessagePromptTemplate,\n", diff --git a/docs/docs/modules/agents/how_to/custom_agent.ipynb b/docs/docs/modules/agents/how_to/custom_agent.ipynb index e49200da22b..8376017632d 100644 --- a/docs/docs/modules/agents/how_to/custom_agent.ipynb +++ b/docs/docs/modules/agents/how_to/custom_agent.ipynb @@ -302,7 +302,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts import MessagesPlaceholder\n", + "from langchain_core.prompts import MessagesPlaceholder\n", "\n", "MEMORY_KEY = \"chat_history\"\n", "prompt = ChatPromptTemplate.from_messages(\n", diff --git a/docs/docs/modules/agents/how_to/streaming.ipynb b/docs/docs/modules/agents/how_to/streaming.ipynb index b88e457752a..a55ed167486 100644 --- a/docs/docs/modules/agents/how_to/streaming.ipynb +++ b/docs/docs/modules/agents/how_to/streaming.ipynb @@ -40,9 +40,9 @@ "source": [ "from langchain import hub\n", "from langchain.agents import AgentExecutor, create_openai_tools_agent\n", - "from langchain.prompts import ChatPromptTemplate\n", "from langchain.tools import tool\n", "from langchain_core.callbacks import Callbacks\n", + "from langchain_core.prompts import ChatPromptTemplate\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/modules/data_connection/document_loaders/pdf.mdx b/docs/docs/modules/data_connection/document_loaders/pdf.mdx index cbb8e289263..936aafdd89c 100644 --- a/docs/docs/modules/data_connection/document_loaders/pdf.mdx +++ b/docs/docs/modules/data_connection/document_loaders/pdf.mdx @@ -298,7 +298,7 @@ snippets.append((cur_text,cur_fs)) ```python -from langchain.docstore.document import Document +from langchain_community.docstore.document import Document cur_idx = -1 semantic_snippets = [] # Assumption: headings have higher font size than their respective content diff --git a/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb b/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb index f52015fb780..f1f52d0fa52 100644 --- a/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb +++ b/docs/docs/modules/data_connection/retrievers/long_context_reorder.ipynb @@ -52,12 +52,12 @@ ], "source": [ "from langchain.chains import LLMChain, StuffDocumentsChain\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_chroma import Chroma\n", "from langchain_community.document_transformers import (\n", " LongContextReorder,\n", ")\n", "from langchain_community.embeddings import HuggingFaceEmbeddings\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI\n", "\n", "# Get embeddings.\n", diff --git a/docs/docs/modules/memory/adding_memory.ipynb b/docs/docs/modules/memory/adding_memory.ipynb index ba994224b07..46574f03371 100644 --- a/docs/docs/modules/memory/adding_memory.ipynb +++ b/docs/docs/modules/memory/adding_memory.ipynb @@ -176,12 +176,12 @@ }, "outputs": [], "source": [ - "from langchain.prompts import (\n", + "from langchain_core.messages import SystemMessage\n", + "from langchain_core.prompts import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " MessagesPlaceholder,\n", ")\n", - "from langchain_core.messages import SystemMessage\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/modules/memory/conversational_customization.ipynb b/docs/docs/modules/memory/conversational_customization.ipynb index 159938b4a0c..23ea0d86066 100644 --- a/docs/docs/modules/memory/conversational_customization.ipynb +++ b/docs/docs/modules/memory/conversational_customization.ipynb @@ -135,7 +135,7 @@ "outputs": [], "source": [ "# Now we can override it and set it to \"AI Assistant\"\n", - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "template = \"\"\"The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.\n", "\n", @@ -250,7 +250,7 @@ "outputs": [], "source": [ "# Now we can override it and set it to \"Friend\"\n", - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "template = \"\"\"The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know.\n", "\n", diff --git a/docs/docs/modules/memory/custom_memory.ipynb b/docs/docs/modules/memory/custom_memory.ipynb index 148c88e02dc..f53b65e0ec1 100644 --- a/docs/docs/modules/memory/custom_memory.ipynb +++ b/docs/docs/modules/memory/custom_memory.ipynb @@ -131,7 +131,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "template = \"\"\"The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. If the AI does not know the answer to a question, it truthfully says it does not know. You are provided with information about entities the Human mentions, if relevant.\n", "\n", diff --git a/docs/docs/modules/memory/index.mdx b/docs/docs/modules/memory/index.mdx index 455dbd85209..6e283b92e28 100644 --- a/docs/docs/modules/memory/index.mdx +++ b/docs/docs/modules/memory/index.mdx @@ -201,7 +201,7 @@ conversation({"question": "hi"}) ```python from langchain_openai import ChatOpenAI -from langchain.prompts import ( +from langchain_core.prompts import ( ChatPromptTemplate, MessagesPlaceholder, SystemMessagePromptTemplate, diff --git a/docs/docs/modules/memory/types/kg.ipynb b/docs/docs/modules/memory/types/kg.ipynb index 8dd648cde97..ec6842411df 100644 --- a/docs/docs/modules/memory/types/kg.ipynb +++ b/docs/docs/modules/memory/types/kg.ipynb @@ -181,7 +181,7 @@ "source": [ "llm = OpenAI(temperature=0)\n", "from langchain.chains import ConversationChain\n", - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "template = \"\"\"The following is a friendly conversation between a human and an AI. The AI is talkative and provides lots of specific details from its context. \n", "If the AI does not know the answer to a question, it truthfully says it does not know. The AI ONLY uses information contained in the \"Relevant Information\" section and does not hallucinate.\n", diff --git a/docs/docs/modules/memory/types/vectorstore_retriever_memory.mdx b/docs/docs/modules/memory/types/vectorstore_retriever_memory.mdx index 14c7ad0bdca..71c474293fd 100644 --- a/docs/docs/modules/memory/types/vectorstore_retriever_memory.mdx +++ b/docs/docs/modules/memory/types/vectorstore_retriever_memory.mdx @@ -23,7 +23,7 @@ Depending on the store you choose, this step may look different. Consult the rel ```python import faiss -from langchain.docstore import InMemoryDocstore +from langchain_community.docstore import InMemoryDocstore from langchain_community.vectorstores import FAISS diff --git a/docs/docs/modules/model_io/index.mdx b/docs/docs/modules/model_io/index.mdx index d6850cdd4f7..7740b88c344 100644 --- a/docs/docs/modules/model_io/index.mdx +++ b/docs/docs/modules/model_io/index.mdx @@ -209,7 +209,7 @@ They bundle up all the logic for going from user input into a fully formatted pr This can start off very simple - for example, a prompt to produce the above string would just be: ```python -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate prompt = PromptTemplate.from_template("What is a good name for a company that makes {product}?") prompt.format(product="colorful socks") @@ -231,7 +231,7 @@ Each `ChatMessageTemplate` contains instructions for how to format that `ChatMes Let's take a look at this below: ```python -from langchain.prompts.chat import ChatPromptTemplate +from langchain_core.prompts.chat import ChatPromptTemplate template = "You are a helpful assistant that translates {input_language} to {output_language}." human_template = "{text}" diff --git a/docs/docs/modules/model_io/output_parsers/quick_start.ipynb b/docs/docs/modules/model_io/output_parsers/quick_start.ipynb index 5a4effdaee2..2f19eaf7dd6 100644 --- a/docs/docs/modules/model_io/output_parsers/quick_start.ipynb +++ b/docs/docs/modules/model_io/output_parsers/quick_start.ipynb @@ -51,7 +51,7 @@ ], "source": [ "from langchain.output_parsers import PydanticOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_core.pydantic_v1 import BaseModel, Field, validator\n", "from langchain_openai import OpenAI\n", "\n", diff --git a/docs/docs/modules/model_io/output_parsers/types/csv.ipynb b/docs/docs/modules/model_io/output_parsers/types/csv.ipynb index aaa607c447c..82f4eb380f1 100644 --- a/docs/docs/modules/model_io/output_parsers/types/csv.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/csv.ipynb @@ -18,7 +18,7 @@ "outputs": [], "source": [ "from langchain.output_parsers import CommaSeparatedListOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import ChatOpenAI\n", "\n", "output_parser = CommaSeparatedListOutputParser()\n", diff --git a/docs/docs/modules/model_io/output_parsers/types/datetime.ipynb b/docs/docs/modules/model_io/output_parsers/types/datetime.ipynb index 51193fa11d6..98333df1c74 100644 --- a/docs/docs/modules/model_io/output_parsers/types/datetime.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/datetime.ipynb @@ -18,7 +18,7 @@ "outputs": [], "source": [ "from langchain.output_parsers import DatetimeOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import OpenAI" ] }, diff --git a/docs/docs/modules/model_io/output_parsers/types/json.ipynb b/docs/docs/modules/model_io/output_parsers/types/json.ipynb index 3c43b97f797..3363c598a10 100644 --- a/docs/docs/modules/model_io/output_parsers/types/json.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/json.ipynb @@ -22,8 +22,8 @@ "source": [ "from typing import List\n", "\n", - "from langchain.prompts import PromptTemplate\n", "from langchain_core.output_parsers import JsonOutputParser\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_core.pydantic_v1 import BaseModel, Field\n", "from langchain_openai import ChatOpenAI" ] diff --git a/docs/docs/modules/model_io/output_parsers/types/pandas_dataframe.ipynb b/docs/docs/modules/model_io/output_parsers/types/pandas_dataframe.ipynb index b007de73ffc..3c0bbc337b2 100644 --- a/docs/docs/modules/model_io/output_parsers/types/pandas_dataframe.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/pandas_dataframe.ipynb @@ -24,7 +24,7 @@ "\n", "import pandas as pd\n", "from langchain.output_parsers import PandasDataFrameOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/modules/model_io/output_parsers/types/pydantic.ipynb b/docs/docs/modules/model_io/output_parsers/types/pydantic.ipynb index 1767a0cfc01..730ae1cbc55 100644 --- a/docs/docs/modules/model_io/output_parsers/types/pydantic.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/pydantic.ipynb @@ -23,7 +23,7 @@ "from typing import List\n", "\n", "from langchain.output_parsers import PydanticOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_core.pydantic_v1 import BaseModel, Field, validator\n", "from langchain_openai import ChatOpenAI" ] diff --git a/docs/docs/modules/model_io/output_parsers/types/retry.ipynb b/docs/docs/modules/model_io/output_parsers/types/retry.ipynb index 779f820fcf8..8eb7857a81a 100644 --- a/docs/docs/modules/model_io/output_parsers/types/retry.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/retry.ipynb @@ -21,7 +21,7 @@ " OutputFixingParser,\n", " PydanticOutputParser,\n", ")\n", - "from langchain.prompts import (\n", + "from langchain_core.prompts import (\n", " PromptTemplate,\n", ")\n", "from langchain_core.pydantic_v1 import BaseModel, Field\n", diff --git a/docs/docs/modules/model_io/output_parsers/types/structured.ipynb b/docs/docs/modules/model_io/output_parsers/types/structured.ipynb index 2bd9c0735a9..237c8c343db 100644 --- a/docs/docs/modules/model_io/output_parsers/types/structured.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/structured.ipynb @@ -18,7 +18,7 @@ "outputs": [], "source": [ "from langchain.output_parsers import ResponseSchema, StructuredOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_openai import ChatOpenAI" ] }, diff --git a/docs/docs/modules/model_io/output_parsers/types/xml.ipynb b/docs/docs/modules/model_io/output_parsers/types/xml.ipynb index bcc56692890..5699e42f6e0 100644 --- a/docs/docs/modules/model_io/output_parsers/types/xml.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/xml.ipynb @@ -21,8 +21,8 @@ "outputs": [], "source": [ "from langchain.output_parsers import XMLOutputParser\n", - "from langchain.prompts import PromptTemplate\n", - "from langchain_community.chat_models import ChatAnthropic" + "from langchain_community.chat_models import ChatAnthropic\n", + "from langchain_core.prompts import PromptTemplate" ] }, { diff --git a/docs/docs/modules/model_io/output_parsers/types/yaml.ipynb b/docs/docs/modules/model_io/output_parsers/types/yaml.ipynb index 3f763114d9e..8918cbb407d 100644 --- a/docs/docs/modules/model_io/output_parsers/types/yaml.ipynb +++ b/docs/docs/modules/model_io/output_parsers/types/yaml.ipynb @@ -23,7 +23,7 @@ "from typing import List\n", "\n", "from langchain.output_parsers import YamlOutputParser\n", - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "from langchain_core.pydantic_v1 import BaseModel, Field\n", "from langchain_openai import ChatOpenAI" ] diff --git a/docs/docs/modules/model_io/prompts/composition.ipynb b/docs/docs/modules/model_io/prompts/composition.ipynb index fe81807cb3f..43c6a7c641b 100644 --- a/docs/docs/modules/model_io/prompts/composition.ipynb +++ b/docs/docs/modules/model_io/prompts/composition.ipynb @@ -37,7 +37,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts import PromptTemplate" + "from langchain_core.prompts import PromptTemplate" ] }, { @@ -339,8 +339,8 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts.pipeline import PipelinePromptTemplate\n", - "from langchain.prompts.prompt import PromptTemplate" + "from langchain_core.prompts.pipeline import PipelinePromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate" ] }, { diff --git a/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb b/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb index 04a567ae9fa..7a92dadbb73 100644 --- a/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb +++ b/docs/docs/modules/model_io/prompts/few_shot_examples.ipynb @@ -38,8 +38,8 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts.few_shot import FewShotPromptTemplate\n", - "from langchain.prompts.prompt import PromptTemplate\n", + "from langchain_core.prompts.few_shot import FewShotPromptTemplate\n", + "from langchain_core.prompts.prompt import PromptTemplate\n", "\n", "examples = [\n", " {\n", @@ -253,8 +253,8 @@ } ], "source": [ - "from langchain.prompts.example_selector import SemanticSimilarityExampleSelector\n", "from langchain_chroma import Chroma\n", + "from langchain_core.prompts.example_selector import SemanticSimilarityExampleSelector\n", "from langchain_openai import OpenAIEmbeddings\n", "\n", "example_selector = SemanticSimilarityExampleSelector.from_examples(\n", diff --git a/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb b/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb index 52cb0ea0b5f..f83ab6344bd 100644 --- a/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb +++ b/docs/docs/modules/model_io/prompts/few_shot_examples_chat.ipynb @@ -52,7 +52,7 @@ }, "outputs": [], "source": [ - "from langchain.prompts import (\n", + "from langchain_core.prompts import (\n", " ChatPromptTemplate,\n", " FewShotChatMessagePromptTemplate,\n", ")" @@ -201,8 +201,8 @@ }, "outputs": [], "source": [ - "from langchain.prompts import SemanticSimilarityExampleSelector\n", "from langchain_chroma import Chroma\n", + "from langchain_core.prompts import SemanticSimilarityExampleSelector\n", "from langchain_openai import OpenAIEmbeddings" ] }, @@ -298,7 +298,7 @@ }, "outputs": [], "source": [ - "from langchain.prompts import (\n", + "from langchain_core.prompts import (\n", " ChatPromptTemplate,\n", " FewShotChatMessagePromptTemplate,\n", ")\n", diff --git a/docs/docs/modules/model_io/prompts/partial.ipynb b/docs/docs/modules/model_io/prompts/partial.ipynb index e34fb7d43f2..ff82434fbae 100644 --- a/docs/docs/modules/model_io/prompts/partial.ipynb +++ b/docs/docs/modules/model_io/prompts/partial.ipynb @@ -45,7 +45,7 @@ } ], "source": [ - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "prompt = PromptTemplate.from_template(\"{foo}{bar}\")\n", "partial_prompt = prompt.partial(foo=\"foo\")\n", diff --git a/docs/docs/modules/model_io/prompts/quick_start.ipynb b/docs/docs/modules/model_io/prompts/quick_start.ipynb index bad77c66cb9..adeae37a438 100644 --- a/docs/docs/modules/model_io/prompts/quick_start.ipynb +++ b/docs/docs/modules/model_io/prompts/quick_start.ipynb @@ -56,7 +56,7 @@ } ], "source": [ - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "prompt_template = PromptTemplate.from_template(\n", " \"Tell me a {adjective} joke about {content}.\"\n", @@ -90,7 +90,7 @@ } ], "source": [ - "from langchain.prompts import PromptTemplate\n", + "from langchain_core.prompts import PromptTemplate\n", "\n", "prompt_template = PromptTemplate.from_template(\"Tell me a joke\")\n", "prompt_template.format()" @@ -201,8 +201,8 @@ } ], "source": [ - "from langchain.prompts import HumanMessagePromptTemplate\n", "from langchain_core.messages import SystemMessage\n", + "from langchain_core.prompts import HumanMessagePromptTemplate\n", "\n", "chat_template = ChatPromptTemplate.from_messages(\n", " [\n", @@ -263,7 +263,7 @@ } ], "source": [ - "from langchain.prompts import ChatMessagePromptTemplate\n", + "from langchain_core.prompts import ChatMessagePromptTemplate\n", "\n", "prompt = \"May the {subject} be with you\"\n", "\n", @@ -290,7 +290,7 @@ "metadata": {}, "outputs": [], "source": [ - "from langchain.prompts import (\n", + "from langchain_core.prompts import (\n", " ChatPromptTemplate,\n", " HumanMessagePromptTemplate,\n", " MessagesPlaceholder,\n", diff --git a/docs/docs/modules/model_io/quick_start.mdx b/docs/docs/modules/model_io/quick_start.mdx index 26d1ed3bfef..6764b75e84a 100644 --- a/docs/docs/modules/model_io/quick_start.mdx +++ b/docs/docs/modules/model_io/quick_start.mdx @@ -161,7 +161,7 @@ They bundle up all the logic for going from user input into a fully formatted pr This can start off very simple - for example, a prompt to produce the above string would just be: ```python -from langchain.prompts import PromptTemplate +from langchain_core.prompts import PromptTemplate prompt = PromptTemplate.from_template("What is a good name for a company that makes {product}?") prompt.format(product="colorful socks") diff --git a/docs/docs/use_cases/data_generation.ipynb b/docs/docs/use_cases/data_generation.ipynb index 82a55cf0524..f96737b2deb 100644 --- a/docs/docs/use_cases/data_generation.ipynb +++ b/docs/docs/use_cases/data_generation.ipynb @@ -64,7 +64,7 @@ "# import dotenv\n", "# dotenv.load_dotenv()\n", "\n", - "from langchain.prompts import FewShotPromptTemplate, PromptTemplate\n", + "from langchain_core.prompts import FewShotPromptTemplate, PromptTemplate\n", "from langchain_core.pydantic_v1 import BaseModel\n", "from langchain_experimental.tabular_synthetic_data.openai import (\n", " OPENAI_TEMPLATE,\n", diff --git a/docs/docs/use_cases/web_scraping.ipynb b/docs/docs/use_cases/web_scraping.ipynb index 9735afa3dba..7af02e71458 100644 --- a/docs/docs/use_cases/web_scraping.ipynb +++ b/docs/docs/use_cases/web_scraping.ipynb @@ -628,8 +628,8 @@ } ], "source": [ - "from langchain.docstore.document import Document\n", "from langchain.indexes import VectorstoreIndexCreator\n", + "from langchain_community.docstore.document import Document\n", "from langchain_community.utilities import ApifyWrapper\n", "\n", "apify = ApifyWrapper()\n",