Harrison/stop importing from init (#10690)

This commit is contained in:
Harrison Chase
2023-09-16 17:22:48 -07:00
committed by GitHub
parent 9749f8ebae
commit 5442d2b1fa
202 changed files with 321 additions and 327 deletions

23
.github/workflows/imports.yml vendored Normal file
View File

@@ -0,0 +1,23 @@
---
name: Imports
on:
push:
branches: [master]
pull_request:
branches: [master]
jobs:
check:
runs-on: ubuntu-latest
steps:
- name: Checkout repository
uses: actions/checkout@v2
- name: Run import check
run: |
# We should not encourage imports directly from main init file
# Expect for __version__ and hub
# And of course expect for this file
git grep 'from langchain import' | grep -vE 'from langchain import (__version__|hub)' | grep -v '.github/workflows/check-imports.yml' && exit 1 || exit 0

View File

@@ -105,7 +105,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n", "from langchain.llms.fake import FakeListLLM\n",
"from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n", "from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n",
"\n", "\n",
@@ -412,7 +412,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n", "from langchain.llms.fake import FakeListLLM\n",
"\n", "\n",
"template = \"\"\"Question: {question}\n", "template = \"\"\"Question: {question}\n",
@@ -572,8 +572,8 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import HuggingFaceHub\n", "from langchain.llms import HuggingFaceHub\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"\"\"Question: {question}\"\"\"\n", "template = \"\"\"Question: {question}\"\"\"\n",
"\n", "\n",
@@ -697,7 +697,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SagemakerEndpoint\n", "from langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n", "from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains import LLMChain\n", "from langchain.chains import LLMChain\n",
"from langchain.prompts import load_prompt, PromptTemplate\n", "from langchain.prompts import load_prompt, PromptTemplate\n",

View File

@@ -97,7 +97,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SerpAPIWrapper\n", "from langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n", "from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI\n", "from langchain.chat_models import ChatOpenAI\n",

View File

@@ -468,7 +468,8 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\n",
"from langchain.chains import LLMChain\n",
"from langchain.chains.prompt_selector import ConditionalPromptSelector\n", "from langchain.chains.prompt_selector import ConditionalPromptSelector\n",
"\n", "\n",
"DEFAULT_LLAMA_SEARCH_PROMPT = PromptTemplate(\n", "DEFAULT_LLAMA_SEARCH_PROMPT = PromptTemplate(\n",
@@ -593,7 +594,7 @@
"name": "python", "name": "python",
"nbconvert_exporter": "python", "nbconvert_exporter": "python",
"pygments_lexer": "ipython3", "pygments_lexer": "ipython3",
"version": "3.9.1" "version": "3.10.1"
} }
}, },
"nbformat": 4, "nbformat": 4,

View File

@@ -19,7 +19,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import LLMChain, OpenAI, Cohere, HuggingFaceHub, PromptTemplate\n", "from langchain.chains import LLMChain\nfrom langchain.llms import OpenAI, Cohere, HuggingFaceHub\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.model_laboratory import ModelLaboratory" "from langchain.model_laboratory import ModelLaboratory"
] ]
}, },
@@ -139,7 +139,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SelfAskWithSearchChain, SerpAPIWrapper\n", "from langchain.chains import SelfAskWithSearchChain\nfrom langchain.utilities import SerpAPIWrapper\n",
"\n", "\n",
"open_ai_llm = OpenAI(temperature=0)\n", "open_ai_llm = OpenAI(temperature=0)\n",
"search = SerpAPIWrapper()\n", "search = SerpAPIWrapper()\n",

View File

@@ -95,7 +95,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n", "from langchain.llms.fake import FakeListLLM\n",
"from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n", "from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n",
"\n", "\n",
@@ -399,7 +399,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n", "from langchain.llms.fake import FakeListLLM\n",
"\n", "\n",
"template = \"\"\"Question: {question}\n", "template = \"\"\"Question: {question}\n",
@@ -564,8 +564,8 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import HuggingFaceHub\n", "from langchain.llms import HuggingFaceHub\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"\"\"Question: {question}\n", "template = \"\"\"Question: {question}\n",
"\n", "\n",
@@ -679,7 +679,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SagemakerEndpoint\n", "from langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n", "from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains import LLMChain\n", "from langchain.chains import LLMChain\n",
"from langchain.prompts import load_prompt, PromptTemplate\n", "from langchain.prompts import load_prompt, PromptTemplate\n",

View File

@@ -123,7 +123,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType" "from langchain.agents import initialize_agent, AgentType"
] ]
}, },

View File

@@ -167,7 +167,7 @@
"import os\n", "import os\n",
"\n", "\n",
"from langchain.chat_models import ChatOpenAI\n", "from langchain.chat_models import ChatOpenAI\n",
"from langchain import LLMChain\n", "from langchain.chains import LLMChain\n",
"from langchain.prompts import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"from langchain.prompts.chat import (\n", "from langchain.prompts.chat import (\n",
" ChatPromptTemplate,\n", " ChatPromptTemplate,\n",

View File

@@ -199,7 +199,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"\n", "\n",
"# Prompt\n", "# Prompt\n",
"template = \"\"\"[INST] <<SYS>> Use the following pieces of context to answer the question at the end. \n", "template = \"\"\"[INST] <<SYS>> Use the following pieces of context to answer the question at the end. \n",

View File

@@ -81,7 +81,7 @@
"import re\n", "import re\n",
"from typing import Iterator, List\n", "from typing import Iterator, List\n",
"\n", "\n",
"from langchain import schema\n", "from langchain.schema import BaseMessage, HumanMessage\n",
"from langchain.chat_loaders import base as chat_loaders\n", "from langchain.chat_loaders import base as chat_loaders\n",
"\n", "\n",
"logger = logging.getLogger()\n", "logger = logging.getLogger()\n",
@@ -117,7 +117,7 @@
" with open(file_path, \"r\", encoding=\"utf-8\") as file:\n", " with open(file_path, \"r\", encoding=\"utf-8\") as file:\n",
" lines = file.readlines()\n", " lines = file.readlines()\n",
"\n", "\n",
" results: List[schema.BaseMessage] = []\n", " results: List[BaseMessage] = []\n",
" current_sender = None\n", " current_sender = None\n",
" current_timestamp = None\n", " current_timestamp = None\n",
" current_content = []\n", " current_content = []\n",
@@ -128,7 +128,7 @@
" ):\n", " ):\n",
" if current_sender and current_content:\n", " if current_sender and current_content:\n",
" results.append(\n", " results.append(\n",
" schema.HumanMessage(\n", " HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n", " content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n", " additional_kwargs={\n",
" \"sender\": current_sender,\n", " \"sender\": current_sender,\n",
@@ -142,7 +142,7 @@
" ]\n", " ]\n",
" elif re.match(r\"\\[\\d{1,2}:\\d{2} (?:AM|PM)\\]\", line.strip()):\n", " elif re.match(r\"\\[\\d{1,2}:\\d{2} (?:AM|PM)\\]\", line.strip()):\n",
" results.append(\n", " results.append(\n",
" schema.HumanMessage(\n", " HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n", " content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n", " additional_kwargs={\n",
" \"sender\": current_sender,\n", " \"sender\": current_sender,\n",
@@ -157,7 +157,7 @@
"\n", "\n",
" if current_sender and current_content:\n", " if current_sender and current_content:\n",
" results.append(\n", " results.append(\n",
" schema.HumanMessage(\n", " HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n", " content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n", " additional_kwargs={\n",
" \"sender\": current_sender,\n", " \"sender\": current_sender,\n",

View File

@@ -78,7 +78,7 @@
"import re\n", "import re\n",
"from typing import Iterator, List\n", "from typing import Iterator, List\n",
"\n", "\n",
"from langchain import schema\n", "from langchain.schema import HumanMessage, BaseMessage\n",
"from langchain.chat_loaders import base as chat_loaders\n", "from langchain.chat_loaders import base as chat_loaders\n",
"\n", "\n",
"logger = logging.getLogger()\n", "logger = logging.getLogger()\n",
@@ -110,7 +110,7 @@
" # skip non-text messages like stickers, images, etc.\n", " # skip non-text messages like stickers, images, etc.\n",
" if not re.match(r\"\\[.*\\]\", content):\n", " if not re.match(r\"\\[.*\\]\", content):\n",
" results.append(\n", " results.append(\n",
" schema.HumanMessage(\n", " HumanMessage(\n",
" content=content,\n", " content=content,\n",
" additional_kwargs={\n", " additional_kwargs={\n",
" \"sender\": current_sender,\n", " \"sender\": current_sender,\n",
@@ -135,7 +135,7 @@
" with open(file_path, \"r\", encoding=\"utf-8\") as file:\n", " with open(file_path, \"r\", encoding=\"utf-8\") as file:\n",
" lines = file.readlines()\n", " lines = file.readlines()\n",
"\n", "\n",
" results: List[schema.BaseMessage] = []\n", " results: List[BaseMessage] = []\n",
" current_sender = None\n", " current_sender = None\n",
" current_timestamp = None\n", " current_timestamp = None\n",
" current_content = []\n", " current_content = []\n",
@@ -292,7 +292,7 @@
"name": "python", "name": "python",
"nbconvert_exporter": "python", "nbconvert_exporter": "python",
"pygments_lexer": "ipython3", "pygments_lexer": "ipython3",
"version": "3.10.11" "version": "3.10.1"
} }
}, },
"nbformat": 4, "nbformat": 4,

View File

@@ -59,7 +59,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import AI21\n", "from langchain.llms import AI21\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -59,7 +59,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import AlephAlpha\n", "from langchain.llms import AlephAlpha\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -41,7 +41,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Anyscale\n", "from langchain.llms import Anyscale\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -154,7 +154,7 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"from langchain.llms.azureml_endpoint import DollyContentFormatter\n", "from langchain.llms.azureml_endpoint import DollyContentFormatter\n",
"from langchain.chains import LLMChain\n", "from langchain.chains import LLMChain\n",
"\n", "\n",

View File

@@ -53,7 +53,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Banana\n", "from langchain.llms import Banana\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -107,7 +107,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.chains import SimpleSequentialChain\n", "from langchain.chains import SimpleSequentialChain\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -80,7 +80,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"import langchain\n", "import langchain\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import NIBittensorLLM\n", "from langchain.llms import NIBittensorLLM\n",
"\n", "\n",
"langchain.debug = True\n", "langchain.debug = True\n",
@@ -123,7 +123,7 @@
" AgentExecutor,\n", " AgentExecutor,\n",
")\n", ")\n",
"from langchain.memory import ConversationBufferMemory\n", "from langchain.memory import ConversationBufferMemory\n",
"from langchain import LLMChain, PromptTemplate\n", "from langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.utilities import GoogleSearchAPIWrapper, SerpAPIWrapper\n", "from langchain.utilities import GoogleSearchAPIWrapper, SerpAPIWrapper\n",
"from langchain.llms import NIBittensorLLM\n", "from langchain.llms import NIBittensorLLM\n",
"\n", "\n",

View File

@@ -44,7 +44,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import CerebriumAI\n", "from langchain.llms import CerebriumAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -22,7 +22,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import ChatGLM\n", "from langchain.llms import ChatGLM\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"# import os" "# import os"
] ]

View File

@@ -82,7 +82,7 @@
"source": [ "source": [
"# Import the required modules\n", "# Import the required modules\n",
"from langchain.llms import Clarifai\n", "from langchain.llms import Clarifai\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -59,7 +59,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Cohere\n", "from langchain.llms import Cohere\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -102,7 +102,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"\"\"Question: {question}\n", "template = \"\"\"Question: {question}\n",
"\n", "\n",

View File

@@ -195,7 +195,7 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"\"\"{question}\n", "template = \"\"\"{question}\n",
"\n", "\n",

View File

@@ -28,7 +28,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import DeepInfra\n", "from langchain.llms import DeepInfra\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -103,7 +103,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"llm=EdenAI(feature=\"text\",provider=\"openai\",model=\"text-davinci-003\",temperature=0.2, max_tokens=250)\n", "llm=EdenAI(feature=\"text\",provider=\"openai\",model=\"text-davinci-003\",temperature=0.2, max_tokens=250)\n",
"\n", "\n",
"prompt = \"\"\"\n", "prompt = \"\"\"\n",

View File

@@ -20,7 +20,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms.fireworks import Fireworks, FireworksChat\n", "from langchain.llms.fireworks import Fireworks, FireworksChat\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.prompts.chat import (\n", "from langchain.prompts.chat import (\n",
" ChatPromptTemplate,\n", " ChatPromptTemplate,\n",
" HumanMessagePromptTemplate,\n", " HumanMessagePromptTemplate,\n",

View File

@@ -27,7 +27,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import ForefrontAI\n", "from langchain.llms import ForefrontAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -66,7 +66,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -43,7 +43,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import GooseAI\n", "from langchain.llms import GooseAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -47,7 +47,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import GPT4All\n", "from langchain.llms import GPT4All\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler" "from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler"
] ]

View File

@@ -91,7 +91,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import HuggingFaceHub" "from langchain.llms import HuggingFaceHub"
] ]
}, },
{ {
@@ -101,7 +101,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -189,7 +189,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import LlamaCpp\n", "from langchain.llms import LlamaCpp\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.callbacks.manager import CallbackManager\n", "from langchain.callbacks.manager import CallbackManager\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler" "from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler"
] ]

View File

@@ -80,7 +80,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"# Map reduce example\n", "# Map reduce example\n",
"from langchain import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"from langchain.text_splitter import CharacterTextSplitter\n", "from langchain.text_splitter import CharacterTextSplitter\n",
"from langchain.chains.mapreduce import MapReduceChain\n", "from langchain.chains.mapreduce import MapReduceChain\n",
"\n", "\n",

View File

@@ -94,7 +94,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Minimax\n", "from langchain.llms import Minimax\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
], ],
"metadata": { "metadata": {
"collapsed": false "collapsed": false

View File

@@ -108,7 +108,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Modal\n", "from langchain.llms import Modal\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -43,7 +43,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import MosaicML\n", "from langchain.llms import MosaicML\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -73,7 +73,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import NLPCloud\n", "from langchain.llms import NLPCloud\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -43,7 +43,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms.octoai_endpoint import OctoAIEndpoint\n", "from langchain.llms.octoai_endpoint import OctoAIEndpoint\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -206,7 +206,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"\n", "\n",
"# Prompt\n", "# Prompt\n",
"template = \"\"\"Use the following pieces of context to answer the question at the end. \n", "template = \"\"\"Use the following pieces of context to answer the question at the end. \n",

View File

@@ -59,7 +59,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"import langchain\n", "import langchain\n",
"from langchain import LLMChain, PromptTemplate\n", "from langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.callbacks.stdout import StdOutCallbackHandler\n", "from langchain.callbacks.stdout import StdOutCallbackHandler\n",
"from langchain.llms import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.memory import ConversationBufferWindowMemory\n", "from langchain.memory import ConversationBufferWindowMemory\n",

View File

@@ -67,7 +67,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -114,7 +114,7 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"What is a good name for a company that makes {product}?\"\n", "template = \"What is a good name for a company that makes {product}?\"\n",
"\n", "\n",

View File

@@ -71,7 +71,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import OpenLM\n", "from langchain.llms import OpenLM\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -45,7 +45,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import Petals\n", "from langchain.llms import Petals\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -50,7 +50,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from langchain.llms import PipelineAI\n", "from langchain.llms import PipelineAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -32,7 +32,7 @@
"\n", "\n",
"import predictionguard as pg\n", "import predictionguard as pg\n",
"from langchain.llms import PredictionGuard\n", "from langchain.llms import PredictionGuard\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
], ],
"id": "7191a5ce" "id": "7191a5ce"
}, },

View File

@@ -104,7 +104,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Replicate\n", "from langchain.llms import Replicate\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -44,7 +44,7 @@
], ],
"source": [ "source": [
"from langchain.llms import SelfHostedPipeline, SelfHostedHuggingFaceLLM\n", "from langchain.llms import SelfHostedPipeline, SelfHostedHuggingFaceLLM\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"import runhouse as rh" "import runhouse as rh"
] ]
}, },

View File

@@ -92,7 +92,7 @@
"source": [ "source": [
"from typing import Dict\n", "from typing import Dict\n",
"\n", "\n",
"from langchain import PromptTemplate, SagemakerEndpoint\n", "from langchain.prompts import PromptTemplate\nfrom langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n", "from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains.question_answering import load_qa_chain\n", "from langchain.chains.question_answering import load_qa_chain\n",
"import json\n", "import json\n",

View File

@@ -80,7 +80,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import StochasticAI\n", "from langchain.llms import StochasticAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -54,7 +54,7 @@
"execution_count": null, "execution_count": null,
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"conversation = \"\"\"Sam: Good morning, team! Let's keep this standup concise. We'll go in the usual order: what you did yesterday, what you plan to do today, and any blockers. Alex, kick us off.\n", "conversation = \"\"\"Sam: Good morning, team! Let's keep this standup concise. We'll go in the usual order: what you did yesterday, what you plan to do today, and any blockers. Alex, kick us off.\n",
"Alex: Morning! Yesterday, I wrapped up the UI for the user dashboard. The new charts and widgets are now responsive. I also had a sync with the design team to ensure the final touchups are in line with the brand guidelines. Today, I'll start integrating the frontend with the new API endpoints Rhea was working on. The only blocker is waiting for some final API documentation, but I guess Rhea can update on that.\n", "Alex: Morning! Yesterday, I wrapped up the UI for the user dashboard. The new charts and widgets are now responsive. I also had a sync with the design team to ensure the final touchups are in line with the brand guidelines. Today, I'll start integrating the frontend with the new API endpoints Rhea was working on. The only blocker is waiting for some final API documentation, but I guess Rhea can update on that.\n",

View File

@@ -44,7 +44,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"import langchain\n", "import langchain\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import TextGen\n", "from langchain.llms import TextGen\n",
"\n", "\n",
"langchain.debug = True\n", "langchain.debug = True\n",
@@ -93,7 +93,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"import langchain\n", "import langchain\n",
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import TextGen\n", "from langchain.llms import TextGen\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n", "from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n",
"\n", "\n",

View File

@@ -157,7 +157,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"llm = TitanTakeoff()\n", "llm = TitanTakeoff()\n",
"\n", "\n",

View File

@@ -76,7 +76,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Tongyi\n", "from langchain.llms import Tongyi\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -128,7 +128,7 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"\"\"Question: {question}\n", "template = \"\"\"Question: {question}\n",
"\n", "\n",

View File

@@ -56,7 +56,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import Writer\n", "from langchain.llms import Writer\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -122,7 +122,7 @@
} }
], ],
"source": [ "source": [
"from langchain import PromptTemplate, LLMChain\n", "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n", "\n",
"template = \"Where can we visit in the capital of {country}?\"\n", "template = \"Where can we visit in the capital of {country}?\"\n",
"\n", "\n",

View File

@@ -20,7 +20,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.memory.motorhead_memory import MotorheadMemory\n", "from langchain.memory.motorhead_memory import MotorheadMemory\n",
"from langchain import OpenAI, LLMChain, PromptTemplate\n", "from langchain.llms import OpenAI\nfrom langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"\n", "\n",
"template = \"\"\"You are a chatbot having a conversation with a human.\n", "template = \"\"\"You are a chatbot having a conversation with a human.\n",
"\n", "\n",

View File

@@ -21,7 +21,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.memory.motorhead_memory import MotorheadMemory\n", "from langchain.memory.motorhead_memory import MotorheadMemory\n",
"from langchain import OpenAI, LLMChain, PromptTemplate\n", "from langchain.llms import OpenAI\nfrom langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"\n", "\n",
"template = \"\"\"You are a chatbot having a conversation with a human.\n", "template = \"\"\"You are a chatbot having a conversation with a human.\n",
"\n", "\n",

View File

@@ -49,7 +49,7 @@
"source": [ "source": [
"from langchain.memory import ZepMemory\n", "from langchain.memory import ZepMemory\n",
"from langchain.retrievers import ZepRetriever\n", "from langchain.retrievers import ZepRetriever\n",
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.schema import HumanMessage, AIMessage\n", "from langchain.schema import HumanMessage, AIMessage\n",
"from langchain.utilities import WikipediaAPIWrapper\n", "from langchain.utilities import WikipediaAPIWrapper\n",
"from langchain.agents import initialize_agent, AgentType, Tool\n", "from langchain.agents import initialize_agent, AgentType, Tool\n",

View File

@@ -31,7 +31,7 @@ Args:
## Examples ## Examples
```python ```python
# Connecting to CnosDB with SQLDatabase Wrapper # Connecting to CnosDB with SQLDatabase Wrapper
from langchain import SQLDatabase from langchain.utilities import SQLDatabase
db = SQLDatabase.from_cnosdb() db = SQLDatabase.from_cnosdb()
``` ```
@@ -45,7 +45,7 @@ llm = ChatOpenAI(temperature=0, model_name="gpt-3.5-turbo")
### SQL Database Chain ### SQL Database Chain
This example demonstrates the use of the SQL Chain for answering a question over a CnosDB. This example demonstrates the use of the SQL Chain for answering a question over a CnosDB.
```python ```python
from langchain import SQLDatabaseChain from langchain.utilities import SQLDatabaseChain
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True) db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)

View File

@@ -80,7 +80,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"# Connecting to Databricks with SQLDatabase wrapper\n", "# Connecting to Databricks with SQLDatabase wrapper\n",
"from langchain import SQLDatabase\n", "from langchain.utilities import SQLDatabase\n",
"\n", "\n",
"db = SQLDatabase.from_databricks(catalog=\"samples\", schema=\"nyctaxi\")" "db = SQLDatabase.from_databricks(catalog=\"samples\", schema=\"nyctaxi\")"
] ]
@@ -115,7 +115,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SQLDatabaseChain\n", "from langchain.utilities import SQLDatabaseChain\n",
"\n", "\n",
"db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)" "db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)"
] ]

View File

@@ -37,7 +37,8 @@ from lcserve import serving
@serving @serving
def ask(input: str) -> str: def ask(input: str) -> str:
from langchain import LLMChain, OpenAI from langchain.chains import LLMChain
from langchain.llms import OpenAI
from langchain.agents import AgentExecutor, ZeroShotAgent from langchain.agents import AgentExecutor, ZeroShotAgent
tools = [...] # list of tools tools = [...] # list of tools

View File

@@ -40,7 +40,7 @@ llm = ChatOpenAI(model_name="gpt-3.5-turbo", callbacks=[log10_callback])
## How to use tags with Log10 ## How to use tags with Log10
```python ```python
from langchain import OpenAI from langchain.llms import OpenAI
from langchain.chat_models import ChatAnthropic from langchain.chat_models import ChatAnthropic
from langchain.chat_models import ChatOpenAI from langchain.chat_models import ChatOpenAI
from langchain.schema import HumanMessage from langchain.schema import HumanMessage
@@ -74,7 +74,7 @@ You can also intermix direct OpenAI calls and Langchain LLM calls:
import os import os
from log10.load import log10, log10_session from log10.load import log10, log10_session
import openai import openai
from langchain import OpenAI from langchain.llms import OpenAI
log10(openai) log10(openai)

View File

@@ -60,7 +60,7 @@ See the [API documentation and examples](https://www.mlflow.org/docs/latest/pyth
```python ```python
import mlflow import mlflow
from langchain import LLMChain, PromptTemplate from langchain.chains import LLMChain, PromptTemplate
from langchain.llms import MlflowAIGateway from langchain.llms import MlflowAIGateway
gateway = MlflowAIGateway( gateway = MlflowAIGateway(
@@ -134,7 +134,8 @@ Databricks MLflow AI Gateway is in private preview.
Please contact a Databricks representative to enroll in the preview. Please contact a Databricks representative to enroll in the preview.
```python ```python
from langchain import LLMChain, PromptTemplate from langchain.chains import LLMChain
from langchain.prompts import PromptTemplate
from langchain.llms import MlflowAIGateway from langchain.llms import MlflowAIGateway
gateway = MlflowAIGateway( gateway = MlflowAIGateway(

View File

@@ -26,7 +26,7 @@ conn_str = f"duckdb:///md:{token}@my_db"
You can use the SQLChain to query data in your Motherduck instance in natural language. You can use the SQLChain to query data in your Motherduck instance in natural language.
``` ```
from langchain import OpenAI, SQLDatabase, SQLDatabaseChain from langchain.llms import OpenAI, SQLDatabase, SQLDatabaseChain
db = SQLDatabase.from_uri(conn_str) db = SQLDatabase.from_uri(conn_str)
db_chain = SQLDatabaseChain.from_llm(OpenAI(temperature=0), db, verbose=True) db_chain = SQLDatabaseChain.from_llm(OpenAI(temperature=0), db, verbose=True)
``` ```

View File

@@ -37,7 +37,7 @@ import os
import predictionguard as pg import predictionguard as pg
from langchain.llms import PredictionGuard from langchain.llms import PredictionGuard
from langchain import PromptTemplate, LLMChain from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
# Your Prediction Guard API key. Get one at predictionguard.com # Your Prediction Guard API key. Get one at predictionguard.com
os.environ["PREDICTIONGUARD_TOKEN"] = "<your Prediction Guard access token>" os.environ["PREDICTIONGUARD_TOKEN"] = "<your Prediction Guard access token>"
@@ -76,7 +76,7 @@ Basic LLM Chaining with the Prediction Guard wrapper:
```python ```python
import os import os
from langchain import PromptTemplate, LLMChain from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
from langchain.llms import PredictionGuard from langchain.llms import PredictionGuard
# Optional, add your OpenAI API Key. This is optional, as Prediction Guard allows # Optional, add your OpenAI API Key. This is optional, as Prediction Guard allows

View File

@@ -108,7 +108,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.llms import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -20,7 +20,7 @@ As of June 2023, the API supports Vicuna-13B by default. We are going to support
For example For example
```python ```python
from langchain.llms import OpenAI from langchain.llms import OpenAI
from langchain import PromptTemplate, LLMChain from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
import os import os
os.environ['OPENAI_API_BASE'] = "https://shale.live/v1" os.environ['OPENAI_API_BASE'] = "https://shale.live/v1"

View File

@@ -204,7 +204,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"retriever = GoogleDriveRetriever(\n", "retriever = GoogleDriveRetriever(\n",
" template=PromptTemplate(input_variables=['query'],\n", " template=PromptTemplate(input_variables=['query'],\n",
" # See https://developers.google.com/drive/api/guides/search-files\n", " # See https://developers.google.com/drive/api/guides/search-files\n",

View File

@@ -150,7 +150,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import LLMChain\n", "from langchain.chains import LLMChain\n",
"from langchain.prompts import PromptTemplate\n", "from langchain.prompts import PromptTemplate\n",
"\n", "\n",
"QUERY_PROMPT = PromptTemplate(\n", "QUERY_PROMPT = PromptTemplate(\n",

View File

@@ -82,7 +82,7 @@
"source": [ "source": [
"# Import the required modules\n", "# Import the required modules\n",
"from langchain.embeddings import ClarifaiEmbeddings\n", "from langchain.embeddings import ClarifaiEmbeddings\n",
"from langchain import PromptTemplate, LLMChain" "from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -81,7 +81,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType" "from langchain.agents import initialize_agent, AgentType"
] ]
}, },

View File

@@ -105,7 +105,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType" "from langchain.agents import initialize_agent, AgentType"
] ]
}, },

View File

@@ -118,7 +118,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType" "from langchain.agents import initialize_agent, AgentType"
] ]
}, },

View File

@@ -167,7 +167,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n", "from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n", "llm = OpenAI(temperature=0)\n",
"agent = initialize_agent(\n", "agent = initialize_agent(\n",

View File

@@ -81,7 +81,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n", "from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n", "llm = OpenAI(temperature=0)\n",
"from langchain.agents.agent_toolkits import MultionToolkit\n", "from langchain.agents.agent_toolkits import MultionToolkit\n",

View File

@@ -93,7 +93,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType" "from langchain.agents import initialize_agent, AgentType"
] ]
}, },

View File

@@ -30,7 +30,7 @@
"from langchain.embeddings.openai import OpenAIEmbeddings\n", "from langchain.embeddings.openai import OpenAIEmbeddings\n",
"from langchain.vectorstores import Chroma\n", "from langchain.vectorstores import Chroma\n",
"from langchain.text_splitter import CharacterTextSplitter\n", "from langchain.text_splitter import CharacterTextSplitter\n",
"from langchain import OpenAI, VectorDBQA\n", "from langchain.llms import OpenAI\nfrom langchain.chains import VectorDBQA\n",
"\n", "\n",
"llm = OpenAI(temperature=0)" "llm = OpenAI(temperature=0)"
] ]

View File

@@ -61,7 +61,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import load_tools, initialize_agent, AgentType\n", "from langchain.agents import load_tools, initialize_agent, AgentType\n",
"\n", "\n",
"llm = OpenAI(temperature=0)\n", "llm = OpenAI(temperature=0)\n",

View File

@@ -126,7 +126,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType, load_tools" "from langchain.agents import initialize_agent, AgentType, load_tools"
] ]
}, },

View File

@@ -167,7 +167,7 @@
}, },
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n", "from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n", "llm = OpenAI(temperature=0)\n",
"agent = initialize_agent(\n", "agent = initialize_agent(\n",

View File

@@ -43,7 +43,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain.agents import load_tools, initialize_agent, AgentType\n", "from langchain.agents import load_tools, initialize_agent, AgentType\n",
"from langchain.utilities import GraphQLAPIWrapper\n", "from langchain.utilities import GraphQLAPIWrapper\n",
"\n", "\n",

View File

@@ -126,7 +126,7 @@
"source": [ "source": [
"import os\n", "import os\n",
"from lemonai import execute_workflow\n", "from lemonai import execute_workflow\n",
"from langchain import OpenAI" "from langchain.llms import OpenAI"
] ]
}, },
{ {

View File

@@ -474,7 +474,7 @@
], ],
"source": [ "source": [
"from langchain.chains import RetrievalQAWithSourcesChain\n", "from langchain.chains import RetrievalQAWithSourcesChain\n",
"from langchain import OpenAI\n", "from langchain.llms import OpenAI\n",
"\n", "\n",
"import os\n", "import os\n",
"import getpass\n", "import getpass\n",

View File

@@ -62,7 +62,7 @@
"from langchain.vectorstores.starrocks import StarRocksSettings\n", "from langchain.vectorstores.starrocks import StarRocksSettings\n",
"from langchain.vectorstores import Chroma\n", "from langchain.vectorstores import Chroma\n",
"from langchain.text_splitter import CharacterTextSplitter, TokenTextSplitter\n", "from langchain.text_splitter import CharacterTextSplitter, TokenTextSplitter\n",
"from langchain import OpenAI, VectorDBQA\n", "from langchain.llms import OpenAI\nfrom langchain.chains import VectorDBQA\n",
"from langchain.document_loaders import DirectoryLoader\n", "from langchain.document_loaders import DirectoryLoader\n",
"from langchain.chains import RetrievalQA\n", "from langchain.chains import RetrievalQA\n",
"from langchain.document_loaders import TextLoader, UnstructuredMarkdownLoader\n", "from langchain.document_loaders import TextLoader, UnstructuredMarkdownLoader\n",

View File

@@ -39,7 +39,7 @@
"\n", "\n",
"import os, sys, torch\n", "import os, sys, torch\n",
"from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel\n", "from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel\n",
"from langchain import HuggingFacePipeline, ConversationChain\n", "from langchain.llms import HuggingFacePipeline\nfrom langchain.chains import ConversationChain\n",
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n", "from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
"from langchain.vectorstores.vearch import VearchDb\n", "from langchain.vectorstores.vearch import VearchDb\n",
"from langchain.document_loaders import TextLoader\n", "from langchain.document_loaders import TextLoader\n",

View File

@@ -332,7 +332,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.chains import RetrievalQAWithSourcesChain\n", "from langchain.chains import RetrievalQAWithSourcesChain\n",
"from langchain import OpenAI" "from langchain.llms import OpenAI"
] ]
}, },
{ {

View File

@@ -23,7 +23,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import SerpAPIWrapper\n", "from langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n", "from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI" "from langchain.chat_models import ChatOpenAI"

View File

@@ -17,7 +17,7 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import OpenAI, Wikipedia\n", "from langchain.llms import OpenAI\from langchain.utilities import Wikipedia\n",
"from langchain.agents import initialize_agent, Tool\n", "from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"from langchain.agents.react.base import DocstoreExplorer\n", "from langchain.agents.react.base import DocstoreExplorer\n",

View File

@@ -45,7 +45,7 @@
} }
], ],
"source": [ "source": [
"from langchain import OpenAI, SerpAPIWrapper\n", "from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n", "from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"\n", "\n",

View File

@@ -17,13 +17,11 @@
"metadata": {}, "metadata": {},
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain import (\n", "from langchain.chains import LLMMathChain\n",
" LLMMathChain,\n", "from langchain.llms import OpenAI\n",
" OpenAI,\n", "from langchain.utilities import SerpAPIWrapper\n",
" SerpAPIWrapper,\n", "from langchain.utilities import SQLDatabase\n",
" SQLDatabase,\n", "from langchain_experimental.sql import SQLDatabaseChain\n",
" SQLDatabaseChain,\n",
")\n",
"from langchain.agents import initialize_agent, Tool\n", "from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI" "from langchain.chat_models import ChatOpenAI"

View File

@@ -29,7 +29,7 @@
"from langchain.agents import AgentExecutor, initialize_agent, AgentType\n", "from langchain.agents import AgentExecutor, initialize_agent, AgentType\n",
"from langchain.schema import AgentFinish\n", "from langchain.schema import AgentFinish\n",
"from langchain.agents.tools import Tool\n", "from langchain.agents.tools import Tool\n",
"from langchain import LLMMathChain\n", "from langchain.chains import LLMMathChain\n",
"from langchain.chat_models import ChatOpenAI" "from langchain.chat_models import ChatOpenAI"
] ]
}, },

View File

@@ -166,7 +166,7 @@
"from langchain.agents import AgentType\n", "from langchain.agents import AgentType\n",
"from langchain.tools import BaseTool\n", "from langchain.tools import BaseTool\n",
"from langchain.llms import OpenAI\n", "from langchain.llms import OpenAI\n",
"from langchain import LLMMathChain, SerpAPIWrapper" "from langchain.chains import LLMMathChain\nfrom langchain.utilities import SerpAPIWrapper"
] ]
}, },
{ {

View File

@@ -47,7 +47,7 @@
} }
], ],
"source": [ "source": [
"from langchain import OpenAI, ConversationChain, LLMChain, PromptTemplate\n", "from langchain.llms import OpenAI\nfrom langchain.chains import ConversationChain, LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.memory import ConversationBufferWindowMemory\n", "from langchain.memory import ConversationBufferWindowMemory\n",
"\n", "\n",
"\n", "\n",

View File

@@ -26,7 +26,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.agents import Tool, AgentExecutor, BaseSingleActionAgent\n", "from langchain.agents import Tool, AgentExecutor, BaseSingleActionAgent\n",
"from langchain import OpenAI, SerpAPIWrapper" "from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper"
] ]
}, },
{ {

View File

@@ -38,7 +38,7 @@
" AgentOutputParser,\n", " AgentOutputParser,\n",
")\n", ")\n",
"from langchain.prompts import StringPromptTemplate\n", "from langchain.prompts import StringPromptTemplate\n",
"from langchain import OpenAI, SerpAPIWrapper, LLMChain\n", "from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\nfrom langchain.chains import LLMChain\n",
"from typing import List, Union\n", "from typing import List, Union\n",
"from langchain.schema import AgentAction, AgentFinish\n", "from langchain.schema import AgentAction, AgentFinish\n",
"import re" "import re"

View File

@@ -48,7 +48,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.agents import ZeroShotAgent, Tool, AgentExecutor\n", "from langchain.agents import ZeroShotAgent, Tool, AgentExecutor\n",
"from langchain import OpenAI, SerpAPIWrapper, LLMChain" "from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\nfrom langchain.chains import LLMChain"
] ]
}, },
{ {

View File

@@ -26,7 +26,7 @@
"outputs": [], "outputs": [],
"source": [ "source": [
"from langchain.agents import Tool, AgentExecutor, BaseMultiActionAgent\n", "from langchain.agents import Tool, AgentExecutor, BaseMultiActionAgent\n",
"from langchain import OpenAI, SerpAPIWrapper" "from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper"
] ]
}, },
{ {

Some files were not shown because too many files have changed in this diff Show More