Compare commits

...

52 Commits

Author SHA1 Message Date
Harrison Chase
26d1cf468c remove from init 2023-09-16 17:17:05 -07:00
Harrison Chase
ccc7e97513 cr 2023-09-16 17:10:33 -07:00
Harrison Chase
d6e245e2e1 cr 2023-09-16 17:07:48 -07:00
Harrison Chase
d3aee0a181 Merge branch 'master' into harrison/stop-importing-from-init 2023-09-16 17:07:05 -07:00
Harrison Chase
e9483bfc88 cr 2023-09-16 17:02:54 -07:00
Harrison Chase
2f718319bb cr 2023-09-16 17:00:38 -07:00
Harrison Chase
40ad26f030 cr 2023-09-16 16:59:48 -07:00
Harrison Chase
4eb649da19 cr 2023-09-16 16:56:05 -07:00
Harrison Chase
71b80e3ba5 cr 2023-09-16 16:55:37 -07:00
Harrison Chase
3d2250820c cr 2023-09-16 16:51:56 -07:00
Harrison Chase
bcb64527ae Update docs/extras/use_cases/more/agents/autonomous_agents/baby_agi_with_agent.ipynb 2023-09-16 16:50:17 -07:00
Harrison Chase
6a0898f89d Update docs/extras/use_cases/more/agents/autonomous_agents/baby_agi_with_agent.ipynb 2023-09-16 16:50:08 -07:00
Harrison Chase
3f991afa15 Update docs/extras/use_cases/more/agents/autonomous_agents/baby_agi.ipynb 2023-09-16 16:49:58 -07:00
Harrison Chase
eb726fa333 Update docs/extras/use_cases/more/agents/agents/wikibase_agent.ipynb 2023-09-16 16:49:49 -07:00
Harrison Chase
9951a6cc75 Update docs/extras/use_cases/more/agents/agents/sales_agent_with_context.ipynb 2023-09-16 16:49:39 -07:00
Harrison Chase
447066163e Update docs/extras/use_cases/more/agents/agents/custom_agent_with_plugin_retrieval_using_plugnplai.ipynb 2023-09-16 16:49:30 -07:00
Harrison Chase
d908818b32 Update docs/extras/use_cases/more/agents/agents/custom_agent_with_plugin_retrieval.ipynb 2023-09-16 16:49:19 -07:00
Harrison Chase
0591854007 Update docs/extras/modules/memory/custom_memory.ipynb 2023-09-16 16:49:10 -07:00
Harrison Chase
f893a82efa Update docs/extras/modules/memory/agent_with_memory_in_db.ipynb 2023-09-16 16:49:02 -07:00
Harrison Chase
fc9e1bcf9d Update libs/experimental/langchain_experimental/autonomous_agents/hugginggpt/repsonse_generator.py 2023-09-16 16:48:51 -07:00
Harrison Chase
3b8f58edfa Update docs/snippets/modules/chains/foundational/llm_chain.mdx 2023-09-16 16:48:46 -07:00
Harrison Chase
baafcbacd9 Update docs/snippets/modules/agents/how_to/mrkl.mdx 2023-09-16 16:48:42 -07:00
Harrison Chase
641183a137 Update docs/snippets/modules/agents/how_to/custom_llm_agent.mdx 2023-09-16 16:48:37 -07:00
Harrison Chase
c8982b1a82 Update docs/extras/use_cases/qa_structured/integrations/myscale_vector_sql.ipynb 2023-09-16 16:48:32 -07:00
Harrison Chase
262b45522d Update docs/extras/use_cases/more/code_writing/llm_math.ipynb 2023-09-16 16:48:20 -07:00
Harrison Chase
5083e7b1e2 Update docs/extras/use_cases/more/agents/autonomous_agents/meta_prompt.ipynb 2023-09-16 16:48:15 -07:00
Harrison Chase
d52a853518 Update docs/extras/modules/memory/agent_with_memory.ipynb 2023-09-16 16:48:10 -07:00
Harrison Chase
3ead58d1c1 Update docs/extras/modules/chains/how_to/serialization.ipynb 2023-09-16 16:47:59 -07:00
Harrison Chase
620be7322f Update docs/extras/modules/chains/how_to/from_hub.ipynb 2023-09-16 16:47:53 -07:00
Harrison Chase
54a0900093 Update docs/extras/modules/agents/tools/custom_tools.ipynb 2023-09-16 16:47:47 -07:00
Harrison Chase
447517e125 Update docs/extras/modules/agents/tools/custom_tools.ipynb 2023-09-16 16:47:42 -07:00
Harrison Chase
563d7806f5 Update docs/extras/modules/agents/how_to/sharedmemory_for_tools.ipynb 2023-09-16 16:47:35 -07:00
Harrison Chase
1055a74dc4 Update docs/extras/modules/agents/how_to/custom_multi_action_agent.ipynb 2023-09-16 16:47:30 -07:00
Harrison Chase
d6663aecec Update docs/extras/modules/agents/how_to/custom_mrkl_agent.ipynb 2023-09-16 16:47:24 -07:00
Harrison Chase
7b692d22bf Update docs/extras/modules/agents/how_to/custom_agent_with_tool_retrieval.ipynb 2023-09-16 16:46:08 -07:00
Harrison Chase
2bf326b77d Update docs/extras/modules/agents/how_to/custom_agent.ipynb 2023-09-16 16:46:02 -07:00
Harrison Chase
e602c5ea95 Update docs/extras/modules/agents/how_to/chatgpt_clone.ipynb 2023-09-16 16:45:53 -07:00
Harrison Chase
6a7c07010d Update docs/extras/modules/agents/how_to/agent_vectorstore.ipynb 2023-09-16 16:45:46 -07:00
Harrison Chase
6609750538 Update docs/extras/modules/agents/agent_types/self_ask_with_search.ipynb 2023-09-16 16:45:39 -07:00
Harrison Chase
c9432e2055 Update docs/extras/modules/agents/agent_types/react_docstore.ipynb 2023-09-16 16:45:32 -07:00
Harrison Chase
865cd63f81 Update docs/extras/integrations/vectorstores/starrocks.ipynb 2023-09-16 16:45:25 -07:00
Harrison Chase
2d3064e6fd Update docs/extras/integrations/toolkits/vectorstore.ipynb 2023-09-16 16:45:18 -07:00
Harrison Chase
87838d0fcd Update docs/extras/integrations/providers/jina.mdx 2023-09-16 16:45:11 -07:00
Harrison Chase
0beb465adb Update docs/extras/integrations/providers/mlflow_ai_gateway.mdx 2023-09-16 16:45:01 -07:00
Harrison Chase
ffa69882c0 Update docs/extras/integrations/memory/motorhead_memory_managed.ipynb 2023-09-16 16:44:53 -07:00
Harrison Chase
c61a05d6b0 Update docs/extras/integrations/memory/motorhead_memory.ipynb 2023-09-16 16:44:46 -07:00
Harrison Chase
f2a56d733c Update docs/extras/integrations/llms/sagemaker.ipynb 2023-09-16 16:44:39 -07:00
Harrison Chase
10e419d84e Update docs/extras/integrations/llms/opaqueprompts.ipynb 2023-09-16 16:44:24 -07:00
Harrison Chase
284b9ae850 Update docs/extras/integrations/llms/bittensor.ipynb 2023-09-16 16:44:14 -07:00
Harrison Chase
bc7535a9c7 Update docs/extras/guides/model_laboratory.ipynb 2023-09-16 15:49:33 -07:00
Harrison Chase
50e9bfce9f stop importing from init 2023-09-16 15:47:54 -07:00
Harrison Chase
f726b428ad stash 2023-09-16 15:41:51 -07:00
203 changed files with 323 additions and 433 deletions

23
.github/workflows/imports.yml vendored Normal file
View File

@@ -0,0 +1,23 @@
---
name: Imports
on:
push:
branches: [master]
pull_request:
branches: [master]
jobs:
check:
runs-on: ubuntu-latest
steps:
- name: Checkout repository
uses: actions/checkout@v2
- name: Run import check
run: |
# We should not encourage imports directly from main init file
# Expect for __version__ and hub
# And of course expect for this file
git grep 'from langchain import' | grep -vE 'from langchain import (__version__|hub)' | grep -v '.github/workflows/check-imports.yml' && exit 1 || exit 0

View File

@@ -105,7 +105,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n",
"from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n",
"\n",
@@ -412,7 +412,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n",
"\n",
"template = \"\"\"Question: {question}\n",
@@ -572,8 +572,8 @@
},
"outputs": [],
"source": [
"from langchain import HuggingFaceHub\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.llms import HuggingFaceHub\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"\"\"Question: {question}\"\"\"\n",
"\n",
@@ -697,7 +697,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import SagemakerEndpoint\n",
"from langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains import LLMChain\n",
"from langchain.prompts import load_prompt, PromptTemplate\n",

View File

@@ -97,7 +97,7 @@
},
"outputs": [],
"source": [
"from langchain import SerpAPIWrapper\n",
"from langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI\n",

View File

@@ -468,7 +468,8 @@
}
],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\n",
"from langchain.chains import LLMChain\n",
"from langchain.chains.prompt_selector import ConditionalPromptSelector\n",
"\n",
"DEFAULT_LLAMA_SEARCH_PROMPT = PromptTemplate(\n",
@@ -593,7 +594,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.9.1"
"version": "3.10.1"
}
},
"nbformat": 4,

View File

@@ -19,7 +19,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import LLMChain, OpenAI, Cohere, HuggingFaceHub, PromptTemplate\n",
"from langchain.chains import LLMChain\nfrom langchain.llms import OpenAI, Cohere, HuggingFaceHub\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.model_laboratory import ModelLaboratory"
]
},
@@ -139,7 +139,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import SelfAskWithSearchChain, SerpAPIWrapper\n",
"from langchain.chains import SelfAskWithSearchChain\nfrom langchain.utilities import SerpAPIWrapper\n",
"\n",
"open_ai_llm = OpenAI(temperature=0)\n",
"search = SerpAPIWrapper()\n",

View File

@@ -95,7 +95,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n",
"from langchain_experimental.comprehend_moderation.base_moderation_exceptions import ModerationPiiError\n",
"\n",
@@ -399,7 +399,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms.fake import FakeListLLM\n",
"\n",
"template = \"\"\"Question: {question}\n",
@@ -564,8 +564,8 @@
},
"outputs": [],
"source": [
"from langchain import HuggingFaceHub\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.llms import HuggingFaceHub\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"\"\"Question: {question}\n",
"\n",
@@ -679,7 +679,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import SagemakerEndpoint\n",
"from langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains import LLMChain\n",
"from langchain.prompts import load_prompt, PromptTemplate\n",

View File

@@ -123,7 +123,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType"
]
},

View File

@@ -167,7 +167,7 @@
"import os\n",
"\n",
"from langchain.chat_models import ChatOpenAI\n",
"from langchain import LLMChain\n",
"from langchain.chains import LLMChain\n",
"from langchain.prompts import PromptTemplate\n",
"from langchain.prompts.chat import (\n",
" ChatPromptTemplate,\n",

View File

@@ -199,7 +199,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate\n",
"from langchain.prompts import PromptTemplate\n",
"\n",
"# Prompt\n",
"template = \"\"\"[INST] <<SYS>> Use the following pieces of context to answer the question at the end. \n",

View File

@@ -81,7 +81,7 @@
"import re\n",
"from typing import Iterator, List\n",
"\n",
"from langchain import schema\n",
"from langchain.schema import BaseMessage, HumanMessage\n",
"from langchain.chat_loaders import base as chat_loaders\n",
"\n",
"logger = logging.getLogger()\n",
@@ -117,7 +117,7 @@
" with open(file_path, \"r\", encoding=\"utf-8\") as file:\n",
" lines = file.readlines()\n",
"\n",
" results: List[schema.BaseMessage] = []\n",
" results: List[BaseMessage] = []\n",
" current_sender = None\n",
" current_timestamp = None\n",
" current_content = []\n",
@@ -128,7 +128,7 @@
" ):\n",
" if current_sender and current_content:\n",
" results.append(\n",
" schema.HumanMessage(\n",
" HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n",
" \"sender\": current_sender,\n",
@@ -142,7 +142,7 @@
" ]\n",
" elif re.match(r\"\\[\\d{1,2}:\\d{2} (?:AM|PM)\\]\", line.strip()):\n",
" results.append(\n",
" schema.HumanMessage(\n",
" HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n",
" \"sender\": current_sender,\n",
@@ -157,7 +157,7 @@
"\n",
" if current_sender and current_content:\n",
" results.append(\n",
" schema.HumanMessage(\n",
" HumanMessage(\n",
" content=\"\".join(current_content).strip(),\n",
" additional_kwargs={\n",
" \"sender\": current_sender,\n",

View File

@@ -78,7 +78,7 @@
"import re\n",
"from typing import Iterator, List\n",
"\n",
"from langchain import schema\n",
"from langchain.schema import HumanMessage, BaseMessage\n",
"from langchain.chat_loaders import base as chat_loaders\n",
"\n",
"logger = logging.getLogger()\n",
@@ -110,7 +110,7 @@
" # skip non-text messages like stickers, images, etc.\n",
" if not re.match(r\"\\[.*\\]\", content):\n",
" results.append(\n",
" schema.HumanMessage(\n",
" HumanMessage(\n",
" content=content,\n",
" additional_kwargs={\n",
" \"sender\": current_sender,\n",
@@ -135,7 +135,7 @@
" with open(file_path, \"r\", encoding=\"utf-8\") as file:\n",
" lines = file.readlines()\n",
"\n",
" results: List[schema.BaseMessage] = []\n",
" results: List[BaseMessage] = []\n",
" current_sender = None\n",
" current_timestamp = None\n",
" current_content = []\n",
@@ -292,7 +292,7 @@
"name": "python",
"nbconvert_exporter": "python",
"pygments_lexer": "ipython3",
"version": "3.10.11"
"version": "3.10.1"
}
},
"nbformat": 4,

View File

@@ -59,7 +59,7 @@
"outputs": [],
"source": [
"from langchain.llms import AI21\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -59,7 +59,7 @@
"outputs": [],
"source": [
"from langchain.llms import AlephAlpha\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -41,7 +41,7 @@
"outputs": [],
"source": [
"from langchain.llms import Anyscale\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -154,7 +154,7 @@
}
],
"source": [
"from langchain import PromptTemplate\n",
"from langchain.prompts import PromptTemplate\n",
"from langchain.llms.azureml_endpoint import DollyContentFormatter\n",
"from langchain.chains import LLMChain\n",
"\n",

View File

@@ -53,7 +53,7 @@
"outputs": [],
"source": [
"from langchain.llms import Banana\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -107,7 +107,7 @@
"outputs": [],
"source": [
"from langchain.chains import SimpleSequentialChain\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -80,7 +80,7 @@
"outputs": [],
"source": [
"import langchain\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import NIBittensorLLM\n",
"\n",
"langchain.debug = True\n",
@@ -123,7 +123,7 @@
" AgentExecutor,\n",
")\n",
"from langchain.memory import ConversationBufferMemory\n",
"from langchain import LLMChain, PromptTemplate\n",
"from langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.utilities import GoogleSearchAPIWrapper, SerpAPIWrapper\n",
"from langchain.llms import NIBittensorLLM\n",
"\n",

View File

@@ -44,7 +44,7 @@
"source": [
"import os\n",
"from langchain.llms import CerebriumAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -22,7 +22,7 @@
"outputs": [],
"source": [
"from langchain.llms import ChatGLM\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"# import os"
]

View File

@@ -82,7 +82,7 @@
"source": [
"# Import the required modules\n",
"from langchain.llms import Clarifai\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -59,7 +59,7 @@
"outputs": [],
"source": [
"from langchain.llms import Cohere\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -102,7 +102,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"\"\"Question: {question}\n",
"\n",

View File

@@ -195,7 +195,7 @@
}
],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"\"\"{question}\n",
"\n",

View File

@@ -28,7 +28,7 @@
"source": [
"import os\n",
"from langchain.llms import DeepInfra\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -103,7 +103,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"llm=EdenAI(feature=\"text\",provider=\"openai\",model=\"text-davinci-003\",temperature=0.2, max_tokens=250)\n",
"\n",
"prompt = \"\"\"\n",

View File

@@ -20,7 +20,7 @@
"outputs": [],
"source": [
"from langchain.llms.fireworks import Fireworks, FireworksChat\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.prompts.chat import (\n",
" ChatPromptTemplate,\n",
" HumanMessagePromptTemplate,\n",

View File

@@ -27,7 +27,7 @@
"source": [
"import os\n",
"from langchain.llms import ForefrontAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -66,7 +66,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -43,7 +43,7 @@
"source": [
"import os\n",
"from langchain.llms import GooseAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -47,7 +47,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import GPT4All\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler"
]

View File

@@ -91,7 +91,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import HuggingFaceHub"
"from langchain.llms import HuggingFaceHub"
]
},
{
@@ -101,7 +101,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -189,7 +189,7 @@
"outputs": [],
"source": [
"from langchain.llms import LlamaCpp\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.callbacks.manager import CallbackManager\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler"
]

View File

@@ -80,7 +80,7 @@
"outputs": [],
"source": [
"# Map reduce example\n",
"from langchain import PromptTemplate\n",
"from langchain.prompts import PromptTemplate\n",
"from langchain.text_splitter import CharacterTextSplitter\n",
"from langchain.chains.mapreduce import MapReduceChain\n",
"\n",

View File

@@ -94,7 +94,7 @@
"outputs": [],
"source": [
"from langchain.llms import Minimax\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
],
"metadata": {
"collapsed": false

View File

@@ -108,7 +108,7 @@
"outputs": [],
"source": [
"from langchain.llms import Modal\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -43,7 +43,7 @@
"outputs": [],
"source": [
"from langchain.llms import MosaicML\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -73,7 +73,7 @@
"outputs": [],
"source": [
"from langchain.llms import NLPCloud\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -43,7 +43,7 @@
"outputs": [],
"source": [
"from langchain.llms.octoai_endpoint import OctoAIEndpoint\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -206,7 +206,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate\n",
"from langchain.prompts import PromptTemplate\n",
"\n",
"# Prompt\n",
"template = \"\"\"Use the following pieces of context to answer the question at the end. \n",

View File

@@ -59,7 +59,7 @@
"outputs": [],
"source": [
"import langchain\n",
"from langchain import LLMChain, PromptTemplate\n",
"from langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.callbacks.stdout import StdOutCallbackHandler\n",
"from langchain.llms import OpenAI\n",
"from langchain.memory import ConversationBufferWindowMemory\n",

View File

@@ -67,7 +67,7 @@
"outputs": [],
"source": [
"from langchain.llms import OpenAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -114,7 +114,7 @@
}
],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"What is a good name for a company that makes {product}?\"\n",
"\n",

View File

@@ -71,7 +71,7 @@
"outputs": [],
"source": [
"from langchain.llms import OpenLM\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -45,7 +45,7 @@
"source": [
"import os\n",
"from langchain.llms import Petals\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -50,7 +50,7 @@
"source": [
"import os\n",
"from langchain.llms import PipelineAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -32,7 +32,7 @@
"\n",
"import predictionguard as pg\n",
"from langchain.llms import PredictionGuard\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
],
"id": "7191a5ce"
},

View File

@@ -104,7 +104,7 @@
"outputs": [],
"source": [
"from langchain.llms import Replicate\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -44,7 +44,7 @@
],
"source": [
"from langchain.llms import SelfHostedPipeline, SelfHostedHuggingFaceLLM\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"import runhouse as rh"
]
},

View File

@@ -92,7 +92,7 @@
"source": [
"from typing import Dict\n",
"\n",
"from langchain import PromptTemplate, SagemakerEndpoint\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.llms import SagemakerEndpoint\n",
"from langchain.llms.sagemaker_endpoint import LLMContentHandler\n",
"from langchain.chains.question_answering import load_qa_chain\n",
"import json\n",

View File

@@ -80,7 +80,7 @@
"outputs": [],
"source": [
"from langchain.llms import StochasticAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -54,7 +54,7 @@
"execution_count": null,
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"conversation = \"\"\"Sam: Good morning, team! Let's keep this standup concise. We'll go in the usual order: what you did yesterday, what you plan to do today, and any blockers. Alex, kick us off.\n",
"Alex: Morning! Yesterday, I wrapped up the UI for the user dashboard. The new charts and widgets are now responsive. I also had a sync with the design team to ensure the final touchups are in line with the brand guidelines. Today, I'll start integrating the frontend with the new API endpoints Rhea was working on. The only blocker is waiting for some final API documentation, but I guess Rhea can update on that.\n",

View File

@@ -44,7 +44,7 @@
"outputs": [],
"source": [
"import langchain\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import TextGen\n",
"\n",
"langchain.debug = True\n",
@@ -93,7 +93,7 @@
"outputs": [],
"source": [
"import langchain\n",
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"from langchain.llms import TextGen\n",
"from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n",
"\n",

View File

@@ -157,7 +157,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"llm = TitanTakeoff()\n",
"\n",

View File

@@ -76,7 +76,7 @@
"outputs": [],
"source": [
"from langchain.llms import Tongyi\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -128,7 +128,7 @@
}
],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"\"\"Question: {question}\n",
"\n",

View File

@@ -56,7 +56,7 @@
"outputs": [],
"source": [
"from langchain.llms import Writer\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -122,7 +122,7 @@
}
],
"source": [
"from langchain import PromptTemplate, LLMChain\n",
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain\n",
"\n",
"template = \"Where can we visit in the capital of {country}?\"\n",
"\n",

View File

@@ -20,7 +20,7 @@
"outputs": [],
"source": [
"from langchain.memory.motorhead_memory import MotorheadMemory\n",
"from langchain import OpenAI, LLMChain, PromptTemplate\n",
"from langchain.llms import OpenAI\nfrom langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"\n",
"template = \"\"\"You are a chatbot having a conversation with a human.\n",
"\n",

View File

@@ -21,7 +21,7 @@
"outputs": [],
"source": [
"from langchain.memory.motorhead_memory import MotorheadMemory\n",
"from langchain import OpenAI, LLMChain, PromptTemplate\n",
"from langchain.llms import OpenAI\nfrom langchain.chains import LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"\n",
"template = \"\"\"You are a chatbot having a conversation with a human.\n",
"\n",

View File

@@ -49,7 +49,7 @@
"source": [
"from langchain.memory import ZepMemory\n",
"from langchain.retrievers import ZepRetriever\n",
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.schema import HumanMessage, AIMessage\n",
"from langchain.utilities import WikipediaAPIWrapper\n",
"from langchain.agents import initialize_agent, AgentType, Tool\n",

View File

@@ -31,7 +31,7 @@ Args:
## Examples
```python
# Connecting to CnosDB with SQLDatabase Wrapper
from langchain import SQLDatabase
from langchain.utilities import SQLDatabase
db = SQLDatabase.from_cnosdb()
```
@@ -45,7 +45,7 @@ llm = ChatOpenAI(temperature=0, model_name="gpt-3.5-turbo")
### SQL Database Chain
This example demonstrates the use of the SQL Chain for answering a question over a CnosDB.
```python
from langchain import SQLDatabaseChain
from langchain.utilities import SQLDatabaseChain
db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)

View File

@@ -80,7 +80,7 @@
"outputs": [],
"source": [
"# Connecting to Databricks with SQLDatabase wrapper\n",
"from langchain import SQLDatabase\n",
"from langchain.utilities import SQLDatabase\n",
"\n",
"db = SQLDatabase.from_databricks(catalog=\"samples\", schema=\"nyctaxi\")"
]
@@ -115,7 +115,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import SQLDatabaseChain\n",
"from langchain.utilities import SQLDatabaseChain\n",
"\n",
"db_chain = SQLDatabaseChain.from_llm(llm, db, verbose=True)"
]

View File

@@ -37,7 +37,8 @@ from lcserve import serving
@serving
def ask(input: str) -> str:
from langchain import LLMChain, OpenAI
from langchain.chains import LLMChain
from langchain.llms import OpenAI
from langchain.agents import AgentExecutor, ZeroShotAgent
tools = [...] # list of tools

View File

@@ -40,7 +40,7 @@ llm = ChatOpenAI(model_name="gpt-3.5-turbo", callbacks=[log10_callback])
## How to use tags with Log10
```python
from langchain import OpenAI
from langchain.llms import OpenAI
from langchain.chat_models import ChatAnthropic
from langchain.chat_models import ChatOpenAI
from langchain.schema import HumanMessage
@@ -74,7 +74,7 @@ You can also intermix direct OpenAI calls and Langchain LLM calls:
import os
from log10.load import log10, log10_session
import openai
from langchain import OpenAI
from langchain.llms import OpenAI
log10(openai)

View File

@@ -60,7 +60,7 @@ See the [API documentation and examples](https://www.mlflow.org/docs/latest/pyth
```python
import mlflow
from langchain import LLMChain, PromptTemplate
from langchain.chains import LLMChain, PromptTemplate
from langchain.llms import MlflowAIGateway
gateway = MlflowAIGateway(
@@ -134,7 +134,8 @@ Databricks MLflow AI Gateway is in private preview.
Please contact a Databricks representative to enroll in the preview.
```python
from langchain import LLMChain, PromptTemplate
from langchain.chains import LLMChain
from langchain.prompts import PromptTemplate
from langchain.llms import MlflowAIGateway
gateway = MlflowAIGateway(

View File

@@ -26,7 +26,7 @@ conn_str = f"duckdb:///md:{token}@my_db"
You can use the SQLChain to query data in your Motherduck instance in natural language.
```
from langchain import OpenAI, SQLDatabase, SQLDatabaseChain
from langchain.llms import OpenAI, SQLDatabase, SQLDatabaseChain
db = SQLDatabase.from_uri(conn_str)
db_chain = SQLDatabaseChain.from_llm(OpenAI(temperature=0), db, verbose=True)
```

View File

@@ -37,7 +37,7 @@ import os
import predictionguard as pg
from langchain.llms import PredictionGuard
from langchain import PromptTemplate, LLMChain
from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
# Your Prediction Guard API key. Get one at predictionguard.com
os.environ["PREDICTIONGUARD_TOKEN"] = "<your Prediction Guard access token>"
@@ -76,7 +76,7 @@ Basic LLM Chaining with the Prediction Guard wrapper:
```python
import os
from langchain import PromptTemplate, LLMChain
from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
from langchain.llms import PredictionGuard
# Optional, add your OpenAI API Key. This is optional, as Prediction Guard allows

View File

@@ -108,7 +108,7 @@
"outputs": [],
"source": [
"from langchain.llms import OpenAI\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -20,7 +20,7 @@ As of June 2023, the API supports Vicuna-13B by default. We are going to support
For example
```python
from langchain.llms import OpenAI
from langchain import PromptTemplate, LLMChain
from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain
import os
os.environ['OPENAI_API_BASE'] = "https://shale.live/v1"

View File

@@ -204,7 +204,7 @@
},
"outputs": [],
"source": [
"from langchain import PromptTemplate\n",
"from langchain.prompts import PromptTemplate\n",
"retriever = GoogleDriveRetriever(\n",
" template=PromptTemplate(input_variables=['query'],\n",
" # See https://developers.google.com/drive/api/guides/search-files\n",

View File

@@ -150,7 +150,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import LLMChain\n",
"from langchain.chains import LLMChain\n",
"from langchain.prompts import PromptTemplate\n",
"\n",
"QUERY_PROMPT = PromptTemplate(\n",

View File

@@ -82,7 +82,7 @@
"source": [
"# Import the required modules\n",
"from langchain.embeddings import ClarifaiEmbeddings\n",
"from langchain import PromptTemplate, LLMChain"
"from langchain.prompts import PromptTemplate\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -81,7 +81,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType"
]
},

View File

@@ -105,7 +105,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType"
]
},

View File

@@ -118,7 +118,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType"
]
},

View File

@@ -167,7 +167,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n",
"agent = initialize_agent(\n",

View File

@@ -81,7 +81,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n",
"from langchain.agents.agent_toolkits import MultionToolkit\n",

View File

@@ -93,7 +93,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType"
]
},

View File

@@ -30,7 +30,7 @@
"from langchain.embeddings.openai import OpenAIEmbeddings\n",
"from langchain.vectorstores import Chroma\n",
"from langchain.text_splitter import CharacterTextSplitter\n",
"from langchain import OpenAI, VectorDBQA\n",
"from langchain.llms import OpenAI\nfrom langchain.chains import VectorDBQA\n",
"\n",
"llm = OpenAI(temperature=0)"
]

View File

@@ -61,7 +61,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import load_tools, initialize_agent, AgentType\n",
"\n",
"llm = OpenAI(temperature=0)\n",

View File

@@ -126,7 +126,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType, load_tools"
]
},

View File

@@ -167,7 +167,7 @@
},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import initialize_agent, AgentType\n",
"llm = OpenAI(temperature=0)\n",
"agent = initialize_agent(\n",

View File

@@ -43,7 +43,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"from langchain.agents import load_tools, initialize_agent, AgentType\n",
"from langchain.utilities import GraphQLAPIWrapper\n",
"\n",

View File

@@ -126,7 +126,7 @@
"source": [
"import os\n",
"from lemonai import execute_workflow\n",
"from langchain import OpenAI"
"from langchain.llms import OpenAI"
]
},
{

View File

@@ -474,7 +474,7 @@
],
"source": [
"from langchain.chains import RetrievalQAWithSourcesChain\n",
"from langchain import OpenAI\n",
"from langchain.llms import OpenAI\n",
"\n",
"import os\n",
"import getpass\n",

View File

@@ -62,7 +62,7 @@
"from langchain.vectorstores.starrocks import StarRocksSettings\n",
"from langchain.vectorstores import Chroma\n",
"from langchain.text_splitter import CharacterTextSplitter, TokenTextSplitter\n",
"from langchain import OpenAI, VectorDBQA\n",
"from langchain.llms import OpenAI\nfrom langchain.chains import VectorDBQA\n",
"from langchain.document_loaders import DirectoryLoader\n",
"from langchain.chains import RetrievalQA\n",
"from langchain.document_loaders import TextLoader, UnstructuredMarkdownLoader\n",

View File

@@ -39,7 +39,7 @@
"\n",
"import os, sys, torch\n",
"from transformers import AutoTokenizer, AutoModelForCausalLM, AutoModel\n",
"from langchain import HuggingFacePipeline, ConversationChain\n",
"from langchain.llms import HuggingFacePipeline\nfrom langchain.chains import ConversationChain\n",
"from langchain.text_splitter import RecursiveCharacterTextSplitter\n",
"from langchain.vectorstores.vearch import VearchDb\n",
"from langchain.document_loaders import TextLoader\n",

View File

@@ -332,7 +332,7 @@
"outputs": [],
"source": [
"from langchain.chains import RetrievalQAWithSourcesChain\n",
"from langchain import OpenAI"
"from langchain.llms import OpenAI"
]
},
{

View File

@@ -23,7 +23,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import SerpAPIWrapper\n",
"from langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI"

View File

@@ -17,7 +17,7 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import OpenAI, Wikipedia\n",
"from langchain.llms import OpenAI\from langchain.utilities import Wikipedia\n",
"from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n",
"from langchain.agents.react.base import DocstoreExplorer\n",

View File

@@ -45,7 +45,7 @@
}
],
"source": [
"from langchain import OpenAI, SerpAPIWrapper\n",
"from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\n",
"from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n",
"\n",

View File

@@ -17,13 +17,11 @@
"metadata": {},
"outputs": [],
"source": [
"from langchain import (\n",
" LLMMathChain,\n",
" OpenAI,\n",
" SerpAPIWrapper,\n",
" SQLDatabase,\n",
" SQLDatabaseChain,\n",
")\n",
"from langchain.chains import LLMMathChain\n",
"from langchain.llms import OpenAI\n",
"from langchain.utilities import SerpAPIWrapper\n",
"from langchain.utilities import SQLDatabase\n",
"from langchain_experimental.sql import SQLDatabaseChain\n",
"from langchain.agents import initialize_agent, Tool\n",
"from langchain.agents import AgentType\n",
"from langchain.chat_models import ChatOpenAI"

View File

@@ -29,7 +29,7 @@
"from langchain.agents import AgentExecutor, initialize_agent, AgentType\n",
"from langchain.schema import AgentFinish\n",
"from langchain.agents.tools import Tool\n",
"from langchain import LLMMathChain\n",
"from langchain.chains import LLMMathChain\n",
"from langchain.chat_models import ChatOpenAI"
]
},

View File

@@ -166,7 +166,7 @@
"from langchain.agents import AgentType\n",
"from langchain.tools import BaseTool\n",
"from langchain.llms import OpenAI\n",
"from langchain import LLMMathChain, SerpAPIWrapper"
"from langchain.chains import LLMMathChain\nfrom langchain.utilities import SerpAPIWrapper"
]
},
{

View File

@@ -47,7 +47,7 @@
}
],
"source": [
"from langchain import OpenAI, ConversationChain, LLMChain, PromptTemplate\n",
"from langchain.llms import OpenAI\nfrom langchain.chains import ConversationChain, LLMChain\nfrom langchain.prompts import PromptTemplate\n",
"from langchain.memory import ConversationBufferWindowMemory\n",
"\n",
"\n",

View File

@@ -26,7 +26,7 @@
"outputs": [],
"source": [
"from langchain.agents import Tool, AgentExecutor, BaseSingleActionAgent\n",
"from langchain import OpenAI, SerpAPIWrapper"
"from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper"
]
},
{

View File

@@ -38,7 +38,7 @@
" AgentOutputParser,\n",
")\n",
"from langchain.prompts import StringPromptTemplate\n",
"from langchain import OpenAI, SerpAPIWrapper, LLMChain\n",
"from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\nfrom langchain.chains import LLMChain\n",
"from typing import List, Union\n",
"from langchain.schema import AgentAction, AgentFinish\n",
"import re"

View File

@@ -48,7 +48,7 @@
"outputs": [],
"source": [
"from langchain.agents import ZeroShotAgent, Tool, AgentExecutor\n",
"from langchain import OpenAI, SerpAPIWrapper, LLMChain"
"from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper\nfrom langchain.chains import LLMChain"
]
},
{

View File

@@ -26,7 +26,7 @@
"outputs": [],
"source": [
"from langchain.agents import Tool, AgentExecutor, BaseMultiActionAgent\n",
"from langchain import OpenAI, SerpAPIWrapper"
"from langchain.llms import OpenAI\nfrom langchain.utilities import SerpAPIWrapper"
]
},
{

Some files were not shown because too many files have changed in this diff Show More