mirror of
https://github.com/hwchase17/langchain.git
synced 2025-07-04 12:18:24 +00:00
fix: update import paths for ChatOllama to use langchain_ollama instead of community (#31721)
This commit is contained in:
parent
e09abf8170
commit
3c3320ae30
@ -47,7 +47,7 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 1,
|
"execution_count": null,
|
||||||
"id": "6a75a5c6-34ee-4ab9-a664-d9b432d812ee",
|
"id": "6a75a5c6-34ee-4ab9-a664-d9b432d812ee",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [
|
"outputs": [
|
||||||
@ -61,7 +61,7 @@
|
|||||||
],
|
],
|
||||||
"source": [
|
"source": [
|
||||||
"# Local\n",
|
"# Local\n",
|
||||||
"from langchain_community.chat_models import ChatOllama\n",
|
"from langchain_ollama import ChatOllama\n",
|
||||||
"\n",
|
"\n",
|
||||||
"llama2_chat = ChatOllama(model=\"llama2:13b-chat\")\n",
|
"llama2_chat = ChatOllama(model=\"llama2:13b-chat\")\n",
|
||||||
"llama2_code = ChatOllama(model=\"codellama:7b-instruct\")\n",
|
"llama2_code = ChatOllama(model=\"codellama:7b-instruct\")\n",
|
||||||
|
@ -204,14 +204,14 @@
|
|||||||
},
|
},
|
||||||
{
|
{
|
||||||
"cell_type": "code",
|
"cell_type": "code",
|
||||||
"execution_count": 4,
|
"execution_count": null,
|
||||||
"id": "523e6ed2-2132-4748-bdb7-db765f20648d",
|
"id": "523e6ed2-2132-4748-bdb7-db765f20648d",
|
||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
"source": [
|
"source": [
|
||||||
"from langchain_community.chat_models import ChatOllama\n",
|
|
||||||
"from langchain_core.output_parsers import StrOutputParser\n",
|
"from langchain_core.output_parsers import StrOutputParser\n",
|
||||||
"from langchain_core.prompts import ChatPromptTemplate"
|
"from langchain_core.prompts import ChatPromptTemplate\n",
|
||||||
|
"from langchain_ollama import ChatOllama"
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
|
@ -215,8 +215,8 @@
|
|||||||
"metadata": {},
|
"metadata": {},
|
||||||
"outputs": [],
|
"outputs": [],
|
||||||
"source": [
|
"source": [
|
||||||
"from langchain_community.chat_models import ChatOllama\n",
|
|
||||||
"from langchain_core.prompts import ChatPromptTemplate\n",
|
"from langchain_core.prompts import ChatPromptTemplate\n",
|
||||||
|
"from langchain_ollama import ChatOllama\n",
|
||||||
"from langchain_openai import ChatOpenAI\n",
|
"from langchain_openai import ChatOpenAI\n",
|
||||||
"\n",
|
"\n",
|
||||||
"# Prompt\n",
|
"# Prompt\n",
|
||||||
|
@ -1505,7 +1505,7 @@ class Runnable(Generic[Input, Output], ABC):
|
|||||||
|
|
||||||
.. code-block:: python
|
.. code-block:: python
|
||||||
|
|
||||||
from langchain_community.chat_models import ChatOllama
|
from langchain_ollama import ChatOllama
|
||||||
from langchain_core.output_parsers import StrOutputParser
|
from langchain_core.output_parsers import StrOutputParser
|
||||||
|
|
||||||
llm = ChatOllama(model='llama2')
|
llm = ChatOllama(model='llama2')
|
||||||
|
Loading…
Reference in New Issue
Block a user